CN116777004B - Non-convex discrimination migration subspace learning method integrating distribution alignment information - Google Patents

Non-convex discrimination migration subspace learning method integrating distribution alignment information Download PDF

Info

Publication number
CN116777004B
CN116777004B CN202310779007.7A CN202310779007A CN116777004B CN 116777004 B CN116777004 B CN 116777004B CN 202310779007 A CN202310779007 A CN 202310779007A CN 116777004 B CN116777004 B CN 116777004B
Authority
CN
China
Prior art keywords
objective function
data
matrix
solution
source domain
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202310779007.7A
Other languages
Chinese (zh)
Other versions
CN116777004A (en
Inventor
罗廷金
刘玥瑛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
National University of Defense Technology
Original Assignee
National University of Defense Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by National University of Defense Technology filed Critical National University of Defense Technology
Priority to CN202310779007.7A priority Critical patent/CN116777004B/en
Publication of CN116777004A publication Critical patent/CN116777004A/en
Application granted granted Critical
Publication of CN116777004B publication Critical patent/CN116777004B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Abstract

The invention discloses a non-convex discrimination migration subspace learning method integrated with distribution alignment information, which comprises the following steps: s1, constructing an objective function by cutting the square of the Frobenius norm; s2, carrying out iterative solution on the objective function through an IALM algorithm to obtain a projection matrix P; s3, projecting the source domain data and the target domain data into a common characteristic subspace through a projection matrix P. Compared with trace norms, the invention adopts a non-convex regularization term which is more compact to rank function approximation as a proxy model, better approximates low-rank constraint by minimizing the first k minimum singular values of a reconstruction matrix, and aligns class information of data by minimizing joint distribution difference between a source domain and a target domain, so that the source domain data can reconstruct the target domain data better, and classification performance is improved.

Description

Non-convex discrimination migration subspace learning method integrating distribution alignment information
Technical Field
The invention relates to the technical field of non-supervision, in particular to a non-convex discrimination migration subspace learning method integrating distribution alignment information.
Background
With the development of big data age, in fields such as image recognition, natural language processing, medical health, etc., data often come from different domains, so that the distribution is different. Meanwhile, mass data is not generally marked. Therefore, how to classify these unlabeled data (i.e., target domain data) is a very important issue, and also draws attention from many expert scholars. Unsupervised domain adaptation provides a natural way to solve the above problem by deep mining the information in the tagged data (i.e., source domain data) to facilitate learning of the target classifier.
In the unsupervised field adaptation, the conventional methods can be divided into two main types, namely, distribution-based adaptation and migration subspace-based learning, and the first type of method reduces the difference between the two domains by jointly distributing Ji Yuanyu and the target domain. Whereas measuring the difference between two domains can be achieved by defining distances in different ways, the MMD criterion is widely adopted for its simplicity and a solid theoretical basis. There are many methods based on distributed adaptation, and several classical methods are presented below. Migration component analysis (TCA) maps data of two domains together into a high-dimensional regenerated kernel hilbert space when solving the problem of different source domain and target domain distributions. In this space, the edge distribution differences between the two are minimized while their respective internal properties are maximally preserved. On the basis, the Joint Distribution Adaptation (JDA) simultaneously considers the edge distribution difference and the class condition distribution difference of the data, and realizes finer alignment of the two-domain distribution from the whole to various types. Wherein the calculation of class conditional distribution differences requires the use of labels for the target domain data, which is precisely our task. To address this problem, JDA proposes the idea of pseudo tags. While taking into account the unreliability of the pseudo tag, an iterative update mechanism is employed to reduce adverse effects. JDA, however, does not make appropriate adjustments to the weights of the edge distribution differences and the condition distribution differences according to the application scenario. Such as when the data sets are more disparate, meaning that the edge distribution differences are greater; and when the data sets are similar, this means that the condition distribution requires more attention. Therefore, if both weights are set to a certain value, the performance of the algorithm is only degraded. For this idea Wang et al propose Balanced Distribution Adaptation (BDA). Meanwhile, the author also proposes a new weighted balance distribution adaptive (W-BDA) algorithm which solves the problem of class imbalance in migration by adaptively changing the weights of the classes. Furthermore, zhang et al believe that the previous approach is to calculate a weighted sum of the edge distribution differences and class condition distribution differences of the two domains, and the more natural measure is to directly calculate their joint probability distribution differences. The authors also propose that the previous approach only increases the mobility between domains, ignoring the discriminability between different classes of data, which may lead to a reduction in model classification performance, and therefore propose Joint Probability Distribution Adaptation (JPDA) to solve this problem. Wang et al search the nature of MMD through theoretical deduction, verify that minimizing the difference of class condition distribution is equivalent to minimizing the inter-class distance of source domain and target domain classes, and also equivalent to maximizing the data variance and inter-class distance of two domain classes, which intuitively illustrates the reason for the reduced feature discriminability in MMD. Through theoretical guidance, wang et al propose a new discriminant MMD with two parallel strategies, alleviating the negative impact of MMD on feature discriminant.
Compared with the first type of methods, the methods based on the migration subspace learning start from the geometric structural features of the data, so that the expression is more concise and effective, and the methods can be divided into two types: the manifold learning retention data structure is introduced by methods represented by geodesic flow Sampling (SGF) and geodesic flow kernel method (GFK). They consider the domains as one point in the glasman manifold, and connect to form a path through d intermediate points in the geodesic distance. Thus, a suitable transformation for each step is found, i.e. a transformation from the source domain to the target domain can be achieved. While another approach considers statistical features on Ji Yuanyu and target domain data to achieve knowledge migration, such as Subspace Alignment (SA), subspace Distribution Alignment (SDA), and variance correlation alignment (CORAL), among others. These methods are susceptible to noise and outliers, and Shao et al therefore propose low rank migration subspace learning (LTSL). The method migrates data in a source domain and a target domain into a unified generalized subspace, realizes linear reconstruction of the target domain sample through the source domain sample, and applies low-rank constraint to a reconstruction matrix to ensure the global structure of the data. In consideration of the influence of noise in the migration process, noise items are introduced to improve the robustness of the method. Fang et al then uses a least squares fit tag. Considering the adverse effect of tag noise on the model, the tag of the source domain data is relaxed, and the operation provides more freedom for adapting the tag and increases the margin between different types of data as much as possible. Discrimination migration subspace learning (DTSL) combines the two, and it is believed that better results can be obtained by only linearly representing the target samples by several similar source domain samples, rather than all the samples, when reconstructed. Therefore, sparse constraint is applied to the reconstruction matrix, and the local structure of the data is described. The joint feature selection and structure preservation method (FSSP) and the joint low rank representation and feature selection (JLRFS) method then impose structured sparse constraints on the projection matrix, causing the model to screen more critical features. And meanwhile, the drawing Laplace term is introduced, so that the data structure is better reserved.
However, these methods still have many limitations, first, the methods based on migration subspace learning typically impose low rank constraints on reconstruction matrices to preserve the global structure of the data. Since rank minimization problem is NP-hard, many scholars approximate rank functions by employing trace norms. However, this approximation method is significantly affected by the maximum singular value, and the rank function is not, so its solution may lead to a serious deviation from the optimal solution. Secondly, the traditional migration subspace learning method only uses the characteristic information of the data to reconstruct the target sample, but ignores the category information of the data. If the target samples are represented linearly by source data of similar characteristics but different classes, it is difficult to split. In order to solve the problems, the invention provides a non-convex discriminant migration subspace learning method integrated with distribution alignment information to solve the problems.
Disclosure of Invention
The invention aims to provide a non-convex discriminant migration subspace learning method integrated with distribution alignment information, so as to overcome the defects existing in the prior art.
In order to achieve the above purpose, the technical scheme adopted by the invention is as follows:
the non-convex discrimination migration subspace learning method integrated with the distribution alignment information is used for image recognition and comprises the following steps of:
s1, constructing an objective function by cutting the square of the Frobenius norm;
s2, carrying out iterative solution on the objective function through an IALM algorithm to obtain a projection matrix P;
s3, projecting the source domain data and the target domain data into a common characteristic subspace through a projection matrix P.
Further, the step S1 specifically includes:
s10, linearly representing the target domain data by the source domain data as follows:
P T X t =P T X s Z+E
wherein X is t And X s Feature matrices of the target domain data and the source domain data respectively; wherein n is t And n s The data quantity of the two domains respectively; z is a reconstruction matrix and E is a noise matrix.
S11, applying low-rank constraint on Z to maintain global structure of data, and applying l 1 The norm promotes the sparsity on Z to ensure the local structure of the data, and simultaneously applies sparsity constraint to E to obtain an objective function as follows:
wherein alpha, beta, lambda are regularization parameters;
s12, the square of the Frobenius norm is truncated to promote the first k minimum singular values to be 0 to approximate a low-rank condition, and an objective function is converted into:
where k=n s -rank(Z),Is a discriminant subspace learning function term that enhances the discriminant of data by fitting source domain labels in subspaces, using the relaxed labels y+=y s +B.sup.M to eliminate the influence of tag noise on the model, Y s For the single thermal encoding of source domain data, M is a non-negative matrix, and by Hadamard operator, matrix B is:
s13, applying structured sparsity on the projection matrix 1,2 The norm converts the objective function in step S12 into:
wherein, gamma is a balance factor;
s14, reducing the distance between a source domain and a target domain in the subspace by learning P, and obtaining the following formula based on a distribution self-adaption method:
where ω is the balance factor,the number of class c samples in the source domain and the target domain, x= [ X ] s ,X t ],M 0 ,M c The definition is as follows
S15, calculating the formula in the step S14 by adopting the pseudo tag, and carrying out iterative refinement on the pseudo tag to enable the pseudo tag to be made The final objective function is:
further, the step S2 specifically includes:
s20, introducing matrixes Q and J, wherein the solution can be equivalently obtained into an objective function by minimizing the augmented Lagrangian function:
wherein Y is 1 ,Y 2 ,Y 3 Is Lagrangian multiplier, μ is penalty parameter;
s21, fixing matrices Q, Z, J, E and M, and rewriting the objective function formula of step S20 to:
by calculating the partial derivative of the objective function in step S21 to make it 0, a closed-form solution of the matrix P can be obtained:
wherein G is 1 =Y s +B⊙M,G 2 =X t -X s Z;
S22, fixing matrices P, Z, J, E and M, updating the objective function of Q as:
wherein the method comprises the steps ofWith a contraction operator, the solution for Q can be expressed as:
wherein [ A ]] :,i Representing the ith column of matrix a;
s23, fixing matrices P, Q, J, E and M, updating the objective function of Z as:
due toIs NP-difficult, based on the Ky Fan theory, which is equivalent toThe objective function may be rewritten as:
and (3) carrying out iterative computation on F and Z until the objective function converges to obtain an optimal solution of Z, and then updating in the step S24, instead of only iterating F and Z once, enabling the partial derivative of the formula to be 0, so as to obtain a closed solution of Z as follows:
wherein,in calculating F, the following formula is optimized:
the optimal solution of F is formed by k left singular vectors corresponding to k minimum singular values of Z, namely U 2 Wherein U is a left singular vector matrix obtained by SVD decomposition of Z, and U= [ U 1 ,U 2 ], Since only FF is needed in computing the closed-form solution of Z T Calculated by the following formula:
s24, fixing matrices P, Q, Z, E and M, updating the objective function of J as:
the iterative solution of J is:
where shrnk (x, c) =sign max (|x| -c, 0).
S25, fixing the matrices P, Q, Z, J and M, updating the objective function of E as:
the iterative solution of E is:
s26, fixing matrices P, Q, Z, J and E, updating the objective function of M to be
Let r=p T X s -Y s The objective function is decomposed into d×n s Sub-optimization problemWherein d, n s The number of rows and columns of the non-negative loose label matrix M are respectively, and the optimal solution is M ij =max(R ij B ij 0), the closed-form solution of the objective function is:
M * =max(R⊙B,0)
s27, updating Lagrangian multiplier Y 1 ,Y 2 ,Y 3 Penalty factor μ:
wherein ρ is learning rate, μ max Is the maximum value that mu can take.
Compared with the prior art, the invention has the advantages that: compared with trace norms, the invention adopts a non-convex regularization term which is more compact to rank function approximation as a proxy model, better approximates low-rank constraint by minimizing the first k minimum singular values of a reconstruction matrix, and aligns class information of data by minimizing joint distribution difference between a source domain and a target domain, so that the source domain data can reconstruct the target domain data better, and classification performance is improved.
Drawings
In order to more clearly illustrate the embodiments of the invention or the technical solutions in the prior art, the drawings that are required in the embodiments or the description of the prior art will be briefly described, it being obvious that the drawings in the following description are only some embodiments of the invention, and that other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is an iterative optimization solution diagram of a non-convex discriminating migration subspace learning method incorporating distribution alignment information of the present invention.
Detailed Description
The preferred embodiments of the present invention will be described in detail below with reference to the accompanying drawings so that the advantages and features of the present invention can be more easily understood by those skilled in the art, thereby making clear and defining the scope of the present invention.
Conventional migratory subspace learning methods typically employ a trace norm to approximate a low rank, however it is too relaxed, resulting in its solution being severely biased from the ideal solution. In addition, the conventional method ignores category information of the alignment data, so that the category performance is reduced. Therefore, the present embodiment proposes a non-convex discrimination migration subspace learning method incorporating distribution alignment information to solve these problems. Specifically, the object of the present embodiment is to learn the projection matrix P to project the source domain and the target domain into one common subspace.
Referring to fig. 1, the embodiment discloses a non-convex discrimination migration subspace learning method integrated with distribution alignment information, which is used for image recognition and comprises the following steps:
step S1, constructing an objective function by cutting off the square of the Frobenius norm, wherein the method specifically comprises the following steps:
in step S10, since the two domain data are in the same manifold, the target domain data can be linearly represented by the source domain data, and the original formula is written as follows:
P T X t =P T X s Z+E (1)
wherein X is t And X s Feature matrices of the target domain data and the source domain data respectively; wherein n is t And n s The data quantity of the two domains respectively; z is a reconstruction matrix and E is a noise matrix.
Step S11, each target sample may be represented linearly by a combination of its source domain neighbor samples. Thus, a low rank constraint is imposed on the reconstruction matrix Z to preserve this property and preserve the global structure of the data. At the same time, the reconstruction of each target sample only needs few source domain samples to participate, and l is applied 1 The norm promotes the sparseness of the data on Z to ensure the local structure of the data, and the obtained objective function is as follows:
where α, β, λ are regularization parameters.
In step S12, since the trace norm is the most tightly convex approximation of the rank function on the unit sphere, many expert scholars approximate the low rank condition as a proxy model. However, according to the definition of the trace norm, it varies drastically with a significant change in the maximum singular value of the matrix, but the rank of the matrix does not change. Thus, the relaxation of the trace norms may cause the solution to deviate strictly from the ideal solution. In the embodiment, the square of the Frobenius norm is truncated, so that the first k minimum singular values are caused to be 0 to approximate a low-rank condition, the influence of other larger singular values on a proxy model is eliminated, and the objective function (2) is converted into:
where k=n s -rank(Z),Is a discriminant subspace learning function term that enhances the discriminant of data by fitting source domain labels in subspaces, using the relaxed labels y+=y s +B.sup.M to eliminate the influence of tag noise on the model, Y s For the single thermal encoding of source domain data, M is a non-negative matrix, and by Hadamard operator, matrix B is:
step S13, additionally, the selected features may be redundant, thus requiring the application of a structured sparsity l to the projection matrix 1,2 The norm converts the objective function in step S12 into:
wherein γ is a balance factor.
In step S14, however, the conventional method and the above formula only use the characteristic information of the data to disregard the influence of the data class information on the model for the subspace of Ji Yuanyu and the target domain. Therefore, the embodiment utilizes semantic information of different categories to reduce the distribution difference between two domains, thereby improving the classification performance of the model. Specifically, the present embodiment reduces the two-domain distance in the subspace by learning P, and obtains the following formula based on the distributed adaptive method:
where ω is the balance factor,the number of class c samples in the source domain and the target domain, x= [ X ] s ,X t ],M 0 ,M c The definition is as follows:
step S15, however, in the calculation of equation (5), since the target field is not labeled, it cannot be known whether the jth sample in the target field belongs to the c-th class, i.e.Unknown. Thus, a pseudo tag is employed for the calculation. Because the reliability of the pseudo tag is lower, the pseudo tag is iteratively refined to make +.>The final objective function is:
and S2, carrying out iterative solution on the objective function through an IALM algorithm to obtain a projection matrix P.
Specifically, the present embodiment designs the IALM algorithm to effectively solve the objective function (6). The step S2 specifically includes:
step S20, introducing matrices Q and J, the solution of which can be equivalently obtained by minimizing the augmented lagrangian function (7):
wherein Y is 1 ,Y 2 ,Y 3 Is Lagrangian multiplier, μ is penalty parameter;
step S21, fixing the matrices Q, Z, J, E and M, and rewriting the objective function formula of step S20 to:
by calculating the partial derivative of the objective function (8) to be 0, the closed-form solution of the matrix P can be obtained:
wherein G is 1 =Y s +B⊙M,G 2 =X t -X s Z;
Step S22, fixing the matrices P, Z, J, E and M, updating the objective function of Q as:
wherein the method comprises the steps ofWith a contraction operator, the solution for Q can be expressed as:
wherein [ A ]] :,i Representing the ith column of matrix a;
step S23, fixing the matrices P, Q, J, E and M, updating the objective function of Z as:
due toIs NP-difficult, based on the Ky Fan theory, which is equivalent toThe objective function may be rewritten as:
and (3) carrying out iterative computation on F and Z until the objective function converges to obtain an optimal solution of Z, and then updating in the step S24, instead of only iterating F and Z once, so that the partial derivative of the formula (13) is 0, and obtaining a closed solution of Z as follows:
wherein,in calculating F, the following formula is optimized:
the optimal solution of F is formed by k left singular vectors corresponding to k minimum singular values of Z, namely U 2 Wherein U is a left singular vector matrix obtained by SVD decomposition of Z, and U= [ U 1 ,U 2 ], Since only FF is needed in computing the closed-form solution of Z T Calculated by the following formula:
step S24, fixing the matrices P, Q, Z, E and M, updating the objective function of J as:
the iterative solution of J is:
where shrnk (x, c) =sign max (|x| -c, 0).
Step S25, fixing the matrices P, Q, Z, J and M, updating the objective function of E as:
the iterative solution of E is:
step S26, fixing matrices P, Q, Z, J and E, updating the objective function of M to
Let r=p T X s -Y s The objective function is decomposed into d×n s Sub-optimization problemWherein d, n s The number of rows and columns of the non-negative loose label matrix M are respectively, and the optimal solution is M ij =max(R ij B ij 0), the closed-form solution of the objective function is:
M * =max(R⊙B,0) (21)
step S27, updating Lagrangian multiplier Y 1 ,Y 2 ,Y 3 Penalty factor μ:
wherein ρ is learning rate, μ max Is the maximum value that mu can take.
In summary, the various optimization steps of the model are summarized in table 1 below:
and S3, projecting the source domain data and the target domain data into a common characteristic subspace through a projection matrix P.
The present embodiment also improves the algorithm described above. Another variation is employed when approximating the rank functionWhich is constructed by truncating the trace norms. Compared with the approximation in the previous text->Variant->The method is closer to the rank function in formula definition, so that the approximation effect is better. The objective function can be obtained as follows:
analysis shows that the proxy model that changes the low-rank condition only affects the computation of the reconstruction matrix Z, so that the objective function transitions to:
in the same way as described above,it is NP-hard to solve. Fortunately, its equivalence can be converted to the following formula
Where r=rank (Z). While in optimizing trace norms, the present invention converts it to Tr (Z T DZ), iteratively solving. Wherein the method comprises the steps of Representing the current optimal solution. Thus, the objective function (24) can be converted into
When the matrices F and G are fixed, Z is updated. The closed-form solution for Z can be obtained as follows:
when Z is fixed, the optimal solutions of F and G are respectively composed of left and right singular vectors corresponding to the first r maximum singular values of Z.
Compared with trace norms, the invention adopts a non-convex regularization term which is more compact to rank function approximation as a proxy model, better approximates low-rank constraint by minimizing the first k minimum singular values of a reconstruction matrix, and aligns class information of data by minimizing joint distribution difference between a source domain and a target domain, so that the source domain data can reconstruct the target domain data better, and classification performance is improved.
Although the embodiments of the present invention have been described with reference to the accompanying drawings, the patentees may make various modifications or alterations within the scope of the appended claims, and are intended to be within the scope of the invention as described in the claims.

Claims (1)

1. The non-convex discrimination migration subspace learning method integrated with the distribution alignment information is used for image recognition and is characterized by comprising the following steps of:
s1, constructing an objective function by cutting the square of the Frobenius norm;
s2, carrying out iterative solution on the objective function through an IALM algorithm to obtain a projection matrix P;
s3, projecting the source domain data and the target domain data into a common feature subspace through a projection matrix P;
the step S1 specifically includes:
s10, linearly representing the target domain data by the source domain data as follows:
P T X t =P T X s Z+E
wherein X is t And X s Feature matrices of the target domain data and the source domain data respectively; wherein n is t And n s The data quantity of the two domains respectively; z is a reconstruction matrix, E is a noise matrix;
s11, applying low-rank constraint on Z to maintain global structure of data, and applying l 1 The norm promotes the sparsity on Z to ensure the local structure of the data, and simultaneously applies sparsity constraint to E to obtain an objective function as follows:
wherein alpha, beta, lambda are regularization parameters;
s12, the square of the Frobenius norm is truncated to promote the first k minimum singular values to be 0 to approximate a low-rank condition, and an objective function is converted into:
where k=n s -rank(Z),Is a discriminant subspace learning function term that enhances the discriminant of data by fitting source domain labels in subspaces, using the relaxed labels y+=y s +B.sup.M to eliminate the influence of tag noise on the model, Y s For the single thermal encoding of source domain data, M is a non-negative matrix, and by Hadamard operator, matrix B is:
s13, applying structured sparsity on the projection matrix 1,2 The norm converts the objective function in step S12 into:
wherein, gamma is a balance factor;
s14, reducing the distance between a source domain and a target domain in the subspace by learning P, and obtaining the following formula based on a distribution self-adaption method:
where ω is the balance factor,the number of class c samples in the source domain and the target domain, x= [ X ] s ,X t ],M 0 ,M c The definition is as follows:
s15, calculating the formula in the step S14 by adopting the pseudo tag, and carrying out iterative refinement on the pseudo tag to enable the pseudo tag to be madeThe final objective function is:
the step S2 specifically includes:
s20, introducing matrixes Q and J, wherein the solution can be equivalently obtained into an objective function by minimizing the augmented Lagrangian function:
wherein Y is 1 ,Y 2 ,Y 3 Is Lagrangian multiplier, μ is penalty parameter;
initializing: p=p 0 ,Q=0,M=1,J=Z=0,E=0,Y 1 =0,Y 2 =0,Y 3 =0,μ 0 =0.05,μ max =10 6 ,ρ=1.1,ε=10 -6 Repeating the following steps until the objective function meets the convergence condition;
s21, fixing matrices Q, Z, J, E and M, and rewriting the objective function formula of step S20 to:
by calculating the partial derivative of the objective function in step S21 to make it 0, a closed-form solution of the matrix P can be obtained:
wherein G is 1 =Y s +B⊙M,G 2 =X t -X s Z;
S22, fixing matrices P, Z, J, E and M, updating the objective function of Q as:
wherein the method comprises the steps ofWith a contraction operator, the solution for Q can be expressed as:
wherein [ A ]] :,i Representing the ith column of matrix a;
s23, fixing matrices P, Q, J, E and M, updating the objective function of Z as:
due toIs NP-difficult, which is equivalent to +.>The objective function may be rewritten as:
and (3) carrying out iterative computation on F and Z until the objective function converges to obtain an optimal solution of Z, and then updating in the step S24, instead of only iterating F and Z once, enabling the partial derivative of the formula to be 0, so as to obtain a closed solution of Z as follows:
wherein,in calculating F, the following formula is optimized:
the optimal solution of F is formed by k left singular vectors corresponding to k minimum singular values of Z, namely U 2 Wherein U is a left singular vector matrix obtained by SVD decomposition of Z, and U= [ U 1 ,U 2 ], Since only FF is needed in computing the closed-form solution of Z T Calculated by the following formula:
s24, fixing matrices P, Q, Z, E and M, updating the objective function of J as:
the iterative solution of J is:
where shrnk (x, c) =sign max (|x| -c, 0);
s25, fixing the matrices P, Q, Z, J and M, updating the objective function of E as:
the iterative solution of E is:
s26, fixing matrices P, Q, Z, J and E, updating the objective function of M to be
Let r=p T X s -Y s The objective function is decomposed into d×n s Sub-optimization problemWherein d, n s The number of rows and columns of the non-negative loose label matrix M are respectively, and the optimal solution is M ij =max(R ij B ij 0), the closed-form solution of the objective function is:
M * =max(R⊙B,0)
s27, updating Lagrangian multiplier Y 1 ,Y 2 ,Y 3 Penalty factor μ:
wherein ρ is learning rate, μ max Is the maximum value that mu can take.
CN202310779007.7A 2023-06-29 2023-06-29 Non-convex discrimination migration subspace learning method integrating distribution alignment information Active CN116777004B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310779007.7A CN116777004B (en) 2023-06-29 2023-06-29 Non-convex discrimination migration subspace learning method integrating distribution alignment information

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310779007.7A CN116777004B (en) 2023-06-29 2023-06-29 Non-convex discrimination migration subspace learning method integrating distribution alignment information

Publications (2)

Publication Number Publication Date
CN116777004A CN116777004A (en) 2023-09-19
CN116777004B true CN116777004B (en) 2024-02-06

Family

ID=87992694

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310779007.7A Active CN116777004B (en) 2023-06-29 2023-06-29 Non-convex discrimination migration subspace learning method integrating distribution alignment information

Country Status (1)

Country Link
CN (1) CN116777004B (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109348229A (en) * 2018-10-11 2019-02-15 武汉大学 Jpeg image mismatch steganalysis method based on the migration of heterogeneous characteristic subspace
CN110659663A (en) * 2019-03-22 2020-01-07 重庆大学 Unsupervised bidirectional reconstruction field self-adaption method
CN114548264A (en) * 2022-02-21 2022-05-27 重庆邮电大学 Unsupervised two-stage field self-adaption method
CN115861708A (en) * 2022-12-20 2023-03-28 重庆邮电大学 Low-rank sparse representation transfer learning method with adaptive graph diffusion
CN116127357A (en) * 2023-02-06 2023-05-16 徐州医科大学 Vibration signal domain adaptation diagnosis method based on Grassmann manifold subspace embedding
CN116701891A (en) * 2023-04-24 2023-09-05 中国人民解放军国防科技大学 Schatten-p norm-based low-rank discrimination migration subspace learning method

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8935308B2 (en) * 2012-01-20 2015-01-13 Mitsubishi Electric Research Laboratories, Inc. Method for recovering low-rank matrices and subspaces from data in high-dimensional matrices
US11534641B2 (en) * 2017-05-03 2022-12-27 Tyco Fire Products Lp Sectional fire protection for attic spaces
US20190102511A1 (en) * 2017-10-02 2019-04-04 Blackthorn Therapeutics, Inc. Methods and tools for detecting, diagnosing, predicting, prognosticating, or treating a neurobehavioral phenotype in a subject
US20220341725A1 (en) * 2021-04-26 2022-10-27 Quality Vision International Inc. Non-invasive alignment method and system for imager-illuminator optical measurement machines

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109348229A (en) * 2018-10-11 2019-02-15 武汉大学 Jpeg image mismatch steganalysis method based on the migration of heterogeneous characteristic subspace
CN110659663A (en) * 2019-03-22 2020-01-07 重庆大学 Unsupervised bidirectional reconstruction field self-adaption method
CN114548264A (en) * 2022-02-21 2022-05-27 重庆邮电大学 Unsupervised two-stage field self-adaption method
CN115861708A (en) * 2022-12-20 2023-03-28 重庆邮电大学 Low-rank sparse representation transfer learning method with adaptive graph diffusion
CN116127357A (en) * 2023-02-06 2023-05-16 徐州医科大学 Vibration signal domain adaptation diagnosis method based on Grassmann manifold subspace embedding
CN116701891A (en) * 2023-04-24 2023-09-05 中国人民解放军国防科技大学 Schatten-p norm-based low-rank discrimination migration subspace learning method

Non-Patent Citations (10)

* Cited by examiner, † Cited by third party
Title
Feiping Nie 等.Unsupervised Feature Selection With Constrained l2,0-Norm and Optimized Graph.IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS.2022,第33卷(第4期),1702-1713. *
Global convergence of ADMM in nonconvex nonsmooth optimization;Wang Yu 等;Journal of Scientific Computing;第78卷;29-63 *
New tight relaxations of rank minimization for multi-task learning;Chang Wei 等;ACM International Conference on Information & Knowledge Management;2910-2914 *
Non-Convex Transfer Subspace Learning for Unsupervised Domain Adaptation;Z. Lin 等;2019 IEEE International Conference on Multimedia and Expo (ICME);1468-1473 *
Rethinking maximum mean discrepancy for visual domain adaptation", IEEE Transactions on Neural Networks and Learning Systems;Wei Wang 等;IEEE Transactions on Neural Networks and Learning Systems;第34卷(第1期);264 - 277 *
Structure preservation and distribution alignment in discriminative transfer subspace learning;Xiao Ting等;Neurocomputing;第337卷;218-234 *
Unsupervised domain adaptation via re-weighted transfer subspace learning with inter-class sparsity;Liran Yang 等;Knowledge-Based Systems;第263卷;1-6 *
基于多源域领域适应字典学习和稀疏表示的脑电情感识别方法;袁凯烽 等;南京信息工程大学学报(自然科学版);第15卷(第4期);412-418 *
基于稀疏优化的学习方法及应用研究;罗廷金;中国博士学位论文全文数据库 (信息科技辑)(第(2020)02期);I138-48 *
基于迁移子空间的半监督领域自适应方法;陶洋 等;计算机工程与设计;第42卷(第8期);2308-2315 *

Also Published As

Publication number Publication date
CN116777004A (en) 2023-09-19

Similar Documents

Publication Publication Date Title
Xu et al. A survey on multi-view learning
Que et al. Back to the future: Radial basis function networks revisited
Weston et al. Connecting language and knowledge bases with embedding models for relation extraction
Zhang et al. Neural networks incorporating dictionaries for Chinese word segmentation
WO2023092961A1 (en) Semi-supervised method and apparatus for public opinion text analysis
CN109784405B (en) Cross-modal retrieval method and system based on pseudo-tag learning and semantic consistency
Cascianelli et al. Full-GRU natural language video description for service robotics applications
CN113535953B (en) Meta learning-based few-sample classification method
Han et al. A survey of multi-label classification based on supervised and semi-supervised learning
Ma et al. Extraction of temporal information from social media messages using the BERT model
Kutuzova et al. Multimodal variational autoencoders for semi-supervised learning: In defense of product-of-experts
Pereverzyev An introduction to artificial intelligence based on reproducing kernel Hilbert spaces
Xu et al. Cycle-reconstructive subspace learning with class discriminability for unsupervised domain adaptation
Xie et al. Semi-supervised partial multi-label learning
Luo et al. Self-supervised learning for semi-supervised temporal language grounding
Lin et al. A deep clustering algorithm based on gaussian mixture model
Sun et al. Partial multi-label learning with noisy side information
Shi et al. Unsupervised domain adaptation based on adaptive local manifold learning
Luo et al. Discriminative label consistent domain adaptation
CN116777004B (en) Non-convex discrimination migration subspace learning method integrating distribution alignment information
Zhang et al. End-to-end temporal attention extraction and human action recognition
CN116701891A (en) Schatten-p norm-based low-rank discrimination migration subspace learning method
WO2023000725A1 (en) Named entity identification method and apparatus for electric power measurement, and computer device
CN114239575B (en) Statement analysis model construction method, statement analysis method, device, medium and computing equipment
Zhu et al. Fast orthogonal locality-preserving projections for unsupervised feature selection

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant