CN107798349B - Transfer learning method based on depth sparse self-coding machine - Google Patents

Transfer learning method based on depth sparse self-coding machine Download PDF

Info

Publication number
CN107798349B
CN107798349B CN201711069171.XA CN201711069171A CN107798349B CN 107798349 B CN107798349 B CN 107798349B CN 201711069171 A CN201711069171 A CN 201711069171A CN 107798349 B CN107798349 B CN 107798349B
Authority
CN
China
Prior art keywords
formula
matrix
data
representing
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201711069171.XA
Other languages
Chinese (zh)
Other versions
CN107798349A (en
Inventor
胡学钢
张玉红
朱毅
李培培
周鹏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hefei University of Technology
Original Assignee
Hefei University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hefei University of Technology filed Critical Hefei University of Technology
Priority to CN201711069171.XA priority Critical patent/CN107798349B/en
Publication of CN107798349A publication Critical patent/CN107798349A/en
Application granted granted Critical
Publication of CN107798349B publication Critical patent/CN107798349B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • G06F18/2136Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on sparsity criteria, e.g. with an overcomplete basis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/217Validation; Performance evaluation; Active pattern learning techniques

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Artificial Intelligence (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a deep sparse self-coding machine-based transfer learning method which comprises the steps of (1) preprocessing a data set vectorization earlier stage, (2) designing and realizing a model, (3) carrying out semi-supervised learning on extracted features obtained by applying a Stacked RICA algorithm, (4) training a classifier on a training set by using a logistic regression analysis model (L R) after the feature extraction is finished, (5) carrying out classification prediction on a test set by using the classifier on the training set, (6) finishing the classification on the test set to obtain a final transfer learning result.

Description

Transfer learning method based on depth sparse self-coding machine
Technical Field
The invention relates to the field of feature extraction and transfer learning methods, in particular to a transfer learning method based on a deep sparse self-coding machine.
Background
Traditional machine learning has achieved significant success in many areas. However, many machine learning algorithms are based on the assumption that the training set and the test set are homologized, independent and homologized, and most machine learning needs to recapture data when the data set distribution changes, which requires a large amount of training data to be collected again. In real-world applications, the environment is constantly changing, data is re-collected, and the model is retrained for each new scenario encountered by the learning system, which is very costly and impractical. It is desirable that the learning system automatically adapt to changes in the environment with little retraining data and retraining time. Under the condition, the migration knowledge which is obtained from the former scene and can be applied to the new scene can help us to speed up the learning process, reduce the cost of collecting new training data and achieve the goal of migration learning. Migration learning emphasizes the transfer of knowledge across domains, tasks, and distributions, which are similar but not identical. For example, learning to recognize an apple may help to learn a pear, or learning to play an electronic organ may help to learn a piano. The research of transfer learning is essentially that people always apply the original knowledge to solve new problems more quickly.
In recent years, Deep learning (Deep L earning) has been used to extract features in the fields of images, text, audio, etc. in terms of human perception, the information processing of the human visual system is hierarchical, edge features are extracted from a low-level V1 region, then the shape of a V2 region or parts of a target, and then the whole target, the behavior of the target, etc. in other words, the features of a high-level are the combination of features of a low-level, and the features from a low-level to a high-level are more abstract and more capable of expressing semantics or intentions.
A sparse self-coding machine is a method for extracting data features. This has the advantage that a set of linearly independent over-complete bases (over-completed bases) can be extracted to reconstruct the sample. The general model for extracting the feature base vectors can only ensure that the base vectors are linearly uncorrelated, and the model cannot be well applied in some applications. For example, some audio is collected, and the audio has personal voices, which are independent of each other, and we want to separate the audio of each person, so the model is disabled at this time. We use the RICA (Reconstruction Independent Component analysis) algorithm, and the goal is to learn a set of mutually Independent overcomplete bases.
The deep sparse self-coding machine is based on the thought of deep learning, the sparse self-coding machine is used as one layer of a model to be superposed, namely the result of the sparse self-coding machine at the previous layer is output and is used as the input of the next layer, so that a multi-layer deep learning structure is formed, and more useful characteristics are extracted. And then, semi-supervised learning is performed according to the extracted features, so that the precision and the accuracy of the transfer learning are improved.
In the research on the feature extraction and the migration learning method, the existing methods are all researches carried out by using self-coding models, and the research work of using sparse coding models is very little. Sparse coding is one of effective means for reducing dimensions in the fields of images, texts and the like, but the application of sparse coding in field adaptation has some problems, and the common problems are as follows: (a) the problem of non-independent linearity between characteristic basis vectors; (b) the problem of the use of tags in the source domain; (c) and (4) the objective function bias term after superposition. If the problems cannot be solved well, the accuracy of feature extraction and transfer learning is inevitably affected, and the invention provides a solution to the problems.
Disclosure of Invention
The invention aims to provide a transfer learning method based on a depth sparse self-coding machine, which aims to solve the problems of the prior art in feature extraction and transfer learning methods.
In order to achieve the purpose, the technical scheme adopted by the invention is as follows:
a transfer learning method based on a depth sparse self-coding machine is characterized by comprising the following steps: the method sequentially comprises the following steps:
(1) and whitening preprocessing is carried out on all images in the image database, and the process is as follows:
(1.1) representing the input dataset as { x }(1),x(2),...,x(n)Calculating the covariance matrix of x
Figure BDA0001456490160000021
Then calculating the characteristic direction of the covariance matrixThe quantity, in terms of column composition matrix U, is shown by the following equation:
Figure BDA0001456490160000031
in the matrix U, U1Is the principal eigenvector, which corresponds to the largest eigenvalue, u2Is a sub-feature vector, analogizes with the same, and records λ12,...,λnCharacteristic values corresponding to all the vector quantities in the matrix U are obtained;
(1.2) representing the input data by the calculated matrix U as shown in the following equation:
Figure BDA0001456490160000032
wherein the subscript rot refers to rotation, which means that it is the result of the rotation processing of the original data, in order to make each input feature have unit variance, the use of
Figure BDA0001456490160000033
Scaling each feature x as a scaling factorrot,iWherein the resulting PCA whitened data is represented by the following formula:
Figure BDA0001456490160000034
(1.3) let R be any orthogonal matrix, i.e. satisfy RRT=RTR is I, then RxPCAwhiteStill with unit covariance, in order to make the input data after whitening processing for all possible R as close to the original input data as possible, let R be U, resulting in formula (1):
xZCAwhite=UxPCAwhite(1),
xZCAwhitenamely processing data of original input data obtained after ZCA whitening;
(2) constructing a depth sparse self-coding machine model to extract high-level abstract features of the image, wherein the process is as follows:
(2.1) constructing a sparse self-encoding machine model, comprising the following steps:
(2.1.1), the sparse self-coding model uses a Reconstruction Independent component analysis algorithm, namely RICA algorithm, and x is obtained by using a formula (1)ZCAwhiteIs input data of the RICA algorithm and substitutes the cost function formula (2):
Figure BDA0001456490160000041
in the cost function formula (2), x is the input data, i.e. xZCAwhiteW is a weighting matrix;
(2.1.2) obtaining the partial derivative of x for the cost function formula (2), wherein the partial derivative of the first term of the cost function formula (2) is obtained
Figure BDA0001456490160000042
As the partial derivative function, the obtained partial derivative function is shown in formula (3):
Figure BDA0001456490160000043
(2.1.3) iteratively calculating a weighting matrix W by utilizing an L-FBGS algorithm to obtain a trained sparse self-coding model.
(2.2) constructing a deep sparse self-coding machine model:
substituting the weighting matrix W obtained in the step (2.1) into the cost function formula (2), and recording the obtained output as
Figure BDA0001456490160000044
Is output data obtained after the training of the single-layer RICA model is finished, and the output data is obtained
Figure BDA0001456490160000045
Repeating step (2.1) as input data to obtain W(i)Training a weighting matrix obtained after stacking the sparse self-coding machine, wherein i is the number of times of the iteration step (2.1);
(2.3) extracting features according to the trained deep sparse self-coding machine model;
pooling the square root of the model square root with the weighting matrix W obtained in step (2.1)(i)Substituting the formula (4) for convolution feature extraction, wherein the formula (4) is as follows:
Figure BDA0001456490160000046
in the formula (4), the first and second groups,
Figure BDA0001456490160000047
represents the input of the l-th layer in the convolutional network,
Figure BDA0001456490160000048
the error term for layer l +1 representing the kth feature in the convolutional network, the output from equation (4) is denoted as xfeaAbstract features extracted from the raw input data;
(3) and optimizing features by semi-supervised learning:
using x obtained in step (2)feaPerforming semi-supervised learning for input to obtain a formula (5), adding K L distance of source domain distribution and target domain distribution and multi-class regression bias terms according to source domain class labels,
Figure BDA0001456490160000051
denotes the output, W, obtained after semi-supervised learningSSLRepresenting a weight matrix in semi-supervised learning, ξ(s)Output representing hidden layers in the source domain, ξ(t)Representing the output of the hidden layer in the target domain, equation (5) is as follows:
Figure BDA0001456490160000052
in the formula (5), the first and second groups,
Figure BDA0001456490160000053
representing the reconstruction error from the original data to the data re-represented after feature extraction;
Figure BDA0001456490160000054
a K L distance representing the source domain distribution and the target domain distribution;
Figure BDA0001456490160000055
representing multi-class regression bias terms made according to the source domain class labels;
Figure BDA0001456490160000056
matrix W representing characteristic parametersSSLThe constraint term of (2);
(4) training a classifier and classifying the test image data set, wherein the process is as follows:
(4.1) training L R classifier with test image data set, in L R classification, note:
Figure BDA0001456490160000057
in formula (6), scale
Figure BDA0001456490160000058
Outputting the finished product of the step (3) for sigmoid function
Figure BDA0001456490160000059
And a label-substitution-in (6) training classifier of y, the training image dataset, of known labels;
(4.2) classifying the test image data set by using the trained classifier; outputting the finished product of the step (3)
Figure BDA0001456490160000061
The output of the test image data set of (1) is substituted for the L R classifier completed in formula (6), and the classification result T of the test image data set is obtainedtestAs shown in equation (7):
Ttest=argmaxP(x) (7)。
the invention provides a transfer learning method based on a depth sparse self-coding machine. According to the method, from the aspect of deep learning, a sparse self-coding machine model applying an RICA algorithm is applied to data set feature extraction, and through a multi-layer superposition idea of deep learning, a deep sparse self-coding machine is constructed through a Stacked RICA algorithm and a linear independent over-complete feature basis vector is trained and extracted. And on the basis of the feature basis vector, a semi-supervised learning method is applied to add the source domain class labels and the bias terms of the multi-class regression, so that the extracted features are further optimized. And finally, training a classifier by applying a support vector machine model according to the extracted features to realize classification prediction of the target domain and finish the target of transfer learning. The method can extract more useful features in the data set, improves the classification precision in the target domain, and obviously improves the accuracy and precision of the transfer learning.
The invention solves the important practical problem of feature extraction and transfer learning, the research result can be directly applied to image classification, text classification, emotion transfer and other applications, and can be expanded to be applied to a plurality of fields such as audio, web pages, videos and the like, and the invention has important application value, and once the research is successful and put into application, huge social and economic benefits can be generated.
Compared with the prior art, the invention has the beneficial effects that:
1. the invention realizes the characteristic representation of the extracted data from the level research of the independent component analysis model, and improves the robustness of the represented characteristic compared with the traditional characteristic extraction algorithm (sparse coding or self-coding).
2. According to the invention, through a method of a hierarchical structure in deep learning, on the basis of analyzing data of a data set, a staged RicA algorithm is researched and provided, a source domain label and a multi-class regression target function are considered in a multilayer structure, the condition of the source domain label is applied to optimization of the target function, more useful characteristics in the data set can be extracted, the classification precision in the target domain is improved, and the accuracy of transfer learning is improved.
3. The invention can be applied to a plurality of fields such as images, texts, audios, videos and the like, and has important application value. Moreover, the results of the research based on the Stacked RICA can also be applied to many pattern classification fields related to transfer learning, such as image recognition, emotion classification, theme classification, voice recognition and robot system.
Drawings
Fig. 1 is a flowchart of a specific study scheme of feature extraction and transfer learning according to the present invention.
FIG. 2 is a schematic diagram of the hierarchy of the RICA model.
FIG. 3 is a schematic diagram of an analysis of a Stacked RICA model according to the present invention.
Detailed Description
As shown in fig. 1, fig. 1 is a flow chart of the method of the present invention, and the specific implementation in fig. 1 is as follows:
(1) in order to train better characteristics, the training data set and the test data set are spliced and vectorized to obtain a vectorized data set.
(2) For the vectorized text data set, a Stacked Reconstruction independent component Analysis (Stacked RICA) model is used for feature extraction, and the specific process is as follows:
1) whitening data with the ZCA method:
ZCA whitening is a data pre-processing method that maps data from x to xZCAwhiteIt has also proven to be a rough model of the processed image of a biological eye (retina). For example, when your eye perceives an image, most adjacent "pixels" are perceived as similar values in the eye, since adjacent parts in an image are very correlated in brightness. Therefore, it is very cost-inefficient if the human eye needs to transmit each pixel value separately (via the optic nerve) into the brain. Instead, the retina performs a decorrelation operation similar to that in ZCA, thereby obtaining a less redundant representation of the input image, and transmits it to the brain. In feature extraction, the input is redundant for training purposes due to the strong correlation between adjacent instances or expressions in the dataset. The purpose of whitening is to reduce the redundancy of the input, and the input of the learning algorithm has the following properties through the whitening process: (i) features ofThe inter-correlation is low; (ii) all features have the same variance. The result of ZCA whitening can be expressed as:
Figure BDA0001456490160000071
2) feature extraction based on Stacked RICA
The method comprises the following specific steps:
① Single layer RICA extraction features
A Reconstruction Independent Component Analysis (RICA) algorithm is designed to extract features according to the idea of fig. 2. Assuming that given an input of x, the present invention is intended to derive a linearly independent set of bases (denoted by W), the objective function can be expressed as:
J(W)=||Wx||1
in the expression, Wx represents the characteristic representation of input x, in RICA, in order to ensure that mutually linearly independent overcomplete bases are obtained, the invention solves the following objective function:
Figure BDA0001456490160000081
where λ is the weight attenuation coefficient, W is the weight proof, and x is the input data. To solve the objective function:
first, the first step requires solving by a method of requiring derivatives
Figure BDA0001456490160000082
I.e. to solve for
Figure BDA0001456490160000083
As shown in fig. 2, the weights and activation functions in the model are as follows:
Figure BDA0001456490160000084
let J (z)(4)) F (x), then J (z)(4))=∑kJ(zk (4))。
Model will beAfter the input of (2) is set to F, the question is converted to a solution
Figure BDA0001456490160000085
Although W appears twice in the model, it can be shown that when W appears multiple times in the neural network, the partial derivative with respect to W is the sum of partial derivatives with respect to each W instance in the network, as follows:
Figure BDA0001456490160000091
as described above, the present invention first derives a partial derivative for each W instance,
with respect to WTThe following steps:
Figure BDA0001456490160000092
regarding W:
Figure BDA0001456490160000093
the final method for solving partial derivatives of W is:
Figure BDA0001456490160000094
the second step is an iteration by the method with l-bfgs. The invention is completed by the following cost function:
Figure BDA0001456490160000095
w finally obtained after multiple iterations is a group of linear independent overcomplete bases of the original input x. From this set of bases we can get a more useful characterization Wx of the original input data x.
② superposition of RICA (stacked RICA) computational feature representation
FIG. 3 is a model diagram of a Stacked RICA model according to the present invention, illustrating that the Stacked RICA model consists of an input layer, two hidden layers, and an output layer. The Stacked RICA model isBased on the idea of deep learning, the RICA structures are superposed, namely a stronger characteristic representation z obtained after the completion of single-layer RICA is used as the input of the next-layer RICA algorithm, and then each layer of iteration optimization parameters is used for optimizing the objective function. Finally obtaining the characteristic representation of the original input data through multilayer superposition
Figure BDA0001456490160000096
(3) After the feature extraction work is done by Stacked RICA, the resulting feature representation is used
Figure BDA0001456490160000097
Instead of the original input data x, pair
Figure BDA0001456490160000098
Semi-Supervised learning (Semi-Supervised L earning) is carried out and consideration of bias terms is added, wherein the bias terms comprise K L distances of source domain distribution and target domain distribution and multi-class regression bias terms made according to source domain class labels, and label information of a source domain is applied to optimization of feature representation.
The objective function can be expressed as:
Figure BDA0001456490160000101
wherein the content of the first and second substances,
Figure BDA0001456490160000102
representing the reconstruction error from the original data to the data re-represented after feature extraction.
Figure BDA0001456490160000103
K L distance representing the source domain distribution and the target domain distribution.
Figure BDA0001456490160000104
Representing according to source domainAnd (4) making multiple types of regression bias items by using the class labels.
Figure BDA0001456490160000105
Representing the constraint terms of the characteristic parameter matrix W.
(4) After all the feature extraction and selection processes are completed, the obtained features of the source domain are used for representing, and a classifier is trained in the source domain, wherein the tool for training the classifier is a Support Vector Machine (SVM), a logistic regression analysis model (L R) or a module classifier.
(5) And (4) carrying out classification prediction in the target domain by using the classifier obtained by the source domain training, thereby applying the classifier in the source domain to the target domain.
(6) And obtaining a final migration learning result.

Claims (1)

1. A transfer learning method based on a depth sparse self-coding machine is characterized by comprising the following steps: the method sequentially comprises the following steps:
(1) and whitening preprocessing is carried out on all images in the image database, and the process is as follows:
(1.1) representing the input dataset as { x }(1),x(2),...,x(n)Calculating the covariance matrix of x
Figure FDA0002477819550000011
Then, calculating the eigenvector of the covariance matrix, and forming a matrix U according to the column as shown in the following formula:
Figure FDA0002477819550000012
in the matrix U, U1Is the principal eigenvector, which corresponds to the largest eigenvalue, u2Is a sub-feature vector, analogizes with the same, and records λ12,...,λnCharacteristic values corresponding to all the vector quantities in the matrix U are obtained;
(1.2) representing the input data by the calculated matrix U as shown in the following equation:
Figure FDA0002477819550000013
wherein the subscript rot refers to rotation, which means that it is the result of the rotation processing of the original data, in order to make each input feature have unit variance, the use of
Figure FDA0002477819550000014
Scaling each feature x as a scaling factorrot,iWherein the resulting PCA whitened data is represented by the following formula:
Figure FDA0002477819550000021
(1.3) let R be any orthogonal matrix, i.e. satisfy RRT=RTR is I, then RxPCAwhiteStill with unit covariance, in order to make the input data after whitening processing for all possible R as close to the original input data as possible, let R be U, resulting in formula (1):
xZCAwhite=UxPCAwhite(1),
xZCAwhitenamely processing data of original input data obtained after ZCA whitening;
(2) constructing a depth sparse self-coding machine model to extract high-level abstract features of the image, wherein the process is as follows:
(2.1) constructing a sparse self-encoding machine model, comprising the following steps:
(2.1.1), the sparse self-coding model uses a Reconstruction Independent Component Analysis algorithm (RICA) to obtain x by using a formula (1)ZCAwhiteIs input data of the RICA algorithm and substitutes the cost function formula (2):
Figure FDA0002477819550000022
in the cost function formula (2), W is a weighting matrix;
(2.1.2) solving for x in the cost function formula (2)ZCAwhiteWherein when the first term of the cost function (2) is subjected to partial derivation, the partial derivation of (1) is adopted
Figure FDA0002477819550000023
As the partial derivative function, the obtained partial derivative function is shown in formula (3):
Figure FDA0002477819550000024
(2.1.3) iteratively calculating a weighting matrix W by using an L-FBGS algorithm to obtain a trained sparse self-coding model;
(2.2) constructing a deep sparse self-coding machine model:
substituting the weighting matrix W obtained in the step (2.1) into the cost function formula (2), and recording the obtained output as
Figure FDA0002477819550000031
Is output data obtained after the training of the single-layer RICA model is finished, and the output data is obtained
Figure FDA0002477819550000032
Repeating step (2.1) as input data to obtain W(l)Training a weighting matrix obtained after stacking the sparse self-coding machine, wherein l is the number of times of the iteration step (2.1);
(2.3) extracting features according to the trained deep sparse self-coding machine model;
pooling the square root of the model square root with the weighting matrix W obtained in step (2.1)(l)Substituting the formula (4) for convolution feature extraction, wherein the formula (4) is as follows:
Figure FDA0002477819550000033
in the formula (4), the first and second groups,
Figure FDA0002477819550000034
represents the input of the z-th layer in the convolutional network,
Figure FDA0002477819550000035
the error term for the z +1 th layer representing the kth feature in the convolutional network, the output from equation (4) is denoted as xfeaAbstract features extracted from the raw input data;
(3) and optimizing features by semi-supervised learning:
using x obtained in step (2)feaPerforming semi-supervised learning for input to obtain a formula (5), adding K L distance of source domain distribution and target domain distribution and multi-class regression bias terms according to source domain class labels,
Figure FDA0002477819550000036
denotes the output, W, obtained after semi-supervised learningSSLRepresenting a weight matrix in semi-supervised learning, ξ(s)Output representing hidden layers in the source domain, ξ(t)Representing the output of the hidden layer in the target domain, equation (5) is as follows:
Figure FDA0002477819550000037
in the formula (5), the first and second groups,
Figure FDA0002477819550000038
representing a reconstruction error from the input data of the step (3) to data re-represented after feature extraction;
Figure FDA0002477819550000039
representing the source domain distribution and the target domain distribution and the K L distance between the target domain distribution and the source domain distribution;
Figure FDA00024778195500000310
representing multi-class regression bias terms made according to the source domain class labels;
Figure FDA0002477819550000041
represents the weight matrix WSSLThe constraint term of (2);
(4) training a classifier and classifying the test image data set, wherein the process is as follows:
(4.1) training L R classifier with test image data set, in L R classification, note:
Figure FDA0002477819550000042
in the formula (6), σ (z') is sigmoid function, and has the same function as input z
Figure FDA0002477819550000043
Outputting the finished product of the step (3)
Figure FDA0002477819550000044
Output of the training data set in (1)
Figure FDA0002477819550000045
And a label substitution (6) training classifier for y, the training image dataset, of known labels;
(4.2) classifying the test image data set by using the trained classifier; outputting the finished product of the step (3)
Figure FDA0002477819550000046
Output of the test data set in (1)
Figure FDA0002477819550000047
Obtaining a classification result T of the test image dataset, in place of the L R classifier performed by equation (6)testAs shown in equation (7):
Figure FDA0002477819550000048
CN201711069171.XA 2017-11-03 2017-11-03 Transfer learning method based on depth sparse self-coding machine Expired - Fee Related CN107798349B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711069171.XA CN107798349B (en) 2017-11-03 2017-11-03 Transfer learning method based on depth sparse self-coding machine

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711069171.XA CN107798349B (en) 2017-11-03 2017-11-03 Transfer learning method based on depth sparse self-coding machine

Publications (2)

Publication Number Publication Date
CN107798349A CN107798349A (en) 2018-03-13
CN107798349B true CN107798349B (en) 2020-07-14

Family

ID=61549046

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711069171.XA Expired - Fee Related CN107798349B (en) 2017-11-03 2017-11-03 Transfer learning method based on depth sparse self-coding machine

Country Status (1)

Country Link
CN (1) CN107798349B (en)

Families Citing this family (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108564121B (en) * 2018-04-09 2022-05-03 南京邮电大学 Unknown class image label prediction method based on self-encoder
CN108595568B (en) * 2018-04-13 2022-05-17 重庆邮电大学 Text emotion classification method based on great irrelevant multiple logistic regression
CN108805160B (en) * 2018-04-17 2020-03-24 平安科技(深圳)有限公司 Transfer learning method and device, computer equipment and storage medium
CN108764281A (en) * 2018-04-18 2018-11-06 华南理工大学 A kind of image classification method learning across task depth network based on semi-supervised step certainly
CN109117793B (en) * 2018-08-16 2021-10-29 厦门大学 Direct-push type radar high-resolution range profile identification method based on deep migration learning
CN109359557B (en) * 2018-09-25 2021-11-09 东北大学 SAR remote sensing image ship detection method based on transfer learning
CN109726742A (en) * 2018-12-11 2019-05-07 中科恒运股份有限公司 The quick training method of disaggregated model and terminal device
CN109816002B (en) * 2019-01-11 2022-09-06 广东工业大学 Single sparse self-encoder weak and small target detection method based on feature self-migration
CN109902861B (en) * 2019-01-31 2023-01-03 南京航空航天大学 Order production progress real-time prediction method based on double-layer transfer learning
CN111046824B (en) * 2019-12-19 2023-04-28 上海交通大学 Efficient denoising and high-precision reconstruction modeling method and system for time series signals
CN111753898B (en) * 2020-06-23 2023-09-22 扬州大学 Representation learning method based on superposition convolution sparse self-encoder
CN111753899B (en) * 2020-06-23 2023-10-17 扬州大学 Self-adaptive unbalanced data field adaptation method
CN111985161B (en) * 2020-08-21 2024-06-14 广东电网有限责任公司清远供电局 Reconstruction method of three-dimensional model of transformer substation
CN112070236B (en) * 2020-09-11 2022-08-16 福州大学 Sparse feature learning method for solving online complex optimization calculation based on transfer learning

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104200224A (en) * 2014-08-28 2014-12-10 西北工业大学 Valueless image removing method based on deep convolutional neural networks
CN104408469A (en) * 2014-11-28 2015-03-11 武汉大学 Firework identification method and firework identification system based on deep learning of image
CN105844331A (en) * 2015-01-15 2016-08-10 富士通株式会社 Neural network system and training method thereof
CN106096652A (en) * 2016-06-12 2016-11-09 西安电子科技大学 Based on sparse coding and the Classification of Polarimetric SAR Image method of small echo own coding device
CN106203506A (en) * 2016-07-11 2016-12-07 上海凌科智能科技有限公司 A kind of pedestrian detection method based on degree of depth learning art
CN106529428A (en) * 2016-10-31 2017-03-22 西北工业大学 Underwater target recognition method based on deep learning
CN106599863A (en) * 2016-12-21 2017-04-26 中国科学院光电技术研究所 Deep face identification method based on transfer learning technology

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104200224A (en) * 2014-08-28 2014-12-10 西北工业大学 Valueless image removing method based on deep convolutional neural networks
CN104408469A (en) * 2014-11-28 2015-03-11 武汉大学 Firework identification method and firework identification system based on deep learning of image
CN105844331A (en) * 2015-01-15 2016-08-10 富士通株式会社 Neural network system and training method thereof
CN106096652A (en) * 2016-06-12 2016-11-09 西安电子科技大学 Based on sparse coding and the Classification of Polarimetric SAR Image method of small echo own coding device
CN106203506A (en) * 2016-07-11 2016-12-07 上海凌科智能科技有限公司 A kind of pedestrian detection method based on degree of depth learning art
CN106529428A (en) * 2016-10-31 2017-03-22 西北工业大学 Underwater target recognition method based on deep learning
CN106599863A (en) * 2016-12-21 2017-04-26 中国科学院光电技术研究所 Deep face identification method based on transfer learning technology

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Text Classification Based on ReLU Activation Function of SAE Algorithm;Jiale Cui et al.;《International Symposium on Neural Network》;20170531;第44-50页 *
一种基于最大熵模型的加权归纳迁移学习方法;梅灿华 等;《计算机研究与发展》;20111231;第48卷(第9期);第1722-1728页 *
基于局部不变特征融合的图像检索技术研究;谢李鹏;《中国优秀硕士学位论文全文数据库 信息科技辑》;20170215;第I138-2970页 *

Also Published As

Publication number Publication date
CN107798349A (en) 2018-03-13

Similar Documents

Publication Publication Date Title
CN107798349B (en) Transfer learning method based on depth sparse self-coding machine
Yang et al. Application of deep convolution neural network
US11645835B2 (en) Hypercomplex deep learning methods, architectures, and apparatus for multimodal small, medium, and large-scale data representation, analysis, and applications
CN111444343B (en) Cross-border national culture text classification method based on knowledge representation
Abdullah et al. Improving face recognition by elman neural network using curvelet transform and HSI color space
CN109189925A (en) Term vector model based on mutual information and based on the file classification method of CNN
CN110110323B (en) Text emotion classification method and device and computer readable storage medium
CN111127146B (en) Information recommendation method and system based on convolutional neural network and noise reduction self-encoder
Taylor et al. Learning invariance through imitation
CN110533024B (en) Double-quadratic pooling fine-grained image classification method based on multi-scale ROI (region of interest) features
CN111538761A (en) Click rate prediction method based on attention mechanism
CN111816156A (en) Many-to-many voice conversion method and system based on speaker style feature modeling
CN104268593A (en) Multiple-sparse-representation face recognition method for solving small sample size problem
CN106919980A (en) A kind of increment type target identification system based on neuromere differentiation
CN109086886A (en) A kind of convolutional neural networks learning algorithm based on extreme learning machine
CN101021900A (en) Method for making human face posture estimation utilizing dimension reduction method
CN108121975A (en) A kind of face identification method combined initial data and generate data
CN112818764A (en) Low-resolution image facial expression recognition method based on feature reconstruction model
Seyyedsalehi et al. A fast and efficient pre-training method based on layer-by-layer maximum discrimination for deep neural networks
CN107369147B (en) Image fusion method based on self-supervision learning
CN113379655B (en) Image synthesis method for generating antagonistic network based on dynamic self-attention
Golovko et al. A new technique for restricted Boltzmann machine learning
Ovcharenko et al. Style transfer for generation of realistically textured subsurface models
CN103646256A (en) Image characteristic sparse reconstruction based image classification method
CN108229571A (en) Apple surface lesion image-recognizing method based on KPCA algorithms Yu depth belief network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20200714