CN109829472B - Semi-supervised classification method based on probability nearest neighbor - Google Patents

Semi-supervised classification method based on probability nearest neighbor Download PDF

Info

Publication number
CN109829472B
CN109829472B CN201811598286.2A CN201811598286A CN109829472B CN 109829472 B CN109829472 B CN 109829472B CN 201811598286 A CN201811598286 A CN 201811598286A CN 109829472 B CN109829472 B CN 109829472B
Authority
CN
China
Prior art keywords
matrix
data
probability
data set
neighbor
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201811598286.2A
Other languages
Chinese (zh)
Other versions
CN109829472A (en
Inventor
马君亮
汪西莉
刘琦
肖冰
唐铭英
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shaanxi Normal University
Original Assignee
Shaanxi Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shaanxi Normal University filed Critical Shaanxi Normal University
Priority to CN201811598286.2A priority Critical patent/CN109829472B/en
Publication of CN109829472A publication Critical patent/CN109829472A/en
Application granted granted Critical
Publication of CN109829472B publication Critical patent/CN109829472B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention relates to a semi-supervised classification method based on probability nearest neighbor, which comprises the steps of preparing a data set, preprocessing the data set, constructing a probability nearest neighbor matrix S, initializing a probability transition matrix P into the probability nearest neighbor matrix S, performing label propagation on an original data set, updating, checking convergence, obtaining classification results and the like. The method solves the problems that in the existing semi-supervised classification based on the graph, the construction of the similarity graph cannot well reflect actual conditions and the classification is inaccurate, and the classification method disclosed by the invention can be closer to the actual conditions, so that the classification result is more accurate.

Description

Semi-supervised classification method based on probability nearest neighbor
Technical Field
The present disclosure relates to data classification methods, and in particular, to a semi-supervised classification method based on probability neighbors.
Background
The existing data classification methods comprise methods of supervised classification, semi-supervised classification, unsupervised classification and the like. The supervision and classification method requires a large amount of marked data to train the model, so that the application scene is limited; the unsupervised classification does not need the class information of the data, and has wide application, but the classification effect is poor due to the lack of the class information. The semi-supervision only needs a small amount of marked data, so that the acquisition cost is low, and a better classification effect can be obtained by learning the data distribution of a large amount of unmarked data, so that the semi-supervision has a wide application scene.
The semi-supervised classification based on the graph is an important branch in the semi-supervised classification, and because the relation between data is fully utilized, a good effect is often obtained, and the great attention is paid. In the methods, the relation between the data is represented in the similarity graph, and the graph structure and the matrix have a one-to-one correspondence relation, so that the similarity graph can be represented by the similarity matrix, and meanwhile, the matrix can be further converted into a probability transition matrix for tag propagation between the data, and further, the tag propagation is carried out by using the category information of the marked data, so that a classification result is obtained.
However, in the current semi-supervised classification method based on the graph, the similarity graph is often constructed by a K-nearest neighbor (KNN) or epsilon-nearest neighbor method, in the process of constructing the graph, only the attribute features of the data are used, the class information of the marked data is not used, the obtained similarity graph cannot well reflect the actual situation, and the classification result is also inaccurate.
Disclosure of Invention
Aiming at the problems, the present disclosure provides a semi-supervised classification method based on probability nearest neighbor, which solves the problems that in the existing semi-supervised classification based on graph, the construction of a similarity graph cannot well reflect actual conditions and classification is inaccurate, and the classification method of the present disclosure can be closer to the actual conditions, so that classification results are more accurate.
Specifically, the disclosure provides a semi-supervised classification method based on probability nearest neighbor, which comprises the following steps:
s100, preparing an original data set, wherein the original data set comprises original data and matched labels thereof, the characteristics of the original data are described through data attributes, and the original data comprises marked data and unmarked data;
S200, preprocessing the data set prepared in the step S100 to obtain a label indication vector matrix V L of marked data;
s300, constructing a probability neighbor matrix S of an original data set;
s400, constructing a category information matrix F and a probability transition matrix P of the original data set, and initializing the probability transition matrix P into a probability neighbor matrix S obtained in the step S300;
s500, based on the initialized probability transition matrix P in the step S400, performing label propagation on the original data set to obtain a new category information matrix F';
S600, updating a marked data matrix F L of a new category information matrix F' after label propagation by using a label indication vector matrix V L of marked data, so as to prevent pollution of label information;
s700, checking whether the new category information matrix F 'obtained in the step S500 is converged, if the new category information matrix F' is converged and is not changed, entering the step S800, otherwise, returning to the step S500;
s800, obtaining a classification result of the original data set through the new category information matrix F' after convergence.
Compared with the prior art, the method has the following beneficial technical effects:
(1) The classification method disclosed by the invention has the advantages that the similarity between two data points is calculated by adding the class information of the data when the data is formed, the actual situation of the data can be well reflected by the constructed similarity graph, and the accuracy of data classification is improved compared with the existing semi-supervised classification method;
(2) The classification method defines a probability neighbor matrix, and the composition problem is regarded as a probability problem, so that the classification result is more accurate.
Drawings
FIG. 1 illustrates a flow chart of a probability neighbor based semi-supervised classification method of the present disclosure;
FIG. 2 is a schematic representation of the values of a probability transition matrix constructed based on the prior KNN method;
FIG. 3 illustrates a schematic representation of the values of a probability transition matrix constructed by the method of the present disclosure;
FIG. 4 is a graph showing accuracy contrast of classifying a plurality of different data sets using the prior KNN method and the classification method of the present disclosure.
Detailed Description
The following describes a specific flow of the probability neighbor based semi-supervised classification method of the present disclosure with reference to fig. 1.
In one embodiment, a semi-supervised classification method based on probabilistic neighbor is provided, comprising:
s100, preparing an original data set, wherein the original data set comprises original data and matched labels thereof, the characteristics of the original data are described through data attributes, and the original data comprises marked data and unmarked data;
S200, preprocessing the data set prepared in the step S100 to obtain a label indication vector matrix V L of marked data;
s300, constructing a probability neighbor matrix S of an original data set;
s400, constructing a category information matrix F and a probability transition matrix P of the original data set, and initializing the probability transition matrix P into a probability neighbor matrix S obtained in the step S300;
s500, based on the initialized probability transition matrix P in the step S400, performing label propagation on the original data set to obtain a new category information matrix F';
S600, updating a marked data matrix F L of a new category information matrix F' after label propagation by using a label indication vector matrix V L of marked data, so as to prevent pollution of label information;
s700, checking whether the new category information matrix F 'obtained in the step S500 is converged, if the new category information matrix F' is converged and is not changed, entering the step S800, otherwise, returning to the step S500;
s800, obtaining a classification result of the original data set through the new category information matrix F' after convergence.
In this embodiment, the tag information pollution in step S600 means that the tag of the marked data may be changed after the processing in step S500, and the tag of the marked data may be wrong, which is called tag information pollution.
In this embodiment, the implementation steps of the classification method proposed in the present disclosure are described in detail, including the steps of preparing a dataset, preprocessing the dataset, constructing a probability neighbor matrix S, initializing a probability transition matrix P to the probability neighbor matrix S, performing label propagation on the original dataset, updating, checking convergence, obtaining classification results, and the like. The method disclosed by the invention can be closer to the actual condition of data classification, so that the classification result is more accurate.
In a preferred embodiment, the step S200 specifically includes:
S201, defining an original dataset as X epsilon R n*d, wherein R n*d represents a matrix of n rows and d columns, n is the number of data points, and d is the attribute number of the data points; defining the marked data matrix as Unlabeled data matrix asWhere n 1 is the number of labeled data points, n 2 is the number of unlabeled data points, n=n 1+n2;
s202, constructing a label indication vector matrix of the marked data according to the label corresponding to the marked data:
Constructing a tag indication vector matrix V epsilon R n*c according to the tag corresponding to the marked data of the original data set, wherein n is the number of data, c is the number of data types, the ith row of the tag indication vector matrix V is the tag indication vector of the ith data, if the type of the ith data point is j, the jth element of the row is 1, and the rest data of the row is 0; partial data points are decimated from the original data set according to a certain proportion, and corresponding rows in the label indication vector matrix V are formed into a matrix according to the label sequence of the decimated data points The labels as marked data indicate a vector matrix.
In this embodiment, a specific description is given of how the data set is preprocessed in step S200 to obtain the tag indication vector matrix V L of the marked data. The selection ratio in step S202 may be selected from the original data and 5% -10% of the data as marked data to construct V L according to the actual situation, and the rest is the unmarked data. In experiments following the present disclosure, experiments were performed with a selection of 10%.
In a preferred embodiment, the step S300 specifically includes:
s301, defining an augmentation matrix A epsilon R n*(d+c) about data, wherein d is the dimension of the data, the number of data attributes is represented, c is the number of data categories, n is the number of data points, and A= [ X, V ];
s302, defining D epsilon R n*n as a similarity matrix between data points, and calculating the similarity between the data points i and j by taking Euclidean distance as a measure Let/>As a specific numerical value of the ith row and the jth column in the matrix D, arranging the matrix D in ascending order, X i represents the ith row of the original data set X, and X j represents the jth row of the original data set X;
s303, defining a probability neighbor matrix of an original data set as S epsilon R n*n, wherein S is a matrix of n x n, a value S i,j of an ith row and a jth column in the S matrix represents possibility that a data point i and a data point j become probability neighbors, defining K as a neighbor number, selecting the first K of a matrix D according to rows according to a K-neighbor (KNN) method to obtain K-neighbors, and constructing the probability neighbors on the basis of the K-neighbors, wherein the probability neighbors are as follows: defining the probability that p (i, k) becomes a probability neighbor between data point i and its kth neighbor, by the formula Obtaining;
Then through the formula Performing assignment, wherein S (i, k) represents the value of a kth neighbor of an ith data point in an S matrix, and an unassigned element in the S matrix is assigned to 0 to obtain an assigned probability neighbor matrix S; wherein the method comprises the steps ofRepresenting the specific value of column (k+1) of row j in matrix D,/>Is a specific value of the j-th column of the i-th row in matrix D,Is the specific numerical value of the ith row and 1 st column in the matrix D, n is the number of data points, and X represents the original data set.
In this embodiment, a specific method of how to construct the probability neighbor matrix S of the original data set in step S300 is described, where the similarity between two data points is calculated by adding the class information of the data during composition, and a probability neighbor matrix is defined, and the composition problem is regarded as a probability problem to be solved, so that the classification method is closer to the actual situation, and the classification result is more accurate.
In this embodiment, a good effect can be obtained by the local characteristics of the data, and the value of the k neighbor number in step S303 can be selected to be 5-20.
In a preferred embodiment, the step S400 specifically includes:
constructing a category information matrix F of the original data set: f (0)=[FL,0],F(0) represents the initialized F matrix:
constructing a probability transition matrix P of the original data set: initializing the probability transition matrix P as the assigned probability neighbor matrix S obtained in the step S300, which includes: p (0) = S; where P ε R n*n is the probability transition matrix, P (0) represents the initialized P matrix, and P i,j represents the value of the ith row and jth column in the P matrix, indicating that the ith data point will propagate with probability P i,j to the jth data point.
In a preferred embodiment, the step S500 specifically includes:
Tag propagation is performed by the formula F ' (t+1)=P*F'(t), where the superscript t denotes the number of iterations, F (t) denotes the F ' matrix obtained by the t-th iteration, F ' (t+1) denotes the F ' matrix obtained by the t+1st iteration, and the formula denotes the tag propagation from the t-th to the t+1st, where when t=0, F ' (t) denotes the original category information matrix F.
In this embodiment, a specific process of tag propagation is described, and a plurality of iterations are performed in combination with a probability transition matrix on the basis of an original category information matrix F, so as to perform tag propagation on unlabeled data according to tags of the labeled data, and classify the unlabeled data.
In a preferred embodiment, the step S600 specifically includes:
F L=VL, reassigning the labeled data F L to a labeled data matrix V L,FL representing the category information matrix F after each iteration of label propagation, and V L representing the label indication vector matrix of the labeled data.
In this embodiment, in order to prevent the tag of the marked data from being contaminated after the process of step S500, reassignment is required.
In a preferred embodiment, in the step S800:
And each row of data of the new category information matrix F 'after convergence represents the category of the data, if the value of F ij is 1, the category of the ith data point is j, so that the category corresponding to each data can be obtained from the F' matrix, and the classification result of the original dataset is obtained.
The method is used for classifying the data according to the method, and can be closer to the actual situation, so that the classification result is more accurate.
Experiment:
In order to verify the advantages of the semi-supervised classification method based on probability neighbor provided by the disclosure compared with the existing classification method based on KNN, experimental comparison and verification are carried out.
In the first experiment, a double-month data set is adopted, the double-month data set is of a known data set type, the number of data points can be selected according to actual needs, and the number of the double-month data points selected in the experiment is 400. 10% of the dataset was selected in this experiment as marked data construct V L, the remainder as unmarked data.
Fig. 2 shows the values of the probability transition matrix constructed based on the existing KNN method, and fig. 3 shows the values of the probability transition matrix constructed by the method of the present invention. Fig. 2 and fig. 3 are values of probability transition matrixes generated on the same data set by two methods, gray scales at points in the graph represent probability values of neighbors among different nodes, and as can be seen from results, the probability transition matrixes generated by the two methods have small proportion of non-zero elements, so that the corresponding graph structures are relatively sparse. The difference is that: in fig. 2, the gray levels corresponding to the non-zero elements are the same, so that in the KNN method, the neighboring points have the same tag propagation probability, which is not consistent with the actual situation; in fig. 3, the gray scales corresponding to the non-zero elements are different, and the label propagation probabilities of the adjacent points are different, so that the method is more in line with the actual situation, and the interrelationship between the data is expressed.
In the second experiment, for the existing multiple different data sets (as shown in the abscissa of fig. 4), the data classification is performed by adopting a method of knn+ tag propagation (knn+lp), and the data classification is performed by adopting a method of pnn+ tag propagation (pnn+lp) in the method of the present invention, and the accuracy of classification results of the two methods on the different data sets is shown by comparing with fig. 4, it can be found that the classification result obtained by adopting the classification method (pnn+lp) of the present disclosure is superior to the prior art, and the classification accuracy is higher.
In summary, the method solves the problems that in the existing semi-supervised classification based on the graph, the construction of the similarity graph cannot well reflect actual conditions and classification is inaccurate, and compared with the prior art, the classification method disclosed by the invention can be closer to the actual conditions, so that the classification result is more accurate.
Although the embodiments of the present invention have been described above with reference to the accompanying drawings, the present invention is not limited to the above-described specific embodiments and application fields, and the above-described specific embodiments are merely illustrative, and not restrictive. Those skilled in the art, having the benefit of this disclosure, may effect numerous forms of the invention without departing from the scope of the invention as claimed.

Claims (8)

1. A semi-supervised classification method based on probability neighbor performed on a computer, the method aiming at the problem that the classification is inaccurate due to the construction of a similarity graph in the existing semi-supervised classification based on the graph, the method is characterized by comprising the following steps:
s100, preparing an original data set, wherein the original data set comprises original data and matched labels thereof, the characteristics of the original data are described through data attributes, and the original data comprises marked data and unmarked data;
S200, preprocessing the data set prepared in the step S100 to obtain a label indication vector matrix V L of marked data;
s300, constructing a probability neighbor matrix S of an original data set;
s400, constructing a category information matrix F and a probability transition matrix P of the original data set, and initializing the probability transition matrix P into a probability neighbor matrix S obtained in the step S300;
s500, based on the initialized probability transition matrix P in the step S400, performing label propagation on the original data set to obtain a new category information matrix F';
S600, updating a marked data matrix F L of a new category information matrix F' after label propagation by using a label indication vector matrix V L of marked data, so as to prevent pollution of label information;
s700, checking whether the new category information matrix F 'obtained in the step S500 is converged, if the new category information matrix F' is converged and is not changed, entering the step S800, otherwise, returning to the step S500;
s800, obtaining a classification result of the original data set through the new category information matrix F' after convergence;
the data set in the step S100 is one of USPS, COIL1 and COIL2 data sets;
The method adds category information of the used data to calculate the similarity between two data points when the composition is carried out;
the method defines a probability neighbor matrix, and the composition problem is regarded as a probability problem, so that the classification result of machine learning is more accurate.
2. The method according to claim 1, wherein the step S200 specifically includes:
s201, defining the original data set as Wherein/>Representing a matrix of n rows and d columns, n being the number of data points and d being the number of attributes of the data points; defining the marked data matrix as/>Unlabeled data matrix is/>Where n 1 is the number of labeled data points, n 2 is the number of unlabeled data points, n=n 1+n2;
s202, constructing a label indication vector matrix of the marked data according to the label corresponding to the marked data:
constructing a label indication vector matrix according to labels corresponding to marked data of an original data set Where n is the number of data, c is the number of data categories, the label indicates that the ith row of vector matrix V is a label indicating vector for the ith data point, if the category of the ith data point is j, then the jth element of the row is 1, and the rest of the data of the row is 0; according to the sequence of marked data in the original data, the corresponding rows in the decimated label indication vector V form a matrix/>The labels as marked data indicate a vector matrix.
3. The method according to claim 2, wherein the step S300 specifically includes:
s301, defining an augmentation matrix for data D is the data dimension, represents the number of data attributes, c represents the number of data categories, n is the number of data points,/>
S302, definitionAs a similarity matrix between data points, the Euclidean distance is used as a measure to calculate the similarity/>, between the data points i and jLet/>As a specific numerical value of the ith row and the jth column in the matrix D, the matrix D is arranged in ascending order, X i represents the ith row of the original data set X, and X j represents the jth row of the original data set X;
S303, defining probability neighbor matrixes of the original data set as S is a matrix of n x n, the value of the j-th column of the i-th row in the S matrix/>Representing the possibility that data point i and data point j become probability neighbors, defining K as a neighbor number, selecting the first K of the matrix D according to a row by a K-neighbor (KNN) method to obtain K-neighbors, and constructing the probability neighbors on the basis of the K-neighbors, wherein the probability neighbors are specifically as follows: definition/>Probability of being a probability neighbor between data point i and its kth neighbor is calculated by the formula/>Obtaining;
Then through the formula Performing assignment, wherein S (i, k) represents the value of a kth neighbor of an ith data point in an S matrix, and an unassigned element in the S matrix is assigned to 0 to obtain an assigned probability neighbor matrix S; wherein the method comprises the steps ofRepresenting the specific value of the ith row and the (k+1) th column in matrix D,/>Is a specific value of the ith row and jth column in matrix D,/>Is the specific numerical value of the ith row and 1 st column in the matrix D, n is the number of data points, and X represents the original data set.
4. A method according to claim 3, wherein said step S400 specifically comprises:
Constructing a category information matrix F of the original data set: f (0) represents the initialized F matrix:
Constructing a probability transition matrix P of the original data set: initializing the probability transition matrix P as the assigned probability neighbor matrix S obtained in the step S300, which includes: p (0) = S; wherein, Is a probability transition matrix,/>Representing the initialized P matrix, P i,j represents the value of the ith row and jth column in the P matrix, indicating that the ith data point will propagate with probability P i,j to the jth data point.
5. The method according to claim 4, wherein the step S500 specifically includes:
By the formula Label propagation is performed, wherein the superscript t denotes the iteration number, F ' (t) denotes the F ' matrix obtained by the t-th iteration process, F ' (t+1) denotes the F ' matrix obtained by the t+1st iteration process, and the above formula denotes the label propagation process from the t-th to the t+1st time, wherein when t=0, F ' (t) denotes the original category information matrix F.
6. The method according to claim 5, wherein the step S600 specifically includes:
f L=VL, reassigning the labeled data F L to a labeled data matrix V L,FL representing the category information matrix F after each iteration of label propagation, and V L representing the label indication vector matrix of the labeled data.
7. The method according to claim 6, wherein in the step S800:
The new category information matrix F' after convergence each row of data represents the category of the data if And if the value is 1, the class of the ith data point is j, so that the class corresponding to each data can be obtained from the F' matrix, and the classification result of the original data set is obtained.
8. In the method according to claim 3, in the step S303, the value of the k neighbor number is selected to be 5-20.
CN201811598286.2A 2018-12-24 2018-12-24 Semi-supervised classification method based on probability nearest neighbor Active CN109829472B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811598286.2A CN109829472B (en) 2018-12-24 2018-12-24 Semi-supervised classification method based on probability nearest neighbor

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811598286.2A CN109829472B (en) 2018-12-24 2018-12-24 Semi-supervised classification method based on probability nearest neighbor

Publications (2)

Publication Number Publication Date
CN109829472A CN109829472A (en) 2019-05-31
CN109829472B true CN109829472B (en) 2024-05-14

Family

ID=66861165

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811598286.2A Active CN109829472B (en) 2018-12-24 2018-12-24 Semi-supervised classification method based on probability nearest neighbor

Country Status (1)

Country Link
CN (1) CN109829472B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111046914B (en) * 2019-11-20 2023-10-27 陕西师范大学 Semi-supervised classification method based on dynamic composition
CN111488923B (en) * 2020-04-03 2023-02-07 陕西师范大学 Enhanced anchor point image semi-supervised classification method

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104463202A (en) * 2014-11-28 2015-03-25 苏州大学 Multi-class image semi-supervised classifying method and system
CN104794489A (en) * 2015-04-23 2015-07-22 苏州大学 Deep label prediction based inducing type image classification method and system
CN105608471A (en) * 2015-12-28 2016-05-25 苏州大学 Robust transductive label estimation and data classification method and system
CN105608690A (en) * 2015-12-05 2016-05-25 陕西师范大学 Graph theory and semi supervised learning combination-based image segmentation method
CN105931253A (en) * 2016-05-16 2016-09-07 陕西师范大学 Image segmentation method combined with semi-supervised learning
CN106596900A (en) * 2016-12-13 2017-04-26 贵州电网有限责任公司电力科学研究院 Transformer fault diagnosis method based on improved semi-supervised classification of graph
CN108009571A (en) * 2017-11-16 2018-05-08 苏州大学 A kind of semi-supervised data classification method of new direct-push and system

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7756800B2 (en) * 2006-12-14 2010-07-13 Xerox Corporation Method for transforming data elements within a classification system based in part on input from a human annotator/expert

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104463202A (en) * 2014-11-28 2015-03-25 苏州大学 Multi-class image semi-supervised classifying method and system
CN104794489A (en) * 2015-04-23 2015-07-22 苏州大学 Deep label prediction based inducing type image classification method and system
CN105608690A (en) * 2015-12-05 2016-05-25 陕西师范大学 Graph theory and semi supervised learning combination-based image segmentation method
CN105608471A (en) * 2015-12-28 2016-05-25 苏州大学 Robust transductive label estimation and data classification method and system
CN105931253A (en) * 2016-05-16 2016-09-07 陕西师范大学 Image segmentation method combined with semi-supervised learning
CN106596900A (en) * 2016-12-13 2017-04-26 贵州电网有限责任公司电力科学研究院 Transformer fault diagnosis method based on improved semi-supervised classification of graph
CN108009571A (en) * 2017-11-16 2018-05-08 苏州大学 A kind of semi-supervised data classification method of new direct-push and system

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
一种基于线性邻域传播的加权K近邻算法;王小攀;马丽;刘福江;;计算机工程;20130715(第07期);全文 *
基于概率超图半监督学习的专利文本分类方法研究;刘桂锋;汪满容;刘海军;情报杂志;20161231(第009期);全文 *
基于近邻密度和半监督KNN的集成自训练方法;黎隽男;吕佳;;计算机工程与应用;20171123(第20期);全文 *
自适应近邻的极小极大标签传播;田勋;汪西莉;小型微型计算机***;20171231(第011期);全文 *

Also Published As

Publication number Publication date
CN109829472A (en) 2019-05-31

Similar Documents

Publication Publication Date Title
CN111191732B (en) Target detection method based on full-automatic learning
CN104298778B (en) A kind of Forecasting Methodology and system of the steel rolling product quality based on correlation rule tree
CN112232413B (en) High-dimensional data feature selection method based on graph neural network and spectral clustering
CN105701509B (en) A kind of image classification method based on across classification migration Active Learning
CN114332649A (en) Cross-scene remote sensing image depth countermeasure transfer learning method based on dual-channel attention mechanism
Kumar et al. Modeling using K-means clustering algorithm
CN109829472B (en) Semi-supervised classification method based on probability nearest neighbor
CN111325264A (en) Multi-label data classification method based on entropy
CN112685504A (en) Production process-oriented distributed migration chart learning method
CN116383422B (en) Non-supervision cross-modal hash retrieval method based on anchor points
CN103455612A (en) Method for detecting non-overlapping network communities and overlapping network communities based on two-stage strategy
CN114373093A (en) Fine-grained image classification method based on direct-push type semi-supervised deep learning
CN112668633B (en) Adaptive graph migration learning method based on fine granularity field
CN103559510B (en) Method for recognizing social group behaviors through related topic model
CN114202671A (en) Image prediction optimization processing method and device
CN116894180B (en) Product manufacturing quality prediction method based on different composition attention network
CN111259176B (en) Cross-modal Hash retrieval method based on matrix decomposition and integrated with supervision information
CN111353525A (en) Modeling and missing value filling method for unbalanced incomplete data set
CN116226404A (en) Knowledge graph construction method and knowledge graph system for intestinal-brain axis
CN115758259A (en) Multi-source domain generalization equipment fault diagnosis method based on feature fusion
CN115995019A (en) Image classification method based on instance-dependent complementary tag learning
Essa et al. A comparison between the hierarchical clustering methods for postgraduate students in Iraqi universities for the year 2019-2020 using the cophenetic and delta correlation coefficients
Russell et al. upclass: An R Package for Updating Model-Based Classification Rules
Saha et al. Unsupervised and supervised learning approaches together for microarray analysis
CN113076976B (en) Small sample image classification method based on local feature relation exploration

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant