CN112101217A - Pedestrian re-identification method based on semi-supervised learning - Google Patents

Pedestrian re-identification method based on semi-supervised learning Download PDF

Info

Publication number
CN112101217A
CN112101217A CN202010970306.5A CN202010970306A CN112101217A CN 112101217 A CN112101217 A CN 112101217A CN 202010970306 A CN202010970306 A CN 202010970306A CN 112101217 A CN112101217 A CN 112101217A
Authority
CN
China
Prior art keywords
samples
sample
pedestrian
new
semi
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010970306.5A
Other languages
Chinese (zh)
Other versions
CN112101217B (en
Inventor
葛永新
高志顺
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhenjiang Qidi Digital World Technology Co ltd
Original Assignee
Zhenjiang Qidi Digital World Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhenjiang Qidi Digital World Technology Co ltd filed Critical Zhenjiang Qidi Digital World Technology Co ltd
Priority to CN202010970306.5A priority Critical patent/CN112101217B/en
Publication of CN112101217A publication Critical patent/CN112101217A/en
Application granted granted Critical
Publication of CN112101217B publication Critical patent/CN112101217B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/23Recognition of whole body movements, e.g. for sport training
    • G06V40/25Recognition of walking or running movements, e.g. gait recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • G06F18/2132Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on discrimination criteria, e.g. discriminant analysis
    • G06F18/21322Rendering the within-class scatter matrix non-singular
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • G06F18/2132Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on discrimination criteria, e.g. discriminant analysis
    • G06F18/21322Rendering the within-class scatter matrix non-singular
    • G06F18/21328Rendering the within-class scatter matrix non-singular involving subspace restrictions, e.g. nullspace techniques

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Human Computer Interaction (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Multimedia (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Mathematical Physics (AREA)
  • Computational Linguistics (AREA)
  • Computing Systems (AREA)
  • Molecular Biology (AREA)
  • Psychiatry (AREA)
  • Software Systems (AREA)
  • Social Psychology (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a pedestrian re-identification method based on semi-supervised learning, which comprises the following steps of S100 learning a projection matrix U belonging to Rd×cProjecting the original d-dimensional feature space to a c-dimensional subspace, so that UTX∈Rc×NSatisfying in the new subspace: the Euclidean distance between sample pairs from the same pedestrian is smaller, and the Euclidean distance between sample pairs from different pedestrians is larger; samples from the same pedestrian are defined as similar samples, and samples from different pedestrians are defined as different samples; s200, adopting a projection matrix U epsilon R for a new sampled×cAnd projecting to the new subspace to obtain a prediction sample sequence, wherein the prediction sample sequence is arranged from small to large according to Euclidean distances between the new samples and samples in the training sample set. The method makes full use of the labeled sample which is accurately labeledAnd by using the contrast loss function, the negative sample pair can be fully utilized while the positive sample is constrained, the identification speed is high, and the identification accuracy is higher.

Description

Pedestrian re-identification method based on semi-supervised learning
Technical Field
The invention relates to the technical field of pedestrian re-identification, in particular to a pedestrian re-identification method based on semi-supervised learning.
Background
Pedestrian re-identification (Person re-identification), also known as pedestrian re-identification, is a technique that uses computer vision techniques to determine whether a particular pedestrian is present in an image or video sequence. Widely considered as a sub-problem for image retrieval, given a monitored pedestrian image, the pedestrian image is retrieved across devices. The visual limitation of a fixed camera is overcome, the pedestrian detection and pedestrian tracking technology can be combined, and the method can be widely applied to the fields of intelligent video monitoring, intelligent security and the like.
Although computer vision practitioners have recently proposed a number of algorithms from different perspectives for the task of pedestrian re-identification, attempting to increase the recognition rate on public data sets, pedestrian re-identification remains a very challenging task due to the effects of several real-world factors.
At present, a semi-supervised learning method is generally used for solving the pedestrian re-identification task, and the flow is approximately as follows: firstly, automatically labeling a sample without a label; secondly, the labeled samples and the automatically labeled samples are subjected to unified training, so that the model is optimized, and the model has better discrimination capability. The utilization of the unlabeled sample after automatic labeling and after labeling has two problems:
the idea of the method used for automatically labeling the unlabeled samples is to label the new mapped space by using a K-Nearest Neighbor (KNN) algorithm. This makes the error after automatic labeling larger if the new space discrimination ability learned is not enough. When the data with larger labeling errors are used for model training, the model is not likely to have better generalization capability due to the increase of training samples, and the discrimination capability of the model is poorer when the model is trained.
Secondly, only the positive sample pairs in the training set are restricted during training, and the negative sample pairs are not concerned, so that the training samples are not fully utilized.
Disclosure of Invention
Aiming at the problems in the prior art, the technical problems to be solved by the invention are as follows: the existing method for solving the pedestrian re-recognition task by using semi-supervised learning has the problems that automatic labeling errors are easily influenced by a new space obtained by learning and the utilization of training samples is insufficient.
In order to solve the technical problems, the invention adopts the following technical scheme: the pedestrian re-identification method based on semi-supervised learning comprises the following steps:
s100: learning a projection matrix U ∈ Rd×cProjecting the original d-dimensional feature space to a c-dimensional subspace, so that UTX∈Rc×NSatisfying in the new subspace: the Euclidean distance between sample pairs from the same pedestrian is smaller, and the Euclidean distance between sample pairs from different pedestrians is larger; samples from the same pedestrian are defined as similar samples, and samples from different pedestrians are defined as different samples;
s200: adopting a projection matrix U epsilon R for a new sampled×cAnd projecting to the new subspace to obtain a prediction sample sequence, wherein the prediction sample sequence is arranged from small to large according to Euclidean distances between the new samples and samples in the training sample set.
Preferably, the learning projection matrix U epsilon R in S100d×cThe method specifically comprises the following steps:
s110, establishing a training sample set, wherein the training sample set comprises a plurality of samples, the plurality of samples comprise labeled samples and unlabeled samples, and the labels of the samples of the same pedestrian in the labeled samples are the same;
let X ═ XL,XU]∈Rd×NRepresents all training samples, where N is the number of all pictures included in the training set, d is the length of the feature vector,
Figure BDA0002683805610000021
represents NLThe sample of the individual label is provided with a label,
Figure BDA0002683805610000022
represents NUEach unlabeled sample;
s120, establishing an objective function as follows:
Figure BDA0002683805610000023
wherein L (U) is a regression function,
Figure BDA0002683805610000024
for a weighted regression function, Ω (U) is a regularized constraint, α, λ > 0 are equilibrium coefficients;
s130, taking the labeled sample loss function as a contrast loss function: for sampled NPA sample pair
Figure BDA0002683805610000025
And
Figure BDA0002683805610000026
if it is
Figure BDA0002683805610000027
And
Figure BDA0002683805610000028
samples from the same pedestrian are then in the new projection space
Figure BDA0002683805610000029
And
Figure BDA00026838056100000210
has an Euclidean distance d betweennShould be as small as possible, close to 0; otherwise, then dnShould be at least greater than a predetermined threshold margin>0, if the above condition is not satisfied, a loss occurs;
s140, labeling the label-free samples by adopting a K mutual nearest neighbor method, wherein the loss function of the label-free samples is as follows:
Figure BDA00026838056100000211
wherein if UTxiAnd UTxjSatisfy K mutual nearest neighbor and xiAnd xjFrom different cameras, then
Figure BDA00026838056100000212
Otherwise Wij=0; (8);
After labeling the label-free samples, further constraining the existing subspace by using the labeled samples, wherein the weight of the constraint is the cosine distance of the two samples in the new projection space;
s150: the regularization term: the projection matrix U is constrained using L2,1 norm:
Ω(U)=||U||2,1 (4)。
preferably, the labeled sample loss function of S130 is:
Figure BDA0002683805610000031
wherein:
Figure BDA0002683805610000032
Figure BDA0002683805610000033
preferably, N sampled in S130PThe sampling strategy of each sample is a sampling strategy for maximizing the top-k recognition rate, namely for each image, all samples of k nearest neighbors are sampled.
Preferably, in S140, the method for labeling the unlabeled sample by using the method of K nearest neighbors includes:
the K nearest neighbors N (x, K) of sample x are defined as follows:
N(x,k)={x1,x2,...,xk},|N(p,k)|=k (5);
where | represents the number of samples in the set, then K is defined as the mutual nearest R (x, K) as follows:
R(x,k)={xi|(xi∈N(x,k))∧(x∈N(xi,k))} (6)。
compared with the prior art, the invention has at least the following advantages:
(1) according to the invention, the K mutual nearest neighbor is used, so that the automatic labeling result of the unlabeled sample is more credible.
(2) And fully utilizing the precisely marked labeled sample. By using the common contrast loss function in the training deep neural network, the negative sample pair can be fully utilized while the positive sample is restrained. It should be noted that any loss for identification or classification may be used as an alternative to the labeled sample loss function.
(3) In order to enable the model to be conveniently migrated to the depth model later, an end-to-end training mode is used, and a random gradient descent method is used in a training strategy. The batch generation strategy for maximizing the top-k recognition rate is provided, and the problems that the convergence speed of a pair training strategy is low in random batches, overfitting of a model is prevented and the like are solved.
Drawings
Fig. 1 shows the K mutual nearest neighbor sampling strategy for the pedestrian re-identification problem. First row: one picture to be retrieved and its 10 nearest neighbors, where P1-P4 are positive samples and N1-N6 are negative samples. A second row: each two columns are the 10 nearest neighbor images to which the first row of images corresponds. The thick-line rectangular frame without chamfers and the thin-line rectangular frame with chamfers respectively represent the retrieved picture and the positive sample picture.
The negative sample closest to the image to be retrieved in FIG. 2 is the most difficult negative sample; the first positive sample just smaller than the hardest negative sample is a moderate positive sample; the samples in the boxes are the sampling strategy herein.
Fig. 3 proper positive sample sampling.
Detailed Description
The present invention will be described in further detail with reference to the accompanying drawings.
Referring to fig. 1-3, the pedestrian re-identification method based on semi-supervised learning includes the following steps:
S100: learning a projection matrix U ∈ Rd×cProjecting the original d-dimensional feature space to a c-dimensional subspace, so that UTX∈Rc×NSatisfying in the new subspace: the Euclidean distance between sample pairs from the same pedestrian is smaller, and the Euclidean distance between sample pairs from different pedestrians is larger; samples from the same pedestrian are defined as homogeneous samples, and samples from different pedestrians are defined as heterogeneous samples.
Learning projection matrix U ∈ Rd×cThe method specifically comprises the following steps:
s110, establishing a training sample set, wherein the training sample set comprises a plurality of samples, the plurality of samples comprise labeled samples and unlabeled samples, and the labels of the samples of the same pedestrian in the labeled samples are the same;
let X ═ XL,XU]∈Rd×NRepresents all training samples, where N is the number of all pictures included in the training set, d is the length of the feature vector,
Figure BDA0002683805610000041
represents NLThe sample of the individual label is provided with a label,
Figure BDA0002683805610000042
represents NUEach unlabeled sample;
s120, establishing an objective function as follows:
Figure BDA0002683805610000043
where L (U) is a regression function, with the goal of making labeled exemplars meet the same labeled exemplar pair closer together, the uncorrelated labeled exemplar pair farther apart in the new space mapped to,
Figure BDA0002683805610000044
the regression function is a weighted regression function which can improve the discrimination of the model by using unlabelled samples, and omega (U) is a regularization constraint which can select more discriminative features from an original feature space to avoid the condition that the more discriminative features are selectedFitting, wherein alpha and lambda are balance coefficients;
s130, a labeled sample loss function: the purpose of this constraint is to fully utilize the label information of the labeled exemplars. To simultaneously use the positive and negative sample pair constraints, we use the training contrast loss function
Figure BDA0002683805610000051
Wherein:
Figure BDA0002683805610000052
Figure BDA0002683805610000053
for sampled NPA sample pair
Figure BDA0002683805610000054
And
Figure BDA0002683805610000055
if it is
Figure BDA0002683805610000056
And
Figure BDA0002683805610000057
samples from the same pedestrian are then in the new projection space
Figure BDA0002683805610000058
And
Figure BDA0002683805610000059
has an Euclidean distance d betweennShould be as small as possible, close to 0; otherwise, then dnShould be at least greater than a predetermined threshold margin>0, if the above condition is not satisfied, a loss occurs;
s140, labeling the label-free samples by adopting a K mutual nearest neighbor method, wherein the loss function of the label-free samples is as follows:
in order to effectively utilize the discrimination information of the unlabeled samples and reduce the adverse effect of wrong labeling on the model, K mutual nearest neighbor is adopted to replace K nearest neighbor to label the unlabeled samples, and only the positive sample pair is constrained in the item. The specific loss function is as follows:
Figure BDA00026838056100000510
wherein if UTxiAnd UTxjSatisfy K mutual nearest neighbor and xiAnd xjFrom different cameras, then
Figure BDA00026838056100000511
Otherwise Wij=0 (8);
The significance of this term is to identify that the K nearest neighbor sample pairs in the learned discriminative subspace are most likely to be from the same pedestrian. And then, after labeling the unlabeled samples, further constraining the existing subspace by using the labeled samples, wherein the weight of the constraint is the cosine distance of the two samples in the new projection space.
S150: the regularization term: the purpose of adding the regularization term is to make the learned projection matrix more sparse while avoiding the occurrence of overfitting. Here, we use L2, a 1 norm to constrain the projection matrix U:
Ω(U)=||U||2,1 (4)。
as an improvement, N sampled in S130PThe sampling strategy of each sample is a sampling strategy for maximizing the top-k recognition rate, namely for each image, all samples of k nearest neighbors are sampled. Therefore, the judgment information of the labeled sample can be utilized to the maximum extent while over-fitting is avoided.
When using random gradient descent for optimization, all samples need to be fed into the model in batches. All samples are sampled randomly, a small fraction of classes, two images per class, are randomly selected each time. In the loss calculation, all sample pairs that all images in each batch may make up participate in the calculation. In this way, although many pairs of samples can be computed at one time, due to the randomness in class sampling, the optimal direction of such sampling may not be the direction that can make the target fall down the fastest. Every time the optimization is completed, the distances of all samples under the current model are calculated. To make the target fall down faster, only one pair of negative examples that are the most difficult under the current model is selected for each image, as in FIG. 2.
It is noted that some positive sample pairs have too large an intra-class difference due to drastic changes, and if these samples are trained, it is highly likely that the model will be overfit, as shown in FIG. 3. To avoid this over-fitting, each picture was sampled with a modest positive sample (a moderate positive sample) in the manner shown in fig. 2, i.e., with the first positive sample being just smaller than the hardest negative sample. To exploit as much of the information provided by labeled exemplars as possible, we propose a sampling strategy that maximizes the top-k recognition rate. As shown in fig. 2, for each image, all samples in k nearest neighbors are sampled, so that discrimination information of labeled samples can be maximally utilized while over-fitting is avoided.
As an improvement, in S140, the method for labeling the label-free samples by using the method of K nearest neighbors includes:
as in fig. 1, P1-P4 are four positive samples of the picture to be retrieved, but do not rank in the first four bits of the nearest neighbor picture, which introduces large errors if the K nearest neighbor result is used directly. It should be noted, however, that the picture to be retrieved and the four positive samples are respectively K-nearest neighbors to each other, which we will refer to as K-nearest neighbors. If unlabeled data is labeled in this manner, error introduction is reduced to some extent.
The K nearest neighbors N (x, K) of sample x are defined as follows:
N(x,k)={x1,x2,...,xk},|N(p,k)|=k (5);
where | represents the number of samples in the set, then K is defined as the mutual nearest R (x, K) as follows:
R(x,k)={xi|(xi∈N(x,k))∧(x∈N(xi,k))} (6)。
s200: adopting a projection matrix U epsilon R for a new sampled×cAnd projecting to the new subspace to obtain a prediction sample sequence, wherein the prediction sample sequence is arranged from small to large according to Euclidean distances between the new samples and samples in the training sample set. The prediction samples are ranked top, indicating the highest likelihood of being the same person between the new sample and the prediction sample.
Experiment and analysis:
selecting characteristics: in order to quickly verify the validity of the proposed method, the LOMO feature and the GOG feature commonly used in the task of pedestrian re-identification are used herein.
Setting parameters: the algorithm is implemented using the theono framework. Wherein the minimum interval margin is 0.5, the balance coefficients alpha and lambda are respectively 0.005 and 0.0001, the mapping to the subspace dimension c is 512, and the batch size, the learning rate and k are respectively 32, 1 and 10.
VIPeR database test results and analysis
The VIPeR database is one of the most popular databases for pedestrian re-identification tasks. It contains 1264 images of 632 pedestrians collected by two cameras with different lighting conditions with 90 deg. view angle change. We freely selected 316 pedestrians to compose the training set, and the remaining 316 pedestrians to compose the testing set, and performed semi-supervised and fully-supervised experimental setup, respectively.
Semi-supervised experiments: for the semi-supervised setup, we randomly took 1/3 pedestrian pictures in the training set to erase the tag as unlabeled samples and the remaining 2/3 pedestrian pictures as tagged samples. The results are shown in Table 4.1. Comparing the method provided by the invention with SSCDL and DLLAP, the method provided by the invention can be found to greatly improve the performance, and particularly the Rank-1 recognition rate can reach 47.5% after the LOMO characteristic and the GOG characteristic are combined.
TABLE 4.1 comparison of recognition rates of semi-supervised learning methods in VIPeR database
Rank 1 5 10 20
SSCDL 25.6 53.7 68.2 83.6
DLLAP 32.5 61.8 74.3 84.1
LOMO+Our 34.2 65.2 76.4 85.4
GOG+Our 42.4 73.4 83.9 91.0
LOMO+GOG+Our 47.5 78.3 86.9 92.1
And (3) full supervision experiment: we also performed a fully supervised setup of the method proposed herein, i.e. using labels of all training samples. The results are shown in Table 4.2. Comparison with DLLAP and L1Graph shows that the methods proposed herein are greatly enhanced when using the GOG feature and using the LOMO and GOG features in combination. Compared with a semi-supervised setting, the method can be seen that when LOMO and GOG characteristics are used, the recognition rate of 47.5% can be achieved by using only the label of 2/3 training samples, which is only 3% different from that in the full-supervised case, and the effectiveness of the method provided by the invention is fully proved.
TABLE 4.2 comparison of recognition rates under fully supervised settings on the VIPeR database
Rank 1 5 10 20
DLLAP[41] 38.5 70.8 78.5 86.1
L1Graph[42] 41.5 - - -
LOMO+Our 36.1 68.2 79.6 88.5
GOG+Our 48.6 77.1 87.3 92.9
LOMO+GOG+Our 50.5 79.6 88.8 94.3
The method of the invention uses the contrast loss function to fully utilize the label information of the labeled sample, and uses the K mutual nearest neighbor method to replace the K nearest neighbor method to label the unlabeled sample. The experimental results on the pedestrian re-identification public data set VIPeR confirm the effectiveness of the method.
Finally, the above embodiments are only for illustrating the technical solutions of the present invention and not for limiting, although the present invention has been described in detail with reference to the preferred embodiments, it should be understood by those skilled in the art that modifications or equivalent substitutions may be made to the technical solutions of the present invention without departing from the spirit and scope of the technical solutions of the present invention, and all of them should be covered in the claims of the present invention.

Claims (5)

1. The pedestrian re-identification method based on semi-supervised learning is characterized by comprising the following steps of:
s100: learning a projection matrix U ∈ Rd×cProjecting the original d-dimensional feature space to a c-dimensional subspace, so that UTX∈Rc ×NSatisfying in the new subspace: the Euclidean distance between sample pairs from the same pedestrian is smaller, and the Euclidean distance between sample pairs from different pedestrians is larger; samples from the same pedestrian are defined as similar samples, and samples from different pedestrians are defined as different samples;
s200: adopting a projection matrix U epsilon R for a new sampled×cAnd projecting to the new subspace to obtain a prediction sample sequence, wherein the prediction sample sequence is arranged from small to large according to Euclidean distances between the new samples and samples in the training sample set.
2. The pedestrian re-identification method based on semi-supervised learning as claimed in claim 1, wherein the learning projection matrix U e R in S100d×cThe method specifically comprises the following steps:
s110, establishing a training sample set, wherein the training sample set comprises a plurality of samples, the plurality of samples comprise labeled samples and unlabeled samples, and the labels of the samples of the same pedestrian in the labeled samples are the same;
let X ═ XL,XU]∈Rd×NRepresents all training samples, where N is the number of all pictures included in the training set, d is the length of the feature vector,
Figure FDA0002683805600000011
represents NLThe sample of the individual label is provided with a label,
Figure FDA0002683805600000012
represents NUEach unlabeled sample;
s120, establishing an objective function as follows:
Figure FDA0002683805600000013
wherein L (U) is a regression function,
Figure FDA0002683805600000014
for a weighted regression function, Ω (U) is a regularized constraint, α, λ > 0 are equilibrium coefficients;
s130, taking the labeled sample loss function as a contrast loss function: for sampled NPA sample pair
Figure FDA0002683805600000015
And
Figure FDA0002683805600000016
if it is
Figure FDA0002683805600000017
And
Figure FDA0002683805600000018
samples from the same pedestrian are then in the new projection space
Figure FDA0002683805600000019
And
Figure FDA00026838056000000110
has an Euclidean distance d betweennShould be as small as possible, close to 0; otherwise, then dnShould be at least greater than a predetermined threshold margin>0, if the above condition is not satisfied, a loss occurs;
s140, labeling the label-free samples by adopting a K mutual nearest neighbor method, wherein the loss function of the label-free samples is as follows:
Figure FDA00026838056000000111
wherein if UTxiAnd UTxjSatisfy K mutual nearest neighbor and xiAnd xjFrom different cameras, then
Figure FDA0002683805600000021
Otherwise Wij=0; (8);
After labeling the label-free samples, further constraining the existing subspace by using the labeled samples, wherein the weight of the constraint is the cosine distance of the two samples in the new projection space;
s150: the regularization term: the projection matrix U is constrained using L2,1 norm:
Ω(U)=||U||2,1 (4)。
3. the pedestrian re-identification method based on semi-supervised learning according to claim 2, wherein the labeled sample loss function of S130 is:
Figure FDA0002683805600000022
wherein:
Figure FDA0002683805600000023
Figure FDA0002683805600000024
4. semi-supervised learning based pedestrian re-identification as claimed in claim 2Method, characterized in that N sampled in S130PThe sampling strategy of each sample is a sampling strategy for maximizing the top-k recognition rate, namely for each image, all samples of k nearest neighbors are sampled.
5. The pedestrian re-identification method based on semi-supervised learning as claimed in claim 2, wherein the method for labeling the unlabeled samples by using the K mutual nearest neighbor method in S140 is as follows:
the K nearest neighbors N (x, K) of sample x are defined as follows:
N(x,k)={x1,x2,...,xk},|N(p,k)|=k (5);
where | represents the number of samples in the set, then K is defined as the mutual nearest R (x, K) as follows:
R(x,k)={xi|(xi∈N(x,k))∧(x∈N(xi,k))} (6)。
CN202010970306.5A 2020-09-15 2020-09-15 Pedestrian re-identification method based on semi-supervised learning Active CN112101217B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010970306.5A CN112101217B (en) 2020-09-15 2020-09-15 Pedestrian re-identification method based on semi-supervised learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010970306.5A CN112101217B (en) 2020-09-15 2020-09-15 Pedestrian re-identification method based on semi-supervised learning

Publications (2)

Publication Number Publication Date
CN112101217A true CN112101217A (en) 2020-12-18
CN112101217B CN112101217B (en) 2024-04-26

Family

ID=73758623

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010970306.5A Active CN112101217B (en) 2020-09-15 2020-09-15 Pedestrian re-identification method based on semi-supervised learning

Country Status (1)

Country Link
CN (1) CN112101217B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113657176A (en) * 2021-07-22 2021-11-16 西南财经大学 Pedestrian re-identification implementation method based on active contrast learning
CN116052095A (en) * 2023-03-31 2023-05-02 松立控股集团股份有限公司 Vehicle re-identification method for smart city panoramic video monitoring

Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020122593A1 (en) * 2000-12-22 2002-09-05 Fuji Xerox Co., Ltd. Pattern recognition method and apparatus
US8527432B1 (en) * 2008-08-08 2013-09-03 The Research Foundation Of State University Of New York Semi-supervised learning based on semiparametric regularization
US20140280232A1 (en) * 2013-03-14 2014-09-18 Xerox Corporation Method and system for tagging objects comprising tag recommendation based on query-based ranking and annotation relationships between objects and tags
US20150117764A1 (en) * 2013-10-29 2015-04-30 Nec Laboratories America, Inc. Efficient distance metric learning for fine-grained visual categorization
CN107145827A (en) * 2017-04-01 2017-09-08 浙江大学 Across the video camera pedestrian recognition methods again learnt based on adaptive distance metric
CN109522956A (en) * 2018-11-16 2019-03-26 哈尔滨理工大学 A kind of low-rank differentiation proper subspace learning method
CN110008828A (en) * 2019-02-21 2019-07-12 上海工程技术大学 Pairs of constraint ingredient assay measures optimization method based on difference regularization
CN110008842A (en) * 2019-03-09 2019-07-12 同济大学 A kind of pedestrian's recognition methods again for more losing Fusion Model based on depth
CN110175511A (en) * 2019-04-10 2019-08-27 杭州电子科技大学 It is a kind of to be embedded in positive negative sample and adjust the distance pedestrian's recognition methods again of distribution
CN111027442A (en) * 2019-12-03 2020-04-17 腾讯科技(深圳)有限公司 Model training method, recognition method, device and medium for pedestrian re-recognition
CN111027421A (en) * 2019-11-26 2020-04-17 西安宏规电子科技有限公司 Graph-based direct-push type semi-supervised pedestrian re-identification method
CN111033509A (en) * 2017-07-18 2020-04-17 视语智能有限公司 Object re-identification
US20200125897A1 (en) * 2018-10-18 2020-04-23 Deepnorth Inc. Semi-Supervised Person Re-Identification Using Multi-View Clustering
CN111144451A (en) * 2019-12-10 2020-05-12 东软集团股份有限公司 Training method, device and equipment of image classification model
CN111353516A (en) * 2018-12-21 2020-06-30 华为技术有限公司 Sample classification method and model updating method for online learning
CN111563424A (en) * 2020-04-20 2020-08-21 清华大学 Pedestrian re-identification method and device based on semi-supervised learning

Patent Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020122593A1 (en) * 2000-12-22 2002-09-05 Fuji Xerox Co., Ltd. Pattern recognition method and apparatus
US8527432B1 (en) * 2008-08-08 2013-09-03 The Research Foundation Of State University Of New York Semi-supervised learning based on semiparametric regularization
US20140280232A1 (en) * 2013-03-14 2014-09-18 Xerox Corporation Method and system for tagging objects comprising tag recommendation based on query-based ranking and annotation relationships between objects and tags
US20150117764A1 (en) * 2013-10-29 2015-04-30 Nec Laboratories America, Inc. Efficient distance metric learning for fine-grained visual categorization
CN107145827A (en) * 2017-04-01 2017-09-08 浙江大学 Across the video camera pedestrian recognition methods again learnt based on adaptive distance metric
CN111033509A (en) * 2017-07-18 2020-04-17 视语智能有限公司 Object re-identification
US20200125897A1 (en) * 2018-10-18 2020-04-23 Deepnorth Inc. Semi-Supervised Person Re-Identification Using Multi-View Clustering
CN109522956A (en) * 2018-11-16 2019-03-26 哈尔滨理工大学 A kind of low-rank differentiation proper subspace learning method
CN111353516A (en) * 2018-12-21 2020-06-30 华为技术有限公司 Sample classification method and model updating method for online learning
CN110008828A (en) * 2019-02-21 2019-07-12 上海工程技术大学 Pairs of constraint ingredient assay measures optimization method based on difference regularization
CN110008842A (en) * 2019-03-09 2019-07-12 同济大学 A kind of pedestrian's recognition methods again for more losing Fusion Model based on depth
CN110175511A (en) * 2019-04-10 2019-08-27 杭州电子科技大学 It is a kind of to be embedded in positive negative sample and adjust the distance pedestrian's recognition methods again of distribution
CN111027421A (en) * 2019-11-26 2020-04-17 西安宏规电子科技有限公司 Graph-based direct-push type semi-supervised pedestrian re-identification method
CN111027442A (en) * 2019-12-03 2020-04-17 腾讯科技(深圳)有限公司 Model training method, recognition method, device and medium for pedestrian re-recognition
CN111144451A (en) * 2019-12-10 2020-05-12 东软集团股份有限公司 Training method, device and equipment of image classification model
CN111563424A (en) * 2020-04-20 2020-08-21 清华大学 Pedestrian re-identification method and device based on semi-supervised learning

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
G. DING等: "Center Based Pseudo-Labeling For Semi-Supervised Person Re-Identification", 2018 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW), 29 November 2018 (2018-11-29), pages 1 - 6 *
张化祥等: "行人重识别研究综述", 山东师范大学学报(自然科学版), vol. 33, no. 04, 31 December 2018 (2018-12-31), pages 379 - 387 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113657176A (en) * 2021-07-22 2021-11-16 西南财经大学 Pedestrian re-identification implementation method based on active contrast learning
CN116052095A (en) * 2023-03-31 2023-05-02 松立控股集团股份有限公司 Vehicle re-identification method for smart city panoramic video monitoring

Also Published As

Publication number Publication date
CN112101217B (en) 2024-04-26

Similar Documents

Publication Publication Date Title
CN111814584B (en) Vehicle re-identification method based on multi-center measurement loss under multi-view environment
Ren et al. Instance-aware, context-focused, and memory-efficient weakly supervised object detection
Liu et al. Stepwise metric promotion for unsupervised video person re-identification
Deng et al. Rethinking triplet loss for domain adaptation
CN106682696B (en) The more example detection networks and its training method refined based on online example classification device
Elhamifar et al. A convex optimization framework for active learning
CN103150580B (en) A kind of high spectrum image semisupervised classification method and device
CN111950372B (en) Unsupervised pedestrian re-identification method based on graph convolution network
US20210319215A1 (en) Method and system for person re-identification
CN111598004B (en) Progressive reinforcement self-learning unsupervised cross-domain pedestrian re-identification method
Wu et al. Vehicle re-identification with the space-time prior
CN113095442B (en) Hail identification method based on semi-supervised learning under multi-dimensional radar data
CN109635708B (en) Unsupervised pedestrian re-identification method based on three-data-set cross migration learning
CN112633382B (en) Method and system for classifying few sample images based on mutual neighbor
CN108921107A (en) Pedestrian's recognition methods again based on sequence loss and Siamese network
CN112101217B (en) Pedestrian re-identification method based on semi-supervised learning
CN113158955B (en) Pedestrian re-recognition method based on clustering guidance and paired measurement triplet loss
Lan et al. Robust multi-modality anchor graph-based label prediction for RGB-infrared tracking
CN112115780A (en) Semi-supervised pedestrian re-identification method based on deep multi-model cooperation
CN116052212A (en) Semi-supervised cross-mode pedestrian re-recognition method based on dual self-supervised learning
Zheng et al. Adaptive boosting for domain adaptation: Toward robust predictions in scene segmentation
CN114495004A (en) Unsupervised cross-modal pedestrian re-identification method
CN112052722A (en) Pedestrian identity re-identification method and storage medium
Liu et al. Automatic inline defect detection for a thin film transistor–liquid crystal display array process using locally linear embedding and support vector data description
Rahimpour et al. Attention-based few-shot person re-identification using meta learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant