CN109635728B - Heterogeneous pedestrian re-identification method based on asymmetric metric learning - Google Patents

Heterogeneous pedestrian re-identification method based on asymmetric metric learning Download PDF

Info

Publication number
CN109635728B
CN109635728B CN201811515924.XA CN201811515924A CN109635728B CN 109635728 B CN109635728 B CN 109635728B CN 201811515924 A CN201811515924 A CN 201811515924A CN 109635728 B CN109635728 B CN 109635728B
Authority
CN
China
Prior art keywords
features
constraint
pedestrian
local
global
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201811515924.XA
Other languages
Chinese (zh)
Other versions
CN109635728A (en
Inventor
赖剑煌
程海杰
张权
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sun Yat Sen University
Original Assignee
Sun Yat Sen University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sun Yat Sen University filed Critical Sun Yat Sen University
Priority to CN201811515924.XA priority Critical patent/CN109635728B/en
Publication of CN109635728A publication Critical patent/CN109635728A/en
Application granted granted Critical
Publication of CN109635728B publication Critical patent/CN109635728B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • G06V20/53Recognition of crowd images, e.g. recognition of crowd congestion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/19Recognition using electronic means
    • G06V30/192Recognition using electronic means using simultaneous comparisons or correlations of the image signals with a plurality of references
    • G06V30/194References adjustable by an adaptive method, e.g. learning

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a heterogeneous pedestrian re-identification method based on asymmetric metric learning, which carries out asymmetric metric on depth features under different modes and comprises the following steps: respectively projecting different modal depth features to a shared space by using two sparse self-encoders which do not share parameters, and simultaneously introducing global constraint and local constraint to constrain the distance between the different modal depth features, so that the intra-class distance between the different modal features is reduced and the inter-class distance between the different modal features is increased; and the constraint results of the global constraint and the local constraint are used as supervision signals to be propagated back to the training network for correcting various parameters. According to the invention, by reducing the modal difference between different modes, the network can ignore modal information as much as possible and pay more attention to identity information, thereby improving the pedestrian characteristic expression and the pedestrian matching accuracy.

Description

Heterogeneous pedestrian re-identification method based on asymmetric metric learning
Technical Field
The invention relates to the field of computer vision, in particular to a heterogeneous pedestrian re-identification method based on asymmetric metric learning.
Background
Along with the rapid development of modern society, the urban population density is higher and higher, and people pay more and more attention to safety problems. In order to prevent and avoid safety incidents in time, a large number of monitoring cameras are installed and applied in public places. In the face of a complex monitoring network and massive monitoring data, how to automatically analyze and interpret information provided by a multi-camera vision field monitoring system is realized, and the system has a positive promoting effect on preventing crimes and maintaining good social security. Therefore, pedestrian re-identification has become a hot research content in the field of computer vision.
Pedestrian re-identification (person re-identification) is used as a key component in the field of video monitoring research, and aims to accurately and quickly identify a certain target pedestrian appearing in the visual field of a monitoring camera from a large number of pedestrians in the visual fields of other cameras of a monitoring network. The application of the pedestrian re-identification technology can greatly reduce the manual participation in video monitoring and realize the rapid and accurate analysis of the pedestrians and the behaviors thereof in the monitoring video. At present, the mainstream pedestrian re-identification method mainly extracts the appearance features and color features (RGB features) of pedestrians for matching, and these methods can be regarded as RGB-RGB single-mode pedestrian matching. However, these methods all have a strong assumption: assuming that the clothing of the same pedestrian in the presence of different cameras remains as unchanged as possible, it can be considered as a short-time pedestrian re-identification. Thus, when the pedestrian's clothing changes significantly or when the color characteristics of the pedestrian become unavailable, the performance of these methods will decline dramatically because the color characteristics at this time more show interference with the model, largely misjudging different pedestrians wearing the same color clothing as the same pedestrian. Therefore, in recent years, in order to overcome the defect of color feature failure under extreme conditions, data of other modes is introduced to make up the deficiency of RGB data, such as infrared data (IR data), which can be regarded as RGB-IR cross-mode pedestrian matching (heterogeneous pedestrian matching), and the biggest challenge of heterogeneous pedestrian re-identification is how to narrow the modal gap between different modes. At present, a scholars proposes a depth zero filling method to reduce the modal gap between different modes, but the identification result of the method is not accurate and can not meet the requirement of practical application.
Disclosure of Invention
The invention aims to overcome the defects and shortcomings of the prior art, and provides a heterogeneous pedestrian re-identification method based on asymmetric metric learning, which can overcome the defects of color feature failure and low accuracy of heterogeneous pedestrian re-identification under extreme conditions, and can reduce modal differences among different modes to enable a network to ignore modal information as much as possible and pay more attention to identity information, thereby improving pedestrian feature expression and pedestrian matching accuracy.
The purpose of the invention is realized by the following technical scheme: a heterogeneous pedestrian re-identification method based on asymmetric metric learning comprises the following steps:
inputting pedestrian images in two modes in the process of training a model, and respectively extracting depth features;
carrying out asymmetric measurement on depth features under different modes, and the method comprises the following steps: respectively projecting different modal depth features to a shared space by using two sparse self-encoders which do not share parameters, and simultaneously introducing global constraint and local constraint to constrain the distance between the different modal depth features, so that the intra-class distance between the different modal features is reduced and the inter-class distance between the different modal features is increased; the constraint results of the global constraint and the local constraint are used as supervision signals to be reversely transmitted to a training network for correcting each parameter;
and calculating the loss of the global feature and the local feature according to the depth feature, and optimizing the training model by using the global loss, the local loss and the sum of the global constraint and the local constraint in the asymmetry metric to achieve minimization.
Through the steps, the heterogeneous pedestrian re-identification model can be trained as long as pedestrian re-identification training data of any two modes are given, and the heterogeneous pedestrian re-identification model has the advantages of high precision and high speed in heterogeneous pedestrian matching.
Preferably, the depth features of the images in different modalities are extracted, and the method comprises the following steps:
firstly, a ResNet50 classification model pre-trained on an ImageNet data set is used as a main network, and the main network is divided into three branches;
then, from top to bottom, each branch extracts the high-level features of the classification model and horizontally and uniformly divides the high-level features into blocks;
then, obtaining a plurality of global features and local features with fixed sizes by each branch through pooling and dimension reduction operations;
and finally, splicing the global features and the local features together in sequence to obtain the depth features of the input image, namely the complete feature expression of the pedestrian.
In order to reduce the modal difference among heterogeneous pedestrian data, the invention carries out asymmetric measurement on depth characteristics under different modes, and the steps are as follows:
first, the extracted depth features are divided into FBAnd FRTwo groups of the first and the second groups of the second,
Figure BDA0001901956920000031
Figure BDA0001901956920000032
b, R respectively representing an RGB mode and an IR mode,
Figure BDA0001901956920000033
representing an ith depth feature vector;
next, two sets of features F are combinedBAnd FRRespectively passing through two sparse self-encoders SAE which do not share parameters, wherein each sparse self-encoder is composed of two fully-connected layers and respectively serves as an encoder E and a decoder D, the encoder E is responsible for projecting different modal characteristics to a shared space, and the decoder D is responsible for remapping the encoded characteristics to a space with the same size as an input characteristic space;
then, a reconstruction loss is constructed, denoted as lrFor constraining the output and input of the SAE to be as consistent as possible:
lr=||fB,DB(EB(fB))||2+||fR,DR(ER(fR))||2
fB,fR,EB,ER,DB,DRfeatures, encoders and decoders representing modes B and R, respectively;
and finally, introducing global constraint for constraining the difference between different modal characteristic distributions in a shared space, introducing local constraint for reducing the intra-class distance and increasing the inter-class distance between different modal characteristics, and reversely transmitting the constraint result serving as a supervision signal back to the training model to correct each parameter.
Further, global constraints are used to constrain the differences between the distributions of different modal features, denoted as lglobal=W(EB(fB),ER(fR))2Where W satisfies X-N (m) for any given two distributionsX,CX) And Y ═ N (m)r,CY) M, C represent the mean and variance of the X and Y distributions, respectively, having
Figure BDA0001901956920000034
Figure BDA0001901956920000035
Further, local constraints are used to reduce the intra-class distance between different modal featuresAnd increasing the distance between classes, denoted as llocalThe distance between the two features is represented by (max (d (f, p)) -min (d (f, n)) + α), p ∈ a (f), n ∈ b (f), a (f), b (f) respectively represent feature sets with the same identity information and different identity information as the feature f, d (·, · denotes the euclidean distance between the two features, and α is a hyper-parameter for controlling the distance between the positive and negative sample pairs.
Further, to make the shared space more efficient, a sparse loss l is constructedsparseUnconstrained hidden layer output,/sparse=||EB(fB)||1+||ER(fR)||1
Furthermore, each sparse autoencoder consists of two fully-connected layers with a ReLU activation function.
Preferably, for the extracted depth features, the triple Loss function is used for calculating the Loss of the global features, the Softmax function is used for calculating the Loss of the local features, and the goal is to optimize the training model by minimizing the sum of the global Loss, the local Loss and the reconstruction Loss, the sparse Loss, the global constraint and the local constraint in the asymmetry measurement. Thereby improving the expression force of the model characteristics.
Compared with the prior art, the invention has the following advantages and beneficial effects:
1. aiming at the defects of large modal information difference and large pedestrian matching difficulty between different modalities in the existing heterogeneous pedestrian re-identification task, the invention provides the asymmetric measurement of depth features under different modalities to reduce the feature distance between the same pedestrians under different modalities, can be applied to any feature extraction network to provide supervision information for a training network and realize end-to-end training, and effectively improves the quality of pedestrian feature extraction and accelerates network convergence.
2. In the invention, different loss functions are adopted for different characteristics in a differentiated mode for collaborative training, and compared with single loss function training, the method purposefully enables the network to ignore modal information and pay more attention to pedestrian identity information as far as possible, so that more complete characteristic expression of pedestrians is obtained, and the method is far better than the existing method in precision.
3. The invention uses the idea of combining the global and local parts to extract the pedestrian characteristics, and compared with the single characteristics, the invention obtains more complete characteristic expression of the pedestrian, thereby improving the precision.
Drawings
Fig. 1 is a general functional framework diagram of the method of the present embodiment.
Detailed Description
The drawings are for illustrative purposes only and are not to be construed as limiting the patent; it will be understood by those skilled in the art that certain well-known structures in the drawings and descriptions thereof may be omitted. The technical solution of the present invention is further described below with reference to the accompanying drawings and examples.
As shown in fig. 1, the method for re-identifying a heterogeneous pedestrian based on asymmetric metric learning mainly includes three steps of feature extraction, asymmetric metric and classification, and each step is specifically described below.
Firstly, feature extraction
The method for extracting the features of the pedestrian comprises the following steps:
firstly, a ResNet50 classification model pre-trained on an ImageNet data set is used as a main network, and the main network is divided into three branches;
then, from top to bottom, each branch extracts the high-level features of the classification model and horizontally and uniformly divides the high-level features into 1 block, 2 blocks and 3 blocks;
then, each branch is subjected to pooling and dimension reduction operation to obtain 3 global features and 5 local features with fixed size of 256 dimensions;
finally, 8 256-dimensional features are spliced together in sequence to obtain a 2048-dimensional feature which is used as the depth feature of the input image, namely the complete feature expression of the pedestrian. The above features are used for later asymmetry measurement and classification.
The invention uses the idea of combining the global and local parts to extract the pedestrian characteristics, and compared with the single characteristics, the invention obtains more complete characteristic expression of the pedestrian, thereby improving the precision.
Two, asymmetric metric
The method comprises the following steps of projecting and reconstructing characteristics under different modes through different projection matrixes to reduce the mode difference between heterogeneous pedestrian data, wherein the steps are as follows:
first, for the extracted features
Figure BDA0001901956920000051
(B/R stands for RGB modality/IR modality, respectively), the extracted features are divided into F by a modality selectorBAnd FRTwo groups;
next, two sets of features F are combinedBAnd FRRespectively passing through two sparse autoencoders SAE without sharing parameters, wherein each sparse autoencoder is composed of two full-connection layers with a ReLU activation function and respectively used as an encoder E and a decoder D, the encoder E is responsible for projecting different modal characteristics to a shared space, the decoder D is responsible for remapping the encoded characteristics to a space with the same size as an input characteristic space, and then reconstruction loss is constructed and recorded as lrAnd is used to constrain the output and input of the SAE to be as consistent as possible. lr=||fB,DB(EB(fB))||2+||fR,DR(ER(fR))||2,fB,fR,EB,ER,DB,DRRepresenting the characteristics of modality B and modality R, respectively, an encoder and a decoder.
Meanwhile, in order to make the shared space more effective, a sparse loss l is constructedsparseUnconstrained hidden layer output,/sparse=||EB(fB)||1+||ER(fR)||1
Finally, in the shared space, a global constraint l is introducedglobalAnd local constraint llocalAnd constraining the distance between different modal characteristics, and reversely transmitting the distance as a supervision signal to other modules of the network to correct parameters of the characteristic extraction step, so that the characteristic extraction ignores modal information as much as possible and pays attention to pedestrian identity information, thereby improving the expression of image characteristics.
Global constraint is used to constrain the difference between different modal feature distributions, denoted as lglobal=W(EB(fB),ER(fR))2Where W satisfies X-N (m) for any given two distributionsX,CX) And Y ═ N (m)Y,CY) M, C represent the mean and variance of the X and Y distributions, respectively, having
Figure BDA0001901956920000052
Figure BDA0001901956920000061
Local constraints are used to reduce the intra-class distance and increase the inter-class distance between different modal features, denoted as llocalThe distance between the two features is represented by (max (d (f, p)) -min (d (f, n)) + α), p ∈ a (f), n ∈ b (f), a (f)/b (f) respectively represent feature sets with the same identity information and different identity information as the feature f, d (·,) represents the euclidean distance between the two features, and α is a hyper-parameter for controlling the distance between the positive sample pairs and the negative sample pairs.
According to the invention, two sparse self-encoders which do not share parameters are used for projecting different modal characteristics to a shared space respectively. Introducing global constraint l simultaneouslyglobalAnd local constraint llocalThe distance between different modal features is constrained, so that the intra-class distance and the inter-class distance between the different modal features are reduced, modal information is effectively ignored in the feature extraction process to pay attention to ID information, and the expression of image features is improved.
Three, classification
The step is that different loss collaborative training is adopted to distinguish input features so as to effectively restrain pedestrian features, and the step is as follows: for the depth features extracted from the feature extraction step, the triple Loss function is used for calculating the Loss of 3 global features, the Softmax function is used for calculating the Loss of five local features, and then the model is optimized by jointly minimizing the global Loss, the local Loss and the sum of reconstruction Loss, sparse Loss, global constraint and local constraint in the asymmetric measurement module, so that the model feature expression is improved.
In the invention, different loss functions are adopted for different characteristics in a differentiated mode for collaborative training, and compared with single loss function training, the method purposefully enables the network to ignore modal information and pay more attention to pedestrian identity information as far as possible, so that more complete characteristic expression of pedestrians is obtained, and the method is far better than the existing method in precision.
Experimental results show that the Rank1 and the mAP of the invention on the current maximum cross-modal pedestrian recognition data set SYSU-MM01 are respectively improved from 24.43% and 26.92% to 66.26% and 66.7%, and the performance is greatly improved compared with other methods.
The techniques described herein may be implemented by various means. For example, these techniques may be implemented in hardware, firmware, software, or a combination thereof. For a hardware implementation, the processing modules may be implemented within one or more Application Specific Integrated Circuits (ASICs), Digital Signal Processors (DSPs), Programmable Logic Devices (PLDs), field-programmable gate arrays (FPGAs), processors, controllers, micro-controllers, electronic devices, other electronic units designed to perform the functions described herein, or a combination thereof.
For a firmware and/or software implementation, the techniques may be implemented with modules (e.g., procedures, steps, flows, and so on) that perform the functions described herein. The firmware and/or software codes may be stored in a memory and executed by a processor. The memory may be implemented within the processor or external to the processor.
Those of ordinary skill in the art will understand that: all or part of the steps for realizing the method embodiments can be completed by hardware related to program instructions, the program can be stored in a computer readable storage medium, and the program executes the steps comprising the method embodiments when executed; and the aforementioned storage medium includes: various media that can store program codes, such as ROM, RAM, magnetic or optical disks.
It should be understood that the above-described embodiments of the present invention are merely examples for clearly illustrating the present invention, and are not intended to limit the embodiments of the present invention. Other variations and modifications will be apparent to persons skilled in the art in light of the above description. And are neither required nor exhaustive of all embodiments. Any modification, equivalent replacement, and improvement made within the spirit and principle of the present invention should be included in the protection scope of the claims of the present invention.

Claims (5)

1. A heterogeneous pedestrian re-identification method based on asymmetric metric learning is characterized by comprising the following steps:
inputting pedestrian images in two modes in the process of training a model, and respectively extracting depth features;
carrying out asymmetric measurement on depth features under different modes, and the method comprises the following steps: respectively projecting different modal depth features to a shared space by using two sparse self-encoders which do not share parameters, and simultaneously introducing global constraint and local constraint to constrain the distance between the different modal depth features, so that the intra-class distance between the different modal features is reduced and the inter-class distance between the different modal features is increased; the constraint results of the global constraint and the local constraint are used as supervision signals to be reversely transmitted to a training network for correcting each parameter;
calculating the loss of the global characteristic and the local characteristic according to the depth characteristic, and optimizing the training model by using the global loss, the local loss and the minimization of the sum of the global constraint and the local constraint in the asymmetry measurement;
carrying out asymmetric measurement on depth features under different modes, and the method comprises the following steps:
first, the extracted depth features are divided into FBAnd FRTwo groups of the first and the second groups of the second,
Figure FDA0002538650360000011
Figure FDA0002538650360000012
b, R represents RGB mode, IR mode, and fi mRepresenting an ith depth feature vector;
next, two sets of features F are combinedBAnd FRRespectively through two notEach sparse self-encoder is composed of two fully-connected layers and serves as an encoder E and a decoder D respectively, the encoder E is responsible for projecting different modal characteristics to a shared space, and the decoder D is responsible for remapping the encoded characteristics to a space with the same size as an input characteristic space;
then, a reconstruction loss is constructed, denoted as lrFor constraining the output and input of the SAE to be as consistent as possible:
lr=||fB,DB(EB(fB))||2+||fR,DR(ER(fR))||2
fB,EB,DBfeatures representing modality B, an encoder and a decoder; f. ofR,ER,DRFeatures representing the modality R, encoder and decoder;
finally, introducing global constraint for constraining the difference between different modal characteristic distributions, introducing local constraint for reducing the intra-class distance and increasing the inter-class distance between different modal characteristics, and reversely transmitting the constraint result serving as a supervision signal back to the training model to correct each parameter in the shared space;
global constraint is used to constrain the difference between different modal feature distributions, denoted as lglobal=W(EB(fB),ER(fR))2Where W satisfies X-N (m) for any given two distributionsX,CX) And Y ═ N (m)Y,CY) M, C represent the mean and variance of the X and Y distributions, respectively, having
Figure FDA0002538650360000013
Local constraints are used to reduce the intra-class distance and increase the inter-class distance between different modal features, denoted as llocal(max (d (f, p)) -min (d (f, n)) + α), p ∈ a (f), n ∈ b (f), a (f), b (f) respectively represent feature sets having the same identity information as feature f and different identity information, d (·,) represents the euclidean distance between the two features,α is a hyperparameter for controlling the spacing of the positive and negative sample pairs.
2. The asymmetric metric learning-based heterogeneous pedestrian re-identification method according to claim 1, wherein depth features of images in different modalities are extracted, and the method comprises the following steps:
firstly, a ResNet50 classification model pre-trained on an ImageNet data set is used as a main network, and the main network is divided into three branches;
then, from top to bottom, each branch extracts the high-level features of the classification model and horizontally and uniformly divides the high-level features into blocks;
then, obtaining a plurality of global features and local features with fixed sizes by each branch through pooling and dimension reduction operations;
and finally, splicing the global features and the local features together in sequence to obtain the depth features of the input image, namely the complete feature expression of the pedestrian.
3. The asymmetric metric learning-based heterogeneous pedestrian re-identification method according to claim 1, wherein a sparse loss l is constructedsparseUnconstrained hidden layer output,/sparse=||EB(fB)||1+||ER(fR)||1
4. The asymmetric metric learning-based heterogeneous pedestrian re-identification method according to claim 1, wherein each sparse self-encoder is composed of two fully-connected layers with a ReLU activation function.
5. The asymmetric metric learning-based heterogeneous pedestrian re-identification method as claimed in claim 1, wherein for the extracted depth features, a Triplet Loss function is used to calculate the Loss of global features, a Softmax function is used to calculate the Loss of local features, and the aim of optimizing the training model is to minimize the global Loss, the local Loss and the sum of reconstruction Loss, sparse Loss, global constraint and local constraint in the asymmetric metric.
CN201811515924.XA 2018-12-12 2018-12-12 Heterogeneous pedestrian re-identification method based on asymmetric metric learning Active CN109635728B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811515924.XA CN109635728B (en) 2018-12-12 2018-12-12 Heterogeneous pedestrian re-identification method based on asymmetric metric learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811515924.XA CN109635728B (en) 2018-12-12 2018-12-12 Heterogeneous pedestrian re-identification method based on asymmetric metric learning

Publications (2)

Publication Number Publication Date
CN109635728A CN109635728A (en) 2019-04-16
CN109635728B true CN109635728B (en) 2020-10-13

Family

ID=66073049

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811515924.XA Active CN109635728B (en) 2018-12-12 2018-12-12 Heterogeneous pedestrian re-identification method based on asymmetric metric learning

Country Status (1)

Country Link
CN (1) CN109635728B (en)

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111079939B (en) * 2019-11-28 2021-04-20 支付宝(杭州)信息技术有限公司 Machine learning model feature screening method and device based on data privacy protection
CN112016401B (en) * 2020-08-04 2024-05-17 杰创智能科技股份有限公司 Cross-mode pedestrian re-identification method and device
CN112001438B (en) * 2020-08-19 2023-01-10 四川大学 Multi-mode data clustering method for automatically selecting clustering number
CN112241682B (en) * 2020-09-14 2022-05-10 同济大学 End-to-end pedestrian searching method based on blocking and multi-layer information fusion
CN112016523B (en) * 2020-09-25 2023-08-29 北京百度网讯科技有限公司 Cross-modal face recognition method, device, equipment and storage medium
CN112200093B (en) * 2020-10-13 2022-08-30 北京邮电大学 Pedestrian re-identification method based on uncertainty estimation
CN112699829B (en) * 2021-01-05 2022-08-30 山东交通学院 Vehicle weight identification method and system based on depth feature and sparse measurement projection
CN113128441B (en) * 2021-04-28 2022-10-14 安徽大学 System and method for identifying vehicle weight by embedding structure of attribute and state guidance
CN114550208A (en) * 2022-02-10 2022-05-27 南通大学 Cross-modal pedestrian re-identification method based on global level and local level combined constraint

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106778921A (en) * 2017-02-15 2017-05-31 张烜 Personnel based on deep learning encoding model recognition methods again
CN108345860A (en) * 2018-02-24 2018-07-31 江苏测联空间大数据应用研究中心有限公司 Personnel based on deep learning and learning distance metric recognition methods again

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108960127B (en) * 2018-06-29 2021-11-05 厦门大学 Shielded pedestrian re-identification method based on adaptive depth measurement learning

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106778921A (en) * 2017-02-15 2017-05-31 张烜 Personnel based on deep learning encoding model recognition methods again
CN108345860A (en) * 2018-02-24 2018-07-31 江苏测联空间大数据应用研究中心有限公司 Personnel based on deep learning and learning distance metric recognition methods again

Also Published As

Publication number Publication date
CN109635728A (en) 2019-04-16

Similar Documents

Publication Publication Date Title
CN109635728B (en) Heterogeneous pedestrian re-identification method based on asymmetric metric learning
CN107273872B (en) Depth discrimination network model method for re-identification of pedestrians in image or video
CN109508663B (en) Pedestrian re-identification method based on multi-level supervision network
CN109359552B (en) Efficient cross-camera bidirectional pedestrian tracking method
Huang et al. Highly accurate moving object detection in variable bit rate video-based traffic monitoring systems
Yang et al. Bi-directional progressive guidance network for RGB-D salient object detection
Wu et al. A one-stage domain adaptation network with image alignment for unsupervised nighttime semantic segmentation
US20240029272A1 (en) Matting network training method and matting method
CN109919073B (en) Pedestrian re-identification method with illumination robustness
WO2023109709A1 (en) Image stiching positioning detection method based on attention mechanism
Niloy et al. CFL-Net: image forgery localization using contrastive learning
CN111160295A (en) Video pedestrian re-identification method based on region guidance and space-time attention
CN112365586B (en) 3D face modeling and stereo judging method and binocular 3D face modeling and stereo judging method of embedded platform
CN110472634A (en) Change detecting method based on multiple dimensioned depth characteristic difference converged network
CN111008608B (en) Night vehicle detection method based on deep learning
Liu et al. Hierarchical multimodal fusion for ground-based cloud classification in weather station networks
CN116342953A (en) Dual-mode target detection model and method based on residual shrinkage attention network
CN117036770A (en) Detection model training and target detection method and system based on cascade attention
CN113627504B (en) Multi-mode multi-scale feature fusion target detection method based on generation of countermeasure network
Peng et al. Point-based multilevel domain adaptation for point cloud segmentation
Kong et al. Hole-robust wireframe detection
CN111191549A (en) Two-stage face anti-counterfeiting detection method
Wang et al. A deep learning-based method for vehicle licenseplate recognition in natural scene
CN114004299A (en) Parent-child relationship verification algorithm based on deep learning
CN113674230A (en) Method and device for detecting key points of indoor backlight face

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
OL01 Intention to license declared
OL01 Intention to license declared