CN110210556B - Pedestrian re-identification data generation method - Google Patents

Pedestrian re-identification data generation method Download PDF

Info

Publication number
CN110210556B
CN110210556B CN201910466234.8A CN201910466234A CN110210556B CN 110210556 B CN110210556 B CN 110210556B CN 201910466234 A CN201910466234 A CN 201910466234A CN 110210556 B CN110210556 B CN 110210556B
Authority
CN
China
Prior art keywords
network
countermeasure
migration
image
generation
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910466234.8A
Other languages
Chinese (zh)
Other versions
CN110210556A (en
Inventor
查正军
刘嘉威
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Science and Technology of China USTC
Original Assignee
University of Science and Technology of China USTC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Science and Technology of China USTC filed Critical University of Science and Technology of China USTC
Priority to CN201910466234.8A priority Critical patent/CN110210556B/en
Publication of CN110210556A publication Critical patent/CN110210556A/en
Application granted granted Critical
Publication of CN110210556B publication Critical patent/CN110210556B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/30Scenes; Scene-specific elements in albums, collections or shared content, e.g. social network photos or video
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/103Static body considered as a whole, e.g. static pedestrian or occupant recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Human Computer Interaction (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a pedestrian re-identification data generation method based on an adaptive migration network, which converts the image style in a source data set by using a countermeasure generation network (general adaptive Networks) to generate an image consistent with the style of the source data set, can effectively reduce the field interval between different data sets, and finally enables a model trained on a certain data set to have strong generalization capability on other data sets.

Description

Pedestrian re-identification data generation method
Technical Field
The invention relates to the technical field of pedestrian re-identification, in particular to a pedestrian re-identification data generation method.
Background
The pedestrian re-identification technology mainly aims to identify images with the same identity as a target person in a large-scale image library acquired by different monitoring cameras at different places. The pedestrian re-identification has very wide application in daily life, such as intelligent security, man-machine interaction, content-based retrieval, behavior analysis and the like. The pedestrian re-identification technology based on deep learning utilizes the strong automatic learning abstract characteristic capability of a neural network in deep learning, so that the performance of a pedestrian re-identification algorithm is greatly improved. The pedestrian re-identification algorithm belongs to a supervised learning algorithm, namely, a large amount of data (pedestrian images) needs to be collected by human in advance to form a data set. Different pedestrians need to be labeled with different identity labels in the data set, then a recognition algorithm is designed, and supervised learning is carried out according to sample data in the training set. Therefore, the quality of the data set and the characteristics of the data set have a great influence on the final performance of the pedestrian re-identification algorithm.
Although the existing pedestrian re-identification methods based on deep learning have good performance on a single data set, the performance of the existing pedestrian re-identification methods is sharply reduced when the existing pedestrian re-identification methods are directly applied to other data sets, and the reason is that the existing pedestrian re-identification methods are poor in generalization performance. The pedestrian image is susceptible to various monitoring environmental factors and interference in the imaging process. In real life, pedestrian image appearances are usually photographed by different monitoring cameras at different angles, at different places, and at different times. This results in significant deviations in the imaging conditions, such as illumination, sharpness, and camera view, during the acquisition of the images in the different data sets. Images in one data set are usually shot by a plurality of cameras at a fixed angle at a specific time, so that a model trained based on a single data set cannot cope with complex shooting conditions in the real world.
Disclosure of Invention
The invention aims to provide a pedestrian re-identification data generation method, and the algorithm aims to generate a source data set image consistent with a target data set format by designing an adaptive migration network, so that the difference existing among data sets is reduced, and the performance of the pedestrian re-identification algorithm on the target data set is improved.
The purpose of the invention is realized by the following technical scheme:
a pedestrian re-identification data generation method comprising:
using the cycleGAN to construct a countermeasure generation network for transferring illumination change, a countermeasure generation network for transferring resolution change, a countermeasure generation network for transferring a photographing view angle and an integrated countermeasure generation network; constructing a sub-network weight fitting network by using a multilayer convolutional neural network and a full-connection network;
the countermeasure generating network of the migration illumination change, the countermeasure generating network of the migration resolution change and the countermeasure generating network of the migration photographing view all comprise a generator consisting of an encoder and a decoder and a discriminator; the integrated countermeasure generating network comprises a decoder and a discriminator; parameters of a decoder and a discriminator in the four confrontation generation networks are mutually shared, and the four discriminators are all used in a training stage and are matched with the output of the corresponding decoder to realize parameter updating;
for an image x from a source data set, respectively performing coding dimension reduction through encoders in the countermeasure generation network for migration illumination change, the countermeasure generation network for migration resolution change and the countermeasure generation network for migration photographing view;
generating the weight of each confrontation generation network according to the output of the three encoders through a sub-network weight fitting network;
after the weights of the three confrontation generation networks are multiplied by the output of the corresponding encoder, a new image with the illumination, the resolution and the photographing visual angle consistent with the target data set is generated through a decoder of the integrated confrontation generation network.
According to the technical scheme provided by the invention, the pedestrian re-identification data generation method based on the adaptive migration network converts the image style in the source data set by using the countermeasure generation network (general adaptive Networks) to generate the image with the style consistent with that of the target data set, so that the field interval between different data sets can be effectively reduced, and the re-identification rate across the data sets is improved. Finally, the model trained on a certain data set has strong generalization capability on other data sets.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings needed to be used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on the drawings without creative efforts.
Fig. 1 is a schematic diagram of a pedestrian re-identification data generation method according to an embodiment of the present invention;
fig. 2 is a schematic diagram of a generation network for migration of illumination change countermeasures according to an embodiment of the present invention;
FIG. 3 is a schematic diagram of a migration resolution change countermeasure generation network provided by an embodiment of the present invention;
fig. 4 is a schematic diagram of a countermeasure generation network for migrating a photographing view according to an embodiment of the present invention;
FIG. 5 is a schematic diagram of a sub-network weight fitting network according to an embodiment of the present invention;
fig. 6 is a schematic diagram of an integrated countermeasure generation network according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention are clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments of the present invention without making any creative effort, shall fall within the protection scope of the present invention.
The invention provides a pedestrian re-identification data generation method, which is a pedestrian re-identification data generation method based on an introduction division and fusion strategy of an adaptive migration network and decoder parameter sharing, and can convert the image style in a source data set into the style of a target data set by introducing four countermeasure generation networks and a sub-network weight fitting network in the training process of a pedestrian re-identification algorithm spanning multiple data sets, thereby adaptively reducing the influence of the characteristics of the data set such as illumination, pixels, shooting positions and the like on the model generalization performance. The pedestrian re-identification data generation method has wide application prospect, and many applications in the intelligent society can not leave pedestrian re-identification algorithms (such as intelligent security, pedestrian path identification and the like), so that the method is a necessary way for human beings to realize strong artificial intelligence.
Fig. 1 is a schematic diagram of a method for generating pedestrian re-identification data according to an embodiment of the present invention, which mainly includes:
1. using the cycleGAN to construct a countermeasure generation network for transferring illumination change, a countermeasure generation network for transferring resolution change, a countermeasure generation network for transferring a photographing view angle and an integrated countermeasure generation network; a sub-network weight-fitting network is constructed using a multi-layer convolutional neural network and a fully-connected network.
As will be appreciated by those skilled in the art, CycleGAN is a basic algorithmic model that can transform image content from one domain to another, such as style migration, etc.
2. The countermeasure generating network of the migration illumination change, the countermeasure generating network of the migration resolution change and the countermeasure generating network of the migration photographing view all comprise a generator consisting of an encoder and a decoder and a discriminator; the integrated countermeasure generating network comprises a decoder and a discriminator; the parameters of the decoder and the discriminator in the four confrontation generation networks are mutually shared, and the four discriminators are all used in the training stage and are matched with the output of the corresponding decoder to realize parameter updating.
In the embodiment of the invention, the encoders in the countermeasure generation networks are mainly used for extracting the relevant features of the input image, and the features extracted by the encoders in different countermeasure generation networks have different emphasis, namely different emphasis on illumination, resolution and visual angle.
3. And for an image x from a source data set, respectively carrying out coding dimension reduction through encoders in the countermeasure generation network for the migration illumination change, the countermeasure generation network for the migration resolution change and the countermeasure generation network for the migration photographing view.
4. And generating the weight of each confrontation generation network according to the output of the three encoders through a sub-network weight fitting network.
5. After the weights of the three confrontation generating networks are multiplied by the output of the corresponding encoder, a new image with the illumination, resolution and photographing visual angle consistent with the target data set is generated by a decoder of the integrated confrontation generating network.
For ease of understanding, each network is described in detail below.
Firstly, shifting the confrontation of illumination change to generate a network.
As shown in fig. 2, the structure of the generation network for migration of the countermeasure against the illumination change mainly includes: a generator and a discriminator, the generator mainly comprises an Encoder (Encoder) and a Decoder (Decode).
For each image x from the source dataset T, the purpose of migrating the antagonistic generation network of illumination changes is to generate a new image based on x that is consistent in illumination characteristics with the target dataset S. The main working process is as follows:
1. for an image x from a source data set, an encoder in a network is generated by shifting the confrontation of illumination change for encoding and dimension reduction, and then a decoder decodes an encoding result to generate a new image with illumination characteristics consistent with a target data set.
2. The new image is mixed with the image in the target data set and input to the discriminator, which discriminates the image, i.e. whether the image is the new image generated or the image in the target data set.
3. And updating parameters of the encoder according to the judgment result of the discriminator.
The countermeasure of the migration illumination variation generates a loss function L of encoder parameter updates in the network total 1 Comprises the following steps:
L total 1 =L gan1 L ill (G,H)
wherein eta is 1 Adjustments may be made during the experiment for over-parameters.
L gan As a basis loss function, expressed as:
L gan =L adv1 L cyc2 L ide
wherein L is adv For ensuring that the distribution of the new image generated is consistent with the distribution of the target data set, L cyc For reversing in Cyc | eGANSwitching, transferring the new image generated back to the original image, L ide For ensuring that the overall colour of the two images remains the same, lambda 1 And λ 2 Adjustments may be made during the experiment for over-parameters.
L ill (G, H) a lighting situation for focusing the countering generation network of the migration lighting changes in the migration image, expressed as:
L ill (G,H)=E x~p(x) [||H(G(x))-H(x)|| 1 ]
wherein G (x) represents the new image generated, E represents the mathematical expectation, p (x) is the data distribution function of the image in the source dataset, H (-) represents the function of extracting the illumination insensitive feature, | | 1 Representing a norm.
And secondly, moving the countermeasure generation network with the changed resolution.
As shown in fig. 3, the structure of the antagonistic generating network for the migration resolution change is also implemented by using CycleGAN, and the structure and workflow are the same as those of the antagonistic generating network for the migration illumination change described above. The differences are that: the opposition to the change in migration resolution generates a new image resolution generated by the network that is consistent with the target data set and that has a different loss function as the encoder parameters are updated.
The countermeasure of the transition resolution change generates a loss function L of encoder parameter updates in the network total 2 Comprises the following steps:
L total 2 =L gan2 L res (G,I)
wherein eta is 2 For the purpose of over-parameters, adjustments can be made during the experiment, L gan The basic loss function is the same as that described above and will not be described again.
L res (G, I) a countermeasure generation network for causing the transition resolution to vary focuses on the resolution in the transition image, which is expressed as:
Figure BDA0002077039960000051
wherein G (x) represents the new image generated, p (x) is the data distribution function of the image in the source data set, I (x) represents the function of extracting the pixel insensitive feature,
Figure BDA0002077039960000052
representing the square of the two norms.
And thirdly, shifting the confrontation of the photographing visual angles to generate a network.
As shown in fig. 4, the structure of the generation network for the countermeasure of the migration photographing perspective is also implemented by using CycleGAN, and the structure and the workflow are the same as those of the countermeasure generation network for the migration illumination change described above. The difference lies in that: the countermeasure of the migration camera view generates a new image camera view generated by the network that is consistent with the target data set and that has a different loss function when the encoder parameters are updated.
Since the migration camera angle does not need to have other constraints, the countermeasure of the migration camera angle generates a loss function L of the encoder parameter update in the network total 3 Comprises the following steps:
L total 3 =L gan
wherein L is gan The basic loss function is the same as that described above and will not be described again.
And fourthly, fitting a network by the sub-network weight.
In the embodiment of the present invention, the sub-network weight fitting network is implemented by using a multilayer convolutional neural network and a full-connection network, and the structure of the sub-network weight fitting network is shown in fig. 5.
The sub-network weight fitting network needs to be trained so that the sub-network weight fitting network can fit the weight of the network generated by each countermeasure;
the training process is as follows:
1) the outputs of the encoders within the three pairs generation network are stacked together as the input to the sub-network weight fitting network.
2) And normalizing the reciprocal of the loss functions of the three confrontation generation networks to be used as a real label.
3) Training by adopting the minimum mean square error as a loss function of a sub-network weight fitting network;
in the testing stage, the weight of each confrontation generation network is directly generated according to the output of the coder in the three confrontation generation networks:
Figure BDA0002077039960000061
wherein,
Figure BDA0002077039960000062
generating a corresponding weight value of the network for the countermeasure of the migration illumination change,
Figure BDA0002077039960000063
generating corresponding weights of the network for the countermeasure of the migration resolution change,
Figure BDA0002077039960000064
and generating a weight corresponding to the network for the confrontation of the migration photographing visual angles.
And fifthly, integrating the countermeasure generation network.
As shown in fig. 6, the integrated countermeasure generation network includes: a decoder and a discriminator, wherein the network structure of the decoder and the discriminator is the same as the three countermeasure generation networks (countermeasure generation network for migrating illumination variation, countermeasure generation network for migrating resolution variation, countermeasure generation network for migrating photo-taking view angle) introduced earlier.
1. And (5) a training stage.
The integrated countermeasure generation network normalizes the reciprocal of the loss function of the three countermeasure generation networks to obtain the weights of the three countermeasure generation networks, and then multiplies the outputs of the encoders of the three countermeasure generation networks by the respective weights to be used as the input z of the integrated countermeasure generation network in the training stage x Expressed as:
Figure BDA0002077039960000065
wherein,
Figure BDA0002077039960000066
the outputs of the encoders of the countermeasure generation network for the migration illumination change, the countermeasure generation network for the migration resolution change, and the countermeasure generation network for the migration photographing view angle are respectively corresponded to;
Figure BDA0002077039960000067
and
Figure BDA0002077039960000068
are the same, can be, for example, 64 x 768, then z x ∈R 64×64×768
By decoder pair z x Decoding to obtain a new image, mixing it with the image of target data, distinguishing the image by discriminator, updating the parameters of decoder and discriminator according to the result of discriminator, and obtaining relative loss function L total 4 Comprises the following steps:
Figure BDA0002077039960000069
wherein eta is 3 For hyper-parameters, can be varied automatically during the course of the experiment, L gan The basic loss function is the same as that described above and will not be described again.
Figure BDA00020770399600000610
f (-) is the Jesen-Shannon divergence between the two distributions,
Figure BDA0002077039960000071
and
Figure BDA0002077039960000072
respectively, are the coding features multiplied by the normalized parameters (i.e. correspond to the above-mentioned ones)
Figure BDA0002077039960000073
)。
2. Testing phase
In the testing stage, the weights of the three confrontation generating networks used by the integrated confrontation generating network are provided by the network weight fitting network, and after the weights are multiplied by the outputs of the encoders corresponding to the three confrontation generating networks, a migrated image is output.
In the embodiment of the invention, the parameters of the decoder and the discriminator in the integrated countermeasure generating network are shared with the decoders of the previous three countermeasure generating networks, and the parameters are updated only in the integrated countermeasure network.
The above description is only for the preferred embodiment of the present invention, but the scope of the present invention is not limited thereto, and any changes or substitutions that can be easily conceived by those skilled in the art within the technical scope of the present invention are included in the scope of the present invention. Therefore, the protection scope of the present invention should be subject to the protection scope of the claims.

Claims (8)

1. A pedestrian re-identification data generation method, characterized by comprising:
constructing a countermeasure generation network for the change of the migration illumination, a countermeasure generation network for the change of the migration resolution, a countermeasure generation network for the migration photographing view and an integrated countermeasure generation network by using the cycleGAN; constructing a sub-network weight fitting network by using a multilayer convolutional neural network and a full-connection network;
the countermeasure generating network of the migration illumination change, the countermeasure generating network of the migration resolution change and the countermeasure generating network of the migration photographing view all comprise a generator consisting of an encoder and a decoder and a discriminator; the integrated countermeasure generation network comprises a decoder and a discriminator; parameters of decoders and discriminators in the four countermeasure generating networks are mutually shared, and the four discriminators are all used in a training stage and are matched with the output of corresponding decoders to realize parameter updating;
for an image x from a source data set, respectively performing coding dimension reduction through encoders in the countermeasure generation network for migration illumination change, the countermeasure generation network for migration resolution change and the countermeasure generation network for migration photographing view;
generating the weight of each confrontation generation network according to the output of the three encoders through a sub-network weight fitting network;
after the weights of the three confrontation generating networks are multiplied by the output of the corresponding encoder, a new image with the illumination, resolution and photographing visual angle consistent with the target data set is generated by a decoder of the integrated confrontation generating network.
2. The pedestrian re-identification data generation method according to claim 1, wherein the structures of the countermeasure generation network for migrating illumination variation, the countermeasure generation network for migrating resolution variation, and the countermeasure generation network for migrating photographing view angle are the same;
for an image x from a source data set, an encoder in a network is generated by shifting the confrontation of illumination change to perform encoding dimensionality reduction, and then a decoder decodes an encoding result to generate a new image with illumination characteristics consistent with a target data set; the new image is mixed with the image in the target data set and then input into a discriminator, and the image is distinguished by the discriminator, so that the parameters of the encoder are updated according to the distinguishing result of the discriminator;
the countermeasure generation network for the transition resolution change and the countermeasure generation network for the transition photographing view both work in the above manner, and the difference is that: the generated new images are respectively consistent with the target data set in the resolution and the photographing view, and the loss functions are different when the parameters of the encoder are updated.
3. The pedestrian re-identification data generation method of claim 1, wherein the countermeasure against the transitional illumination change generates a loss function L of the encoder parameter update in the network total 1 Comprises the following steps:
L total 1 =L gan1 L ill (G,H)
wherein eta 1 Is a hyperparameter, L gan Is a base loss function; l is ill (G, H) is used to makeCountering transitional lighting changes generates lighting conditions, L, in which the network focuses on the transitional image ill (G,H)=E x~p(x) [||H(G(x))-H(x)|| 1 ]G (x) represents a new generated image, E represents a mathematical expectation, p (x) is a data distribution function of the image in the source dataset, H (·) represents a function that extracts the illumination insensitive features, | | 1 Representing a norm.
4. The pedestrian re-identification data generation method according to claim 1, wherein the countermeasure against the change in the transition resolution generates a loss function L of the encoder parameter update in the network total 2 Comprises the following steps:
L total 2 =L gan2 L res (G,I)
wherein eta is 2 Is hyperparametric, L gan Is a base loss function; l is res (G, I) a countermeasure generation network for causing the transition resolution to vary focuses on the resolution in the transition image,
Figure FDA0002077039950000021
g (x) represents the new image generated, E represents the mathematical expectation, p (x) is the data distribution function of the source data set image, I (x) represents the function of extracting the pixel insensitive features,
Figure FDA0002077039950000022
representing the square of the two norms.
5. The pedestrian re-identification data generation method of claim 1, wherein the countermeasure of the migration photo-angle generates a loss function L of the encoder parameter update in the network total 3 Comprises the following steps:
L total 3 =L gan
wherein L is gan Is the basis loss function.
6. The pedestrian re-identification data generation method according to claim 1,
the sub-network weight fitting network needs to be trained so that the sub-network weight fitting network can fit the weight of the network generated by each countermeasure;
the training process is as follows: stacking outputs of encoders in the three confrontation generation networks to serve as inputs of a sub-network weight fitting network, and normalizing reciprocal numbers of loss functions of the three confrontation generation networks to serve as real labels; then, training by adopting the minimum mean square error as a loss function of a sub-network weight fitting network;
in the testing stage, the weight of each confrontation generation network is directly generated according to the output of the coder in the three confrontation generation networks:
Figure FDA0002077039950000023
wherein,
Figure FDA0002077039950000024
generating a corresponding weight value of the network for the countermeasure of the migration illumination change,
Figure FDA0002077039950000025
generate corresponding network weights for the confrontation of the migration resolution change,
Figure FDA0002077039950000026
and generating a weight corresponding to the network for the confrontation of the migration photographing visual angles.
7. The pedestrian re-identification data generation method of claim 1, wherein in the training phase, the integrated countermeasure generation network normalizes reciprocal of loss functions of the three countermeasure generation networks to obtain weights of the three countermeasure generation networks, and then multiplies outputs of encoders of the three countermeasure generation networks by their respective weights as inputs z of the integrated countermeasure generation network in the training phase x Expressed as:
Figure FDA0002077039950000027
wherein,
Figure FDA0002077039950000031
the outputs of the encoders respectively correspond to a countermeasure generation network for migration illumination change, a countermeasure generation network for migration resolution change, and a countermeasure generation network for migration photographing view;
by decoder pair z x Decoding to obtain a new image, mixing it with the image of target data, distinguishing the image by discriminator, updating the parameters of decoder and discriminator according to the result of discriminator, and obtaining corresponding loss function L total 4 Comprises the following steps:
Figure FDA0002077039950000032
wherein eta is 3 Is a hyperparameter, L gan As a basis loss function;
Figure FDA0002077039950000033
Figure FDA0002077039950000034
f (-) is the Jesen-Shannon divergence between the two distributions,
Figure FDA0002077039950000035
and
Figure FDA0002077039950000036
respectively the coding characteristics multiplied by the normalized parameters;
after the integrated countermeasure generating network finishes the parameter updating of the decoder and the discriminator, the updated related parameters are shared to the countermeasure generating network of the migration illumination change, the countermeasure generating network of the migration resolution change and the countermeasure generating network of the migration photographing view;
in the testing stage, the weights of the three confrontation generating networks used by the integrated confrontation generating network are provided by a network weight fitting network.
8. The pedestrian re-identification data generation method according to claim 3, 4, 5 or 7, wherein the basic loss function L gan Comprises the following steps:
L gan =L adv1 L cyc2 L ide
wherein L is adv For ensuring that the distribution of the new image generated is consistent with the distribution of the target data set, L cyc For reverse conversion in CycleGAN, with new images being generated and migrated back to the original, L ide For ensuring that the overall colour of the two images remains the same, lambda 1 And λ 2 Is a hyper-parameter.
CN201910466234.8A 2019-05-29 2019-05-29 Pedestrian re-identification data generation method Active CN110210556B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910466234.8A CN110210556B (en) 2019-05-29 2019-05-29 Pedestrian re-identification data generation method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910466234.8A CN110210556B (en) 2019-05-29 2019-05-29 Pedestrian re-identification data generation method

Publications (2)

Publication Number Publication Date
CN110210556A CN110210556A (en) 2019-09-06
CN110210556B true CN110210556B (en) 2022-09-06

Family

ID=67789683

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910466234.8A Active CN110210556B (en) 2019-05-29 2019-05-29 Pedestrian re-identification data generation method

Country Status (1)

Country Link
CN (1) CN110210556B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111382675B (en) * 2020-02-24 2024-02-27 江苏大学 Generation countermeasure network system for pedestrian recognition data set enhancement training
CN112116104B (en) * 2020-09-17 2024-06-18 京东科技控股股份有限公司 Method, device, medium and electronic equipment for automatically integrating machine learning

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108288072A (en) * 2018-01-26 2018-07-17 深圳市唯特视科技有限公司 A kind of facial expression synthetic method based on generation confrontation network
CN109376769A (en) * 2018-09-21 2019-02-22 广东技术师范学院 The information transfer method of multitask classification is used for based on production confrontation neural network
CN109447906A (en) * 2018-11-08 2019-03-08 北京印刷学院 A kind of picture synthetic method based on generation confrontation network
CN109815893A (en) * 2019-01-23 2019-05-28 中山大学 The normalized method in colorized face images illumination domain of confrontation network is generated based on circulation

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6435581B2 (en) * 2015-01-29 2018-12-12 パナソニックIpマネジメント株式会社 Transfer learning device, transfer learning system, transfer learning method and program

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108288072A (en) * 2018-01-26 2018-07-17 深圳市唯特视科技有限公司 A kind of facial expression synthetic method based on generation confrontation network
CN109376769A (en) * 2018-09-21 2019-02-22 广东技术师范学院 The information transfer method of multitask classification is used for based on production confrontation neural network
CN109447906A (en) * 2018-11-08 2019-03-08 北京印刷学院 A kind of picture synthetic method based on generation confrontation network
CN109815893A (en) * 2019-01-23 2019-05-28 中山大学 The normalized method in colorized face images illumination domain of confrontation network is generated based on circulation

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
Pose Transferrable Person Re-identification;Jinxian Liu 等;《2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition》;20181217;第4099-4108页 *
Visual Domain Adaptation: A survey of recent advances;Vishal M Patel 等;《IEEE Signal Processing Magazine》;20150402;第32卷(第3期);第53-69页 *
基于深度自编码的SAR图像配准方法研究;宁梦丹;《中国优秀硕士论文全文数据库信息科技辑》;20190215;第2019年卷(第2期);I136-1247 *
基于生成对抗网络的图片风格迁移;许哲豪 等;《软件导刊》;20181231;第17卷(第6期);第207-209+212+228页 *

Also Published As

Publication number Publication date
CN110210556A (en) 2019-09-06

Similar Documents

Publication Publication Date Title
KR101803471B1 (en) Deep learning system and learning method using of convolutional neural network based image patterning
Chen et al. Saliency detection via the improved hierarchical principal component analysis method
CN110191299B (en) Multi-frame interpolation method based on convolutional neural network
Lu et al. Evolving block-based convolutional neural network for hyperspectral image classification
KR102138657B1 (en) Apparatus and method for robust face recognition via hierarchical collaborative representation
CN111950649A (en) Attention mechanism and capsule network-based low-illumination image classification method
CN110210556B (en) Pedestrian re-identification data generation method
Wozniak et al. A multiscale image compressor with rbfnn and discrete wavelet decomposition
CN112560865B (en) Semantic segmentation method for point cloud under outdoor large scene
CN113743544A (en) Cross-modal neural network construction method, pedestrian retrieval method and system
CN103268484A (en) Design method of classifier for high-precision face recognitio
CN117743946B (en) Signal type identification method and system based on fusion characteristic and group convolution ViT network
Chen et al. Cumulative attribute space regression for head pose estimation and color constancy
CN117218351A (en) Three-dimensional point cloud semantic segmentation method based on local and global context awareness
CN114170426A (en) Algorithm model for classifying rare tumor category small samples based on cost sensitivity
Casagrande et al. Abnormal motion analysis for tracking-based approaches using region-based method with mobile grid
CN113378598B (en) Dynamic bar code detection method based on deep learning
Palomo et al. Image compression and video segmentation using hierarchical self-organization
Gomes et al. A Deep Learning Approach for Reconstruction of Color Images in Different Lighting Conditions Based on Autoencoder Technique
CN113205175A (en) Multi-layer attribute network representation learning method based on mutual information maximization
Li et al. Automated deep learning system for power line inspection image analysis and processing: Architecture and design issues
Psaltis et al. Deep 3d flow features for human action recognition
Zhang et al. GARBAGE CLASSIFICATION BASED ON A CASCADE NEURAL NETWORK.
Mousa et al. Identification the modulation type in cognitive radio network based on Alexnet architecture
Faye et al. Context Normalization Layer with Applications

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant