CN110210556B - Pedestrian re-identification data generation method - Google Patents
Pedestrian re-identification data generation method Download PDFInfo
- Publication number
- CN110210556B CN110210556B CN201910466234.8A CN201910466234A CN110210556B CN 110210556 B CN110210556 B CN 110210556B CN 201910466234 A CN201910466234 A CN 201910466234A CN 110210556 B CN110210556 B CN 110210556B
- Authority
- CN
- China
- Prior art keywords
- network
- countermeasure
- migration
- image
- generation
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 30
- 230000005012 migration Effects 0.000 claims abstract description 56
- 238000013508 migration Methods 0.000 claims abstract description 56
- 230000008859 change Effects 0.000 claims description 42
- 238000005286 illumination Methods 0.000 claims description 37
- 230000006870 function Effects 0.000 claims description 30
- 238000012549 training Methods 0.000 claims description 13
- 230000007704 transition Effects 0.000 claims description 8
- 230000000007 visual effect Effects 0.000 claims description 7
- 238000009826 distribution Methods 0.000 claims description 6
- 230000008569 process Effects 0.000 claims description 5
- 230000009467 reduction Effects 0.000 claims description 5
- 238000012360 testing method Methods 0.000 claims description 5
- 238000013527 convolutional neural network Methods 0.000 claims description 4
- 238000005315 distribution function Methods 0.000 claims description 4
- 238000006243 chemical reaction Methods 0.000 claims 1
- 239000000284 extract Substances 0.000 claims 1
- 230000003044 adaptive effect Effects 0.000 abstract description 6
- 238000010586 diagram Methods 0.000 description 7
- 238000002474 experimental method Methods 0.000 description 4
- 230000003042 antagnostic effect Effects 0.000 description 3
- 238000013135 deep learning Methods 0.000 description 3
- 238000012544 monitoring process Methods 0.000 description 3
- 241000282414 Homo sapiens Species 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000003384 imaging method Methods 0.000 description 2
- 238000013473 artificial intelligence Methods 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/30—Scenes; Scene-specific elements in albums, collections or shared content, e.g. social network photos or video
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/103—Static body considered as a whole, e.g. static pedestrian or occupant recognition
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- Molecular Biology (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Human Computer Interaction (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a pedestrian re-identification data generation method based on an adaptive migration network, which converts the image style in a source data set by using a countermeasure generation network (general adaptive Networks) to generate an image consistent with the style of the source data set, can effectively reduce the field interval between different data sets, and finally enables a model trained on a certain data set to have strong generalization capability on other data sets.
Description
Technical Field
The invention relates to the technical field of pedestrian re-identification, in particular to a pedestrian re-identification data generation method.
Background
The pedestrian re-identification technology mainly aims to identify images with the same identity as a target person in a large-scale image library acquired by different monitoring cameras at different places. The pedestrian re-identification has very wide application in daily life, such as intelligent security, man-machine interaction, content-based retrieval, behavior analysis and the like. The pedestrian re-identification technology based on deep learning utilizes the strong automatic learning abstract characteristic capability of a neural network in deep learning, so that the performance of a pedestrian re-identification algorithm is greatly improved. The pedestrian re-identification algorithm belongs to a supervised learning algorithm, namely, a large amount of data (pedestrian images) needs to be collected by human in advance to form a data set. Different pedestrians need to be labeled with different identity labels in the data set, then a recognition algorithm is designed, and supervised learning is carried out according to sample data in the training set. Therefore, the quality of the data set and the characteristics of the data set have a great influence on the final performance of the pedestrian re-identification algorithm.
Although the existing pedestrian re-identification methods based on deep learning have good performance on a single data set, the performance of the existing pedestrian re-identification methods is sharply reduced when the existing pedestrian re-identification methods are directly applied to other data sets, and the reason is that the existing pedestrian re-identification methods are poor in generalization performance. The pedestrian image is susceptible to various monitoring environmental factors and interference in the imaging process. In real life, pedestrian image appearances are usually photographed by different monitoring cameras at different angles, at different places, and at different times. This results in significant deviations in the imaging conditions, such as illumination, sharpness, and camera view, during the acquisition of the images in the different data sets. Images in one data set are usually shot by a plurality of cameras at a fixed angle at a specific time, so that a model trained based on a single data set cannot cope with complex shooting conditions in the real world.
Disclosure of Invention
The invention aims to provide a pedestrian re-identification data generation method, and the algorithm aims to generate a source data set image consistent with a target data set format by designing an adaptive migration network, so that the difference existing among data sets is reduced, and the performance of the pedestrian re-identification algorithm on the target data set is improved.
The purpose of the invention is realized by the following technical scheme:
a pedestrian re-identification data generation method comprising:
using the cycleGAN to construct a countermeasure generation network for transferring illumination change, a countermeasure generation network for transferring resolution change, a countermeasure generation network for transferring a photographing view angle and an integrated countermeasure generation network; constructing a sub-network weight fitting network by using a multilayer convolutional neural network and a full-connection network;
the countermeasure generating network of the migration illumination change, the countermeasure generating network of the migration resolution change and the countermeasure generating network of the migration photographing view all comprise a generator consisting of an encoder and a decoder and a discriminator; the integrated countermeasure generating network comprises a decoder and a discriminator; parameters of a decoder and a discriminator in the four confrontation generation networks are mutually shared, and the four discriminators are all used in a training stage and are matched with the output of the corresponding decoder to realize parameter updating;
for an image x from a source data set, respectively performing coding dimension reduction through encoders in the countermeasure generation network for migration illumination change, the countermeasure generation network for migration resolution change and the countermeasure generation network for migration photographing view;
generating the weight of each confrontation generation network according to the output of the three encoders through a sub-network weight fitting network;
after the weights of the three confrontation generation networks are multiplied by the output of the corresponding encoder, a new image with the illumination, the resolution and the photographing visual angle consistent with the target data set is generated through a decoder of the integrated confrontation generation network.
According to the technical scheme provided by the invention, the pedestrian re-identification data generation method based on the adaptive migration network converts the image style in the source data set by using the countermeasure generation network (general adaptive Networks) to generate the image with the style consistent with that of the target data set, so that the field interval between different data sets can be effectively reduced, and the re-identification rate across the data sets is improved. Finally, the model trained on a certain data set has strong generalization capability on other data sets.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings needed to be used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on the drawings without creative efforts.
Fig. 1 is a schematic diagram of a pedestrian re-identification data generation method according to an embodiment of the present invention;
fig. 2 is a schematic diagram of a generation network for migration of illumination change countermeasures according to an embodiment of the present invention;
FIG. 3 is a schematic diagram of a migration resolution change countermeasure generation network provided by an embodiment of the present invention;
fig. 4 is a schematic diagram of a countermeasure generation network for migrating a photographing view according to an embodiment of the present invention;
FIG. 5 is a schematic diagram of a sub-network weight fitting network according to an embodiment of the present invention;
fig. 6 is a schematic diagram of an integrated countermeasure generation network according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention are clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments of the present invention without making any creative effort, shall fall within the protection scope of the present invention.
The invention provides a pedestrian re-identification data generation method, which is a pedestrian re-identification data generation method based on an introduction division and fusion strategy of an adaptive migration network and decoder parameter sharing, and can convert the image style in a source data set into the style of a target data set by introducing four countermeasure generation networks and a sub-network weight fitting network in the training process of a pedestrian re-identification algorithm spanning multiple data sets, thereby adaptively reducing the influence of the characteristics of the data set such as illumination, pixels, shooting positions and the like on the model generalization performance. The pedestrian re-identification data generation method has wide application prospect, and many applications in the intelligent society can not leave pedestrian re-identification algorithms (such as intelligent security, pedestrian path identification and the like), so that the method is a necessary way for human beings to realize strong artificial intelligence.
Fig. 1 is a schematic diagram of a method for generating pedestrian re-identification data according to an embodiment of the present invention, which mainly includes:
1. using the cycleGAN to construct a countermeasure generation network for transferring illumination change, a countermeasure generation network for transferring resolution change, a countermeasure generation network for transferring a photographing view angle and an integrated countermeasure generation network; a sub-network weight-fitting network is constructed using a multi-layer convolutional neural network and a fully-connected network.
As will be appreciated by those skilled in the art, CycleGAN is a basic algorithmic model that can transform image content from one domain to another, such as style migration, etc.
2. The countermeasure generating network of the migration illumination change, the countermeasure generating network of the migration resolution change and the countermeasure generating network of the migration photographing view all comprise a generator consisting of an encoder and a decoder and a discriminator; the integrated countermeasure generating network comprises a decoder and a discriminator; the parameters of the decoder and the discriminator in the four confrontation generation networks are mutually shared, and the four discriminators are all used in the training stage and are matched with the output of the corresponding decoder to realize parameter updating.
In the embodiment of the invention, the encoders in the countermeasure generation networks are mainly used for extracting the relevant features of the input image, and the features extracted by the encoders in different countermeasure generation networks have different emphasis, namely different emphasis on illumination, resolution and visual angle.
3. And for an image x from a source data set, respectively carrying out coding dimension reduction through encoders in the countermeasure generation network for the migration illumination change, the countermeasure generation network for the migration resolution change and the countermeasure generation network for the migration photographing view.
4. And generating the weight of each confrontation generation network according to the output of the three encoders through a sub-network weight fitting network.
5. After the weights of the three confrontation generating networks are multiplied by the output of the corresponding encoder, a new image with the illumination, resolution and photographing visual angle consistent with the target data set is generated by a decoder of the integrated confrontation generating network.
For ease of understanding, each network is described in detail below.
Firstly, shifting the confrontation of illumination change to generate a network.
As shown in fig. 2, the structure of the generation network for migration of the countermeasure against the illumination change mainly includes: a generator and a discriminator, the generator mainly comprises an Encoder (Encoder) and a Decoder (Decode).
For each image x from the source dataset T, the purpose of migrating the antagonistic generation network of illumination changes is to generate a new image based on x that is consistent in illumination characteristics with the target dataset S. The main working process is as follows:
1. for an image x from a source data set, an encoder in a network is generated by shifting the confrontation of illumination change for encoding and dimension reduction, and then a decoder decodes an encoding result to generate a new image with illumination characteristics consistent with a target data set.
2. The new image is mixed with the image in the target data set and input to the discriminator, which discriminates the image, i.e. whether the image is the new image generated or the image in the target data set.
3. And updating parameters of the encoder according to the judgment result of the discriminator.
The countermeasure of the migration illumination variation generates a loss function L of encoder parameter updates in the network total 1 Comprises the following steps:
L total 1 =L gan +η 1 L ill (G,H)
wherein eta is 1 Adjustments may be made during the experiment for over-parameters.
L gan As a basis loss function, expressed as:
L gan =L adv +λ 1 L cyc +λ 2 L ide
wherein L is adv For ensuring that the distribution of the new image generated is consistent with the distribution of the target data set, L cyc For reversing in Cyc | eGANSwitching, transferring the new image generated back to the original image, L ide For ensuring that the overall colour of the two images remains the same, lambda 1 And λ 2 Adjustments may be made during the experiment for over-parameters.
L ill (G, H) a lighting situation for focusing the countering generation network of the migration lighting changes in the migration image, expressed as:
L ill (G,H)=E x~p(x) [||H(G(x))-H(x)|| 1 ]
wherein G (x) represents the new image generated, E represents the mathematical expectation, p (x) is the data distribution function of the image in the source dataset, H (-) represents the function of extracting the illumination insensitive feature, | | 1 Representing a norm.
And secondly, moving the countermeasure generation network with the changed resolution.
As shown in fig. 3, the structure of the antagonistic generating network for the migration resolution change is also implemented by using CycleGAN, and the structure and workflow are the same as those of the antagonistic generating network for the migration illumination change described above. The differences are that: the opposition to the change in migration resolution generates a new image resolution generated by the network that is consistent with the target data set and that has a different loss function as the encoder parameters are updated.
The countermeasure of the transition resolution change generates a loss function L of encoder parameter updates in the network total 2 Comprises the following steps:
L total 2 =L gan +η 2 L res (G,I)
wherein eta is 2 For the purpose of over-parameters, adjustments can be made during the experiment, L gan The basic loss function is the same as that described above and will not be described again.
L res (G, I) a countermeasure generation network for causing the transition resolution to vary focuses on the resolution in the transition image, which is expressed as:
wherein G (x) represents the new image generated, p (x) is the data distribution function of the image in the source data set, I (x) represents the function of extracting the pixel insensitive feature,representing the square of the two norms.
And thirdly, shifting the confrontation of the photographing visual angles to generate a network.
As shown in fig. 4, the structure of the generation network for the countermeasure of the migration photographing perspective is also implemented by using CycleGAN, and the structure and the workflow are the same as those of the countermeasure generation network for the migration illumination change described above. The difference lies in that: the countermeasure of the migration camera view generates a new image camera view generated by the network that is consistent with the target data set and that has a different loss function when the encoder parameters are updated.
Since the migration camera angle does not need to have other constraints, the countermeasure of the migration camera angle generates a loss function L of the encoder parameter update in the network total 3 Comprises the following steps:
L total 3 =L gan
wherein L is gan The basic loss function is the same as that described above and will not be described again.
And fourthly, fitting a network by the sub-network weight.
In the embodiment of the present invention, the sub-network weight fitting network is implemented by using a multilayer convolutional neural network and a full-connection network, and the structure of the sub-network weight fitting network is shown in fig. 5.
The sub-network weight fitting network needs to be trained so that the sub-network weight fitting network can fit the weight of the network generated by each countermeasure;
the training process is as follows:
1) the outputs of the encoders within the three pairs generation network are stacked together as the input to the sub-network weight fitting network.
2) And normalizing the reciprocal of the loss functions of the three confrontation generation networks to be used as a real label.
3) Training by adopting the minimum mean square error as a loss function of a sub-network weight fitting network;
in the testing stage, the weight of each confrontation generation network is directly generated according to the output of the coder in the three confrontation generation networks:
wherein,generating a corresponding weight value of the network for the countermeasure of the migration illumination change,generating corresponding weights of the network for the countermeasure of the migration resolution change,and generating a weight corresponding to the network for the confrontation of the migration photographing visual angles.
And fifthly, integrating the countermeasure generation network.
As shown in fig. 6, the integrated countermeasure generation network includes: a decoder and a discriminator, wherein the network structure of the decoder and the discriminator is the same as the three countermeasure generation networks (countermeasure generation network for migrating illumination variation, countermeasure generation network for migrating resolution variation, countermeasure generation network for migrating photo-taking view angle) introduced earlier.
1. And (5) a training stage.
The integrated countermeasure generation network normalizes the reciprocal of the loss function of the three countermeasure generation networks to obtain the weights of the three countermeasure generation networks, and then multiplies the outputs of the encoders of the three countermeasure generation networks by the respective weights to be used as the input z of the integrated countermeasure generation network in the training stage x Expressed as:
wherein,the outputs of the encoders of the countermeasure generation network for the migration illumination change, the countermeasure generation network for the migration resolution change, and the countermeasure generation network for the migration photographing view angle are respectively corresponded to;andare the same, can be, for example, 64 x 768, then z x ∈R 64×64×768 。
By decoder pair z x Decoding to obtain a new image, mixing it with the image of target data, distinguishing the image by discriminator, updating the parameters of decoder and discriminator according to the result of discriminator, and obtaining relative loss function L total 4 Comprises the following steps:
wherein eta is 3 For hyper-parameters, can be varied automatically during the course of the experiment, L gan The basic loss function is the same as that described above and will not be described again.f (-) is the Jesen-Shannon divergence between the two distributions,andrespectively, are the coding features multiplied by the normalized parameters (i.e. correspond to the above-mentioned ones))。
2. Testing phase
In the testing stage, the weights of the three confrontation generating networks used by the integrated confrontation generating network are provided by the network weight fitting network, and after the weights are multiplied by the outputs of the encoders corresponding to the three confrontation generating networks, a migrated image is output.
In the embodiment of the invention, the parameters of the decoder and the discriminator in the integrated countermeasure generating network are shared with the decoders of the previous three countermeasure generating networks, and the parameters are updated only in the integrated countermeasure network.
The above description is only for the preferred embodiment of the present invention, but the scope of the present invention is not limited thereto, and any changes or substitutions that can be easily conceived by those skilled in the art within the technical scope of the present invention are included in the scope of the present invention. Therefore, the protection scope of the present invention should be subject to the protection scope of the claims.
Claims (8)
1. A pedestrian re-identification data generation method, characterized by comprising:
constructing a countermeasure generation network for the change of the migration illumination, a countermeasure generation network for the change of the migration resolution, a countermeasure generation network for the migration photographing view and an integrated countermeasure generation network by using the cycleGAN; constructing a sub-network weight fitting network by using a multilayer convolutional neural network and a full-connection network;
the countermeasure generating network of the migration illumination change, the countermeasure generating network of the migration resolution change and the countermeasure generating network of the migration photographing view all comprise a generator consisting of an encoder and a decoder and a discriminator; the integrated countermeasure generation network comprises a decoder and a discriminator; parameters of decoders and discriminators in the four countermeasure generating networks are mutually shared, and the four discriminators are all used in a training stage and are matched with the output of corresponding decoders to realize parameter updating;
for an image x from a source data set, respectively performing coding dimension reduction through encoders in the countermeasure generation network for migration illumination change, the countermeasure generation network for migration resolution change and the countermeasure generation network for migration photographing view;
generating the weight of each confrontation generation network according to the output of the three encoders through a sub-network weight fitting network;
after the weights of the three confrontation generating networks are multiplied by the output of the corresponding encoder, a new image with the illumination, resolution and photographing visual angle consistent with the target data set is generated by a decoder of the integrated confrontation generating network.
2. The pedestrian re-identification data generation method according to claim 1, wherein the structures of the countermeasure generation network for migrating illumination variation, the countermeasure generation network for migrating resolution variation, and the countermeasure generation network for migrating photographing view angle are the same;
for an image x from a source data set, an encoder in a network is generated by shifting the confrontation of illumination change to perform encoding dimensionality reduction, and then a decoder decodes an encoding result to generate a new image with illumination characteristics consistent with a target data set; the new image is mixed with the image in the target data set and then input into a discriminator, and the image is distinguished by the discriminator, so that the parameters of the encoder are updated according to the distinguishing result of the discriminator;
the countermeasure generation network for the transition resolution change and the countermeasure generation network for the transition photographing view both work in the above manner, and the difference is that: the generated new images are respectively consistent with the target data set in the resolution and the photographing view, and the loss functions are different when the parameters of the encoder are updated.
3. The pedestrian re-identification data generation method of claim 1, wherein the countermeasure against the transitional illumination change generates a loss function L of the encoder parameter update in the network total 1 Comprises the following steps:
L total 1 =L gan +η 1 L ill (G,H)
wherein eta 1 Is a hyperparameter, L gan Is a base loss function; l is ill (G, H) is used to makeCountering transitional lighting changes generates lighting conditions, L, in which the network focuses on the transitional image ill (G,H)=E x~p(x) [||H(G(x))-H(x)|| 1 ]G (x) represents a new generated image, E represents a mathematical expectation, p (x) is a data distribution function of the image in the source dataset, H (·) represents a function that extracts the illumination insensitive features, | | 1 Representing a norm.
4. The pedestrian re-identification data generation method according to claim 1, wherein the countermeasure against the change in the transition resolution generates a loss function L of the encoder parameter update in the network total 2 Comprises the following steps:
L total 2 =L gan +η 2 L res (G,I)
wherein eta is 2 Is hyperparametric, L gan Is a base loss function; l is res (G, I) a countermeasure generation network for causing the transition resolution to vary focuses on the resolution in the transition image,g (x) represents the new image generated, E represents the mathematical expectation, p (x) is the data distribution function of the source data set image, I (x) represents the function of extracting the pixel insensitive features,representing the square of the two norms.
5. The pedestrian re-identification data generation method of claim 1, wherein the countermeasure of the migration photo-angle generates a loss function L of the encoder parameter update in the network total 3 Comprises the following steps:
L total 3 =L gan
wherein L is gan Is the basis loss function.
6. The pedestrian re-identification data generation method according to claim 1,
the sub-network weight fitting network needs to be trained so that the sub-network weight fitting network can fit the weight of the network generated by each countermeasure;
the training process is as follows: stacking outputs of encoders in the three confrontation generation networks to serve as inputs of a sub-network weight fitting network, and normalizing reciprocal numbers of loss functions of the three confrontation generation networks to serve as real labels; then, training by adopting the minimum mean square error as a loss function of a sub-network weight fitting network;
in the testing stage, the weight of each confrontation generation network is directly generated according to the output of the coder in the three confrontation generation networks:
wherein,generating a corresponding weight value of the network for the countermeasure of the migration illumination change,generate corresponding network weights for the confrontation of the migration resolution change,and generating a weight corresponding to the network for the confrontation of the migration photographing visual angles.
7. The pedestrian re-identification data generation method of claim 1, wherein in the training phase, the integrated countermeasure generation network normalizes reciprocal of loss functions of the three countermeasure generation networks to obtain weights of the three countermeasure generation networks, and then multiplies outputs of encoders of the three countermeasure generation networks by their respective weights as inputs z of the integrated countermeasure generation network in the training phase x Expressed as:
wherein,the outputs of the encoders respectively correspond to a countermeasure generation network for migration illumination change, a countermeasure generation network for migration resolution change, and a countermeasure generation network for migration photographing view;
by decoder pair z x Decoding to obtain a new image, mixing it with the image of target data, distinguishing the image by discriminator, updating the parameters of decoder and discriminator according to the result of discriminator, and obtaining corresponding loss function L total 4 Comprises the following steps:
wherein eta is 3 Is a hyperparameter, L gan As a basis loss function; f (-) is the Jesen-Shannon divergence between the two distributions,andrespectively the coding characteristics multiplied by the normalized parameters;
after the integrated countermeasure generating network finishes the parameter updating of the decoder and the discriminator, the updated related parameters are shared to the countermeasure generating network of the migration illumination change, the countermeasure generating network of the migration resolution change and the countermeasure generating network of the migration photographing view;
in the testing stage, the weights of the three confrontation generating networks used by the integrated confrontation generating network are provided by a network weight fitting network.
8. The pedestrian re-identification data generation method according to claim 3, 4, 5 or 7, wherein the basic loss function L gan Comprises the following steps:
L gan =L adv +λ 1 L cyc +λ 2 L ide
wherein L is adv For ensuring that the distribution of the new image generated is consistent with the distribution of the target data set, L cyc For reverse conversion in CycleGAN, with new images being generated and migrated back to the original, L ide For ensuring that the overall colour of the two images remains the same, lambda 1 And λ 2 Is a hyper-parameter.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910466234.8A CN110210556B (en) | 2019-05-29 | 2019-05-29 | Pedestrian re-identification data generation method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910466234.8A CN110210556B (en) | 2019-05-29 | 2019-05-29 | Pedestrian re-identification data generation method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110210556A CN110210556A (en) | 2019-09-06 |
CN110210556B true CN110210556B (en) | 2022-09-06 |
Family
ID=67789683
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910466234.8A Active CN110210556B (en) | 2019-05-29 | 2019-05-29 | Pedestrian re-identification data generation method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110210556B (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111382675B (en) * | 2020-02-24 | 2024-02-27 | 江苏大学 | Generation countermeasure network system for pedestrian recognition data set enhancement training |
CN112116104B (en) * | 2020-09-17 | 2024-06-18 | 京东科技控股股份有限公司 | Method, device, medium and electronic equipment for automatically integrating machine learning |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108288072A (en) * | 2018-01-26 | 2018-07-17 | 深圳市唯特视科技有限公司 | A kind of facial expression synthetic method based on generation confrontation network |
CN109376769A (en) * | 2018-09-21 | 2019-02-22 | 广东技术师范学院 | The information transfer method of multitask classification is used for based on production confrontation neural network |
CN109447906A (en) * | 2018-11-08 | 2019-03-08 | 北京印刷学院 | A kind of picture synthetic method based on generation confrontation network |
CN109815893A (en) * | 2019-01-23 | 2019-05-28 | 中山大学 | The normalized method in colorized face images illumination domain of confrontation network is generated based on circulation |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6435581B2 (en) * | 2015-01-29 | 2018-12-12 | パナソニックIpマネジメント株式会社 | Transfer learning device, transfer learning system, transfer learning method and program |
-
2019
- 2019-05-29 CN CN201910466234.8A patent/CN110210556B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108288072A (en) * | 2018-01-26 | 2018-07-17 | 深圳市唯特视科技有限公司 | A kind of facial expression synthetic method based on generation confrontation network |
CN109376769A (en) * | 2018-09-21 | 2019-02-22 | 广东技术师范学院 | The information transfer method of multitask classification is used for based on production confrontation neural network |
CN109447906A (en) * | 2018-11-08 | 2019-03-08 | 北京印刷学院 | A kind of picture synthetic method based on generation confrontation network |
CN109815893A (en) * | 2019-01-23 | 2019-05-28 | 中山大学 | The normalized method in colorized face images illumination domain of confrontation network is generated based on circulation |
Non-Patent Citations (4)
Title |
---|
Pose Transferrable Person Re-identification;Jinxian Liu 等;《2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition》;20181217;第4099-4108页 * |
Visual Domain Adaptation: A survey of recent advances;Vishal M Patel 等;《IEEE Signal Processing Magazine》;20150402;第32卷(第3期);第53-69页 * |
基于深度自编码的SAR图像配准方法研究;宁梦丹;《中国优秀硕士论文全文数据库信息科技辑》;20190215;第2019年卷(第2期);I136-1247 * |
基于生成对抗网络的图片风格迁移;许哲豪 等;《软件导刊》;20181231;第17卷(第6期);第207-209+212+228页 * |
Also Published As
Publication number | Publication date |
---|---|
CN110210556A (en) | 2019-09-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR101803471B1 (en) | Deep learning system and learning method using of convolutional neural network based image patterning | |
Chen et al. | Saliency detection via the improved hierarchical principal component analysis method | |
CN110191299B (en) | Multi-frame interpolation method based on convolutional neural network | |
Lu et al. | Evolving block-based convolutional neural network for hyperspectral image classification | |
KR102138657B1 (en) | Apparatus and method for robust face recognition via hierarchical collaborative representation | |
CN111950649A (en) | Attention mechanism and capsule network-based low-illumination image classification method | |
CN110210556B (en) | Pedestrian re-identification data generation method | |
Wozniak et al. | A multiscale image compressor with rbfnn and discrete wavelet decomposition | |
CN112560865B (en) | Semantic segmentation method for point cloud under outdoor large scene | |
CN113743544A (en) | Cross-modal neural network construction method, pedestrian retrieval method and system | |
CN103268484A (en) | Design method of classifier for high-precision face recognitio | |
CN117743946B (en) | Signal type identification method and system based on fusion characteristic and group convolution ViT network | |
Chen et al. | Cumulative attribute space regression for head pose estimation and color constancy | |
CN117218351A (en) | Three-dimensional point cloud semantic segmentation method based on local and global context awareness | |
CN114170426A (en) | Algorithm model for classifying rare tumor category small samples based on cost sensitivity | |
Casagrande et al. | Abnormal motion analysis for tracking-based approaches using region-based method with mobile grid | |
CN113378598B (en) | Dynamic bar code detection method based on deep learning | |
Palomo et al. | Image compression and video segmentation using hierarchical self-organization | |
Gomes et al. | A Deep Learning Approach for Reconstruction of Color Images in Different Lighting Conditions Based on Autoencoder Technique | |
CN113205175A (en) | Multi-layer attribute network representation learning method based on mutual information maximization | |
Li et al. | Automated deep learning system for power line inspection image analysis and processing: Architecture and design issues | |
Psaltis et al. | Deep 3d flow features for human action recognition | |
Zhang et al. | GARBAGE CLASSIFICATION BASED ON A CASCADE NEURAL NETWORK. | |
Mousa et al. | Identification the modulation type in cognitive radio network based on Alexnet architecture | |
Faye et al. | Context Normalization Layer with Applications |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |