CN108960124B - Image processing method and device for pedestrian re-identification - Google Patents

Image processing method and device for pedestrian re-identification Download PDF

Info

Publication number
CN108960124B
CN108960124B CN201810691348.8A CN201810691348A CN108960124B CN 108960124 B CN108960124 B CN 108960124B CN 201810691348 A CN201810691348 A CN 201810691348A CN 108960124 B CN108960124 B CN 108960124B
Authority
CN
China
Prior art keywords
pedestrian
image
training
network model
position information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810691348.8A
Other languages
Chinese (zh)
Other versions
CN108960124A (en
Inventor
郭英强
张默
孙海涌
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Moshanghua Technology Co ltd
Original Assignee
Beijing Moshanghua Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Moshanghua Technology Co ltd filed Critical Beijing Moshanghua Technology Co ltd
Priority to CN201810691348.8A priority Critical patent/CN108960124B/en
Publication of CN108960124A publication Critical patent/CN108960124A/en
Application granted granted Critical
Publication of CN108960124B publication Critical patent/CN108960124B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • G06F18/24155Bayesian classification

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Probability & Statistics with Applications (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The application discloses an image processing method and device for pedestrian re-identification. The image processing method for pedestrian re-identification comprises the steps of extracting position information of a pedestrian in an image to be identified; segmenting a pedestrian image according to the position information; executing a preset network model test on the pedestrian image, wherein the preset network model at least comprises an attention branch when being tested; and identifying a target image according to the preset network model training result. The technical problem that the recognition accuracy is low in pedestrian image re-recognition is solved. The application can realize the following beneficial effects: the representation capability of the pedestrian image in the pedestrian picture is improved, the key information in the image is extracted, and the influence brought by the background is weakened.

Description

Image processing method and device for pedestrian re-identification
Technical Field
The application relates to the technical field of image recognition, in particular to an image processing method for pedestrian re-recognition.
Background
With the advance of smart city construction and the innovation of intelligent security technology, an intelligent video analysis technology in a video monitoring system becomes an effective means for saving manpower and material resources and improving the monitoring safety level. And the pedestrian re-identification task can complete the retrieval of related pedestrians in the monitoring field, and the labor cost is reduced. The pedestrian picture that usually detects out disturbs greatly, and the background is noisy, and the regional position that occupies of pedestrian is unfixed, also is difficult to accomplish the alignment like face identification.
Aiming at the problem of low identification precision in pedestrian image re-identification in the related technology, no effective solution is provided at present.
Disclosure of Invention
The present application mainly aims to provide an image processing method and an image processing device for pedestrian re-identification, so as to solve the problem of low identification accuracy in pedestrian image re-identification.
In order to achieve the above object, according to one aspect of the present application, there is provided an image processing method for pedestrian re-recognition.
The image processing method for pedestrian re-identification according to the application comprises the following steps:
extracting the position information of the pedestrian in the image to be identified;
segmenting a pedestrian image according to the position information;
executing a preset network model test on the pedestrian image, wherein the preset network model at least comprises an attention branch when being tested; and
and identifying a target image according to the preset network model training result.
Further, the extracting the position information of the pedestrian in the image to be recognized comprises:
acquiring a pedestrian video screenshot in an image acquisition device;
training a model at least comprising a DPM algorithm for detecting pedestrian position information according to the pedestrian video screenshot; and
and obtaining the position information of the pedestrian in the image to be recognized by the model execution position detection task at least comprising the DPM algorithm.
Further, the segmenting the pedestrian image according to the position information includes:
and segmenting the pedestrian from the image to be recognized according to the trained model for extracting the position of the pedestrian image, and storing the segmented pedestrian image.
Further, the performing of the preset network model training on the pedestrian image comprises:
training a depth residual error network model through the pedestrian image; and
and extracting high-dimensional features of the pedestrian image according to the depth residual error network model.
Further, the identifying the target image according to the preset network model training result includes:
training a combined Bayesian matrix according to the preset network model training result;
calculating the similarity between the pedestrian image and a target image according to the combined Bayesian matrix;
and screening the pedestrian images meeting the preset conditions.
In order to achieve the above object, according to another aspect of the present application, there is provided an image processing apparatus for pedestrian re-recognition.
An image processing apparatus for pedestrian re-recognition according to the present application includes:
the extraction module is used for extracting the position information of the pedestrian in the image to be identified;
the segmentation module is used for segmenting a pedestrian image according to the position information;
the training module is used for executing preset network model training on the pedestrian image, wherein the training of the preset network model at least comprises an attention branch; and
and the recognition module is used for recognizing the target image according to the preset network model training result.
Further, the extraction module comprises:
the acquisition unit is used for acquiring a pedestrian video screenshot in the image acquisition device;
the first training unit is used for training a model which is used for detecting pedestrian position information and at least comprises a DPM algorithm according to the pedestrian video screenshot; and
and the detection unit is used for executing a position detection task through the model at least comprising the DPM algorithm to obtain the position information of the pedestrian in the image to be recognized.
Further, the segmentation module comprises:
and the segmentation and storage unit is used for segmenting the pedestrian from the image to be recognized according to the trained model for extracting the position of the pedestrian image and storing the segmented pedestrian image.
Further, the training module comprises:
the second training unit is used for training a depth residual error network model through the pedestrian image; and
and the extraction unit is used for extracting the high-dimensional characteristics of the pedestrian image according to the depth residual error network model.
Further, the identification module comprises:
the third training unit is used for training a combined Bayesian matrix according to the preset network model training result;
the calculating unit is used for calculating the similarity between the pedestrian image and the target image according to the combined Bayesian matrix;
and the screening unit is used for screening the pedestrian images meeting the preset conditions.
In the embodiment of the application, the pedestrian image is segmented according to the position information by extracting the position information of the pedestrian in the image to be recognized, and a deep learning network model is adopted for testing, so that the purpose of extracting the image characteristics of the pedestrian is achieved, the similarity between the pedestrian image and the target image is further calculated through the combined Bayesian matrix, and finally a plurality of pedestrian images with higher similarity are output for manual comparison, so that the technical effect of finding the target image from mass video data is realized, and the technical problem of lower recognition precision in the re-recognition of the pedestrian image is solved.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this application, serve to provide a further understanding of the application and to enable other features, objects, and advantages of the application to be more apparent. The drawings and their description illustrate the embodiments of the invention and do not limit it. In the drawings:
FIG. 1 is a schematic diagram of an image processing method for pedestrian re-identification according to a first embodiment of the present application;
FIG. 2 is a schematic diagram of an image processing method for pedestrian re-identification according to a second embodiment of the present application;
FIG. 3 is a schematic diagram of an image processing method for pedestrian re-identification according to a third embodiment of the present application;
FIG. 4 is a schematic diagram of an image processing method for pedestrian re-identification according to a fourth embodiment of the present application;
FIG. 5 is a schematic diagram of an image processing apparatus for pedestrian re-identification according to a first embodiment of the present application;
FIG. 6 is a schematic diagram of an image processing apparatus for pedestrian re-identification according to a second embodiment of the present application;
fig. 7 is a schematic diagram of an image processing apparatus for pedestrian re-recognition according to a third embodiment of the present application;
FIG. 8 is a schematic diagram of an image processing apparatus for pedestrian re-identification according to a fourth embodiment of the present application;
FIG. 9 is a flowchart of an image processing method for pedestrian re-identification according to an embodiment of the present application;
FIG. 10a is a diagram of an attention branch structure according to an embodiment of the present application; and
fig. 10b is a diagram of a prior convolutional neural network structure according to an embodiment of the present application.
Detailed Description
In order to make the technical solutions better understood by those skilled in the art, the technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are only partial embodiments of the present application, but not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
It should be noted that the terms "first," "second," and the like in the description and claims of this application and in the drawings described above are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It should be understood that the data so used may be interchanged under appropriate circumstances such that embodiments of the application described herein may be used. Furthermore, the terms "comprises," "comprising," and "having," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed, but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
In this application, the terms "upper", "lower", "left", "right", "front", "rear", "top", "bottom", "inner", "outer", "middle", "vertical", "horizontal", "lateral", "longitudinal", and the like indicate orientations or positional relationships based on the orientations or positional relationships shown in the drawings. These terms are used primarily to better describe the present application and its embodiments, and are not used to limit the indicated devices, elements or components to a particular orientation or to be constructed and operated in a particular orientation.
Moreover, some of the above terms may be used to indicate other meanings besides the orientation or positional relationship, for example, the term "on" may also be used to indicate some kind of attachment or connection relationship in some cases. The specific meaning of these terms in this application will be understood by those of ordinary skill in the art as appropriate.
Furthermore, the terms "mounted," "disposed," "provided," "connected," and "sleeved" are to be construed broadly. For example, it may be a fixed connection, a removable connection, or a unitary construction; can be a mechanical connection, or an electrical connection; may be directly connected, or indirectly connected through intervening media, or may be in internal communication between two devices, elements or components. The specific meaning of the above terms in the present application can be understood by those of ordinary skill in the art as appropriate.
It should be noted that the embodiments and features of the embodiments in the present application may be combined with each other without conflict. The present application will be described in detail below with reference to the embodiments with reference to the attached drawings.
According to an embodiment of the present invention, there is provided an image processing method for pedestrian re-recognition, as shown in fig. 1, the method including steps S102 to S108 as follows:
step S102, extracting the position information of the pedestrian in the image to be identified;
preferably, the image to be recognized may be a video shot containing a pedestrian.
The position information of the pedestrian may be a specific motion frame position of the pedestrian image in the video.
The step of extracting the position information of the pedestrian in the image to be identified can be to acquire a video image containing the pedestrian, and the video image containing the pedestrian can be acquired through an image acquisition device such as a camera.
Step S104, segmenting a pedestrian image according to the position information;
preferably, the segmenting of the pedestrian image according to the position information may be a picture with the pedestrian image being cut out according to a specific motion frame position of the pedestrian image in the video.
Step S106, executing a preset network model test on the pedestrian image, wherein the preset network model at least comprises an attention branch when being tested; and
preferably, the preset network model may be a deep learning network model fused with a convolutional neural network and an attention branch.
Preferably, the preset network model can be a ResNet-50 neural network model, the network of the preset network model comprises 49 convolutional layers and 1 full-connection layer, a residual error network structure is introduced into the network, a large number of network stacking can be carried out, the feature vector of the higher layer is extracted, and finally the pedestrian picture represented by the 512-dimensional vector is obtained. And training the network using the training set.
As shown in fig. 10, an attention mechanism is added to an original convolutional neural network structure, in the attention mechanism network, after a feature map is obtained by a network structure, two routes of processing are performed on the network structure, one route is directly transmitted to Scale of a next layer as shown in fig. 10a, and the other route is subjected to full connection and Softmax operation to perform two classifications to obtain a probability about each region in the feature map, then Scale operation is performed on two routes, that is, each region in the feature map is assigned with different weights, and finally feature representation about a pedestrian picture is obtained through global pooling.
And S108, identifying a target image according to the preset network model training result.
Preferably, the training results may be network model parameters.
The target image may be an image to be matched, for example, an image of a suspect when a police department searches for the suspect, and an image of a missing person when looking for the missing person.
And enabling the pedestrian image to be retrieved to flow through the trained network model, calculating the similarity between the pedestrian feature obtained as a result and the target pedestrian image, and judging whether the pedestrian image and the target image are the same target or not.
As shown in fig. 2, the extracting of the position information of the pedestrian in the image to be recognized includes steps S202 to S206 as follows:
step S202, acquiring a pedestrian video screenshot in an image acquisition device;
preferably, the image acquisition device may be a monitoring camera or the like.
The pedestrian video screenshot in the image acquisition device can be acquired through model training, and the video screenshot containing the pedestrian image in the mass video can be acquired.
Step S204, training a model at least comprising a DPM algorithm for detecting pedestrian position information according to the pedestrian video screenshot; and
preferably, the DPM algorithm is a deformable component model, and is a component-based detection algorithm, and the gradient direction histogram is calculated first, and then the support vector machine is used for training to obtain the gradient model of the object.
The model which is trained according to the pedestrian video screenshot and used for detecting the pedestrian position information and at least comprises a DPM algorithm can detect the pedestrian position by utilizing the DPM characteristics. And cutting the pedestrian picture into 384x128 size for subsequent steps.
And step S206, executing a position detection task through the model at least comprising the DPM algorithm to obtain the position information of the pedestrian in the image to be identified.
Preferably, the pedestrian position is detected using the DPM feature. And cutting the pedestrian picture into 384x128 size for subsequent steps.
As an embodiment of the present invention, the segmenting the pedestrian image according to the position information includes:
and segmenting the pedestrian from the image to be recognized according to the trained model for extracting the position of the pedestrian image, and storing the segmented pedestrian image.
Preferably, the pedestrian picture is cropped to a size of 384x128 and saved in the corresponding file location.
As shown in fig. 3, the performing of the preset network model training on the pedestrian image includes steps S302 to S304 as follows:
step S302, training a depth residual error network model through the pedestrian image; and
preferably, a ResNet-50 neural network model is used, the network comprises 49 convolutional layers and 1 full-connection layer, a residual error network structure is introduced into the network, a large number of network stacking can be carried out, the feature vector of a higher layer is extracted, and finally a 512-dimensional vector representation is obtained to obtain a pedestrian picture. And training the network using the training set.
And step S304, extracting the high-dimensional characteristics of the pedestrian image according to the depth residual error network model.
Preferably, the high-dimensional features may be facial features or other body part features of a pedestrian, may also be wearing features, or the like.
And (4) extracting pedestrian picture features by using the neural network combined by the convolutional neural network and the attention system, and storing the pedestrian feature library.
As shown in fig. 4, the identifying a target image according to the preset network model training result includes steps S402 to S406 as follows:
step S402, training a combined Bayesian matrix according to the preset network model training result;
preferably, the training of the joint bayesian matrix according to the preset network model training result may be a training of the joint bayesian matrix according to the high-dimensional features of the pedestrian extracted by the depth residual error network model.
It should be noted that joint bayes is used to measure the pedestrian feature distance. The specific principle and the calculation process are as follows:
according to the joint Bayesian algorithm, the distribution of target pedestrians can be described as x ═ mu +Epsilon, where x is defined as the representation of a pedestrian, mu represents the difference of different people, and epsilon represents the difference of the same person under different light, posture and expression. The two latent variables μ and ε distributions obey two Gaussian distributions: n (0, S)μ) And N (0, S)ε).SμAnd SεRepresenting the covariance matrix to be determined, can be obtained by training.
Step S404, calculating the similarity between the pedestrian image and the target image according to the combined Bayesian matrix;
preferably, further, according to the joint bayesian algorithm, the formula of the calculated distance is finally obtained as follows:
Figure GDA0002948103130000091
wherein A is: a ═ Sμ+Sε)-1-(F+G)
G can be found by the following equation:
Figure GDA0002948103130000092
and step S406, screening the pedestrian images meeting the preset conditions.
Preferably, the preset condition may be how many pedestrian images with the highest similarity rank are set by the system, for example, 20 images with the highest similarity to the target pedestrian image are automatically screened out and displayed through an interface, and then manually determined.
As shown in fig. 9, the flow of the image processing method for pedestrian re-identification is as follows:
collecting a video image;
and detecting the pedestrians by the pictures containing the pedestrians, and cutting out the pictures only containing the single pedestrians.
The CNN is trained.
Training a network fusing the CNN and attention mechanism.
And (4) extracting pedestrian picture features by using the neural network luo fused with the CNN and the attention mechanism, and storing a pedestrian feature library.
And training a combined Bayesian matrix by using the obtained pedestrian features.
And calculating the similarity between the pedestrian features by using a combined Bayesian method.
And selecting 20 pieces with the highest similarity to be displayed through an interface, and judging by a person.
From the above description, it can be seen that the present invention achieves the following technical effects: the invention discloses an image processing method for pedestrian re-identification, which is based on a convolutional neural network and an attention mechanism and comprises the following steps: extracting pedestrian features by using a neural network fused with the CNN and the attention mechanism and storing the pedestrian features in a feature library; training a combined Bayesian matrix by using the pedestrian characteristics; measuring the similarity of the target pedestrian and the pedestrian in the contained library by using a combined Bayesian matrix; displaying the first 20 most similar ones through an interface according to the obtained similarity, and judging by a person; the above process is repeated to complete the pedestrian recognition. The pedestrian re-identification method can accurately execute the pedestrian re-identification task, has better representation on the pedestrian image, and greatly improves the identification precision.
It should be noted that the steps illustrated in the flowcharts of the figures may be performed in a computer system such as a set of computer-executable instructions and that, although a logical order is illustrated in the flowcharts, in some cases, the steps illustrated or described may be performed in an order different than presented herein.
According to an embodiment of the present invention, there is also provided an image processing apparatus for implementing the above-described image processing method for pedestrian re-recognition, as shown in fig. 5, the apparatus including:
the extraction module 10 is used for extracting the position information of the pedestrian in the image to be identified;
the segmentation module 20 is used for segmenting the pedestrian image according to the position information;
the training module 30 is configured to execute a preset network model test on the pedestrian image, where the preset network model test at least includes an attention branch; and
and the recognition module 40 is configured to recognize a target image according to the preset network model training result.
As shown in fig. 6, the extraction module includes:
the acquisition unit is used for acquiring a pedestrian video screenshot in the image acquisition device;
the first training unit is used for training a model which is used for detecting pedestrian position information and at least comprises a DPM algorithm according to the pedestrian video screenshot; and
and the detection unit is used for executing a position detection task through the model at least comprising the DPM algorithm to obtain the position information of the pedestrian in the image to be recognized.
Preferably, the segmentation module comprises:
and the segmentation and storage unit is used for segmenting the pedestrian from the image to be recognized according to the trained model for extracting the position of the pedestrian image and storing the segmented pedestrian image.
As shown in fig. 7, the training module 30 includes:
a second training unit 301, configured to train a depth residual error network model through the pedestrian image; and
an extracting unit 302, configured to extract a high-dimensional feature of the pedestrian image according to the depth residual error network model.
As shown in fig. 8, the identification module 40 includes:
a third training unit 401, configured to train a joint bayesian matrix according to the preset network model training result;
a calculating unit 402, configured to calculate a similarity between the pedestrian image and a target image according to the joint bayesian matrix;
a screening unit 403, configured to screen pedestrian images that meet preset conditions.
It will be apparent to those skilled in the art that the modules or steps of the present invention described above may be implemented by a general purpose computing device, they may be centralized on a single computing device or distributed across a network of multiple computing devices, and they may alternatively be implemented by program code executable by a computing device, such that they may be stored in a storage device and executed by a computing device, or fabricated separately as individual integrated circuit modules, or fabricated as a single integrated circuit module from multiple modules or steps. Thus, the present invention is not limited to any specific combination of hardware and software.
The above description is only a preferred embodiment of the present application and is not intended to limit the present application, and various modifications and changes may be made by those skilled in the art. Any modification, equivalent replacement, improvement and the like made within the spirit and principle of the present application shall be included in the protection scope of the present application.

Claims (8)

1. An image processing method for pedestrian re-identification, comprising:
extracting the position information of the pedestrian in the image to be identified;
segmenting a pedestrian image according to the position information;
executing a preset network model test on the pedestrian image, wherein the preset network model at least comprises an attention branch when being tested; and
identifying a target image according to the preset network model training result;
the extracting of the position information of the pedestrian in the image to be recognized comprises:
acquiring a pedestrian video screenshot in an image acquisition device;
training a model at least comprising a DPM algorithm for detecting pedestrian position information according to the pedestrian video screenshot; and
obtaining the position information of the pedestrian in the image to be recognized by the model execution position detection task at least comprising the DPM algorithm; cropping the pedestrian picture to 384x128 size;
wherein, training the combined Bayesian matrix through the preset network model training result comprises:
according to the joint bayesian measure of pedestrian feature distance, the distribution of target pedestrians can be described as x ═ μ + e, where x is defined as the representation of the pedestrian, μ represents the difference of different people, and e represents the same person in different light raysThe distribution of the two potential variables mu and epsilon obeys two Gaussian distributions: n (0, S)μ) And N (0, S)ε).SμAnd SεRepresents the covariance matrix to be determined;
calculating the similarity between the pedestrian image and a target image according to the combined Bayesian matrix; and
screening satisfies pedestrian's image of predetermineeing the condition, include: screening 20 pedestrian images with the highest similarity to the target pedestrian image;
wherein the position information of the pedestrian further includes: the specific action frame position of the pedestrian image in the video;
executing a preset network model test on the pedestrian image, wherein the preset network model at least comprises an attention branch when being tested, and the method comprises the following steps:
the method comprises the steps that a preset network model is a ResNet-50 neural network model, the network model comprises 49 convolutional layers and 1 full-connection layer, a residual error network structure is introduced into a network, and the network model is trained by using a training set;
adding an attention mechanism into the neural network model, processing the pedestrian image based on the attention mechanism network, and obtaining a characteristic representation of the pedestrian image;
in the attention mechanism network, after a network structure obtains a feature map, processing two routes on the feature map, directly transmitting the feature map to Scale of a next layer in one route, performing full connection and Softmax operation on the other route, performing secondary classification on the feature map to obtain the probability of each region in the feature map, performing Scale operation on the two routes, namely allocating different weights to each region in the feature map, and finally performing global pooling to obtain feature representation of a pedestrian picture;
training a combined Bayesian matrix through the preset network model training result, and further comprising:
training a combined Bayesian matrix according to the high-dimensional characteristics of the pedestrian image extracted by the depth residual error network model;
the image processing method for pedestrian re-identification comprises the following steps:
collecting a video image;
detecting the pedestrians by the pictures containing the pedestrians, and cutting out the pictures only containing the single pedestrians;
training a CNN;
training a network fusing the CNN and the attention mechanism;
extracting pedestrian picture features by using a neural network fused with a CNN (central nervous system) and an attention mechanism, and storing a pedestrian feature library;
training a combined Bayesian matrix by using the obtained pedestrian characteristics;
calculating the similarity between pedestrian features by using a combined Bayesian method;
selecting 20 pieces with highest similarity to be displayed through an interface;
and repeating the image processing process of pedestrian re-identification to obtain a pedestrian identification result.
2. The image processing method according to claim 1, wherein the segmenting the pedestrian image according to the position information includes:
and segmenting the pedestrian from the image to be recognized according to the trained model for extracting the position of the pedestrian image, and storing the segmented pedestrian image.
3. The image processing method according to claim 1, wherein the performing of the preset network model training on the pedestrian image comprises:
training a depth residual error network model through the pedestrian image; and
and extracting high-dimensional features of the pedestrian image according to the depth residual error network model.
4. The image processing method according to claim 1, wherein the identifying a target image according to the preset network model training result comprises:
training a combined Bayesian matrix according to the preset network model training result;
calculating the similarity between the pedestrian image and a target image according to the combined Bayesian matrix;
and screening the pedestrian images meeting the preset conditions.
5. An image processing apparatus for pedestrian re-recognition, comprising:
the extraction module is used for extracting the position information of the pedestrian in the image to be identified;
the segmentation module is used for segmenting a pedestrian image according to the position information;
the training module is used for executing preset network model training on the pedestrian image, wherein the training of the preset network model at least comprises an attention branch; and
the recognition module is used for recognizing a target image according to the preset network model training result;
the extraction module comprises:
the acquisition unit is used for acquiring a pedestrian video screenshot in the image acquisition device;
the first training unit is used for training a model which is used for detecting pedestrian position information and at least comprises a DPM algorithm according to the pedestrian video screenshot; and
the detection unit is used for executing a position detection task through the model at least comprising the DPM algorithm to obtain the position information of the pedestrian in the image to be identified; cropping the pedestrian picture to 384x128 size;
wherein, training the combined Bayesian matrix through the preset network model training result comprises:
according to the combined Bayes measure pedestrian characteristic distance, the distribution of target pedestrians can be described as x ═ mu + epsilon, wherein x is defined as the representation of the pedestrian, mu represents the difference of different people, epsilon represents the difference of the same people under different light rays, postures and expressions, and the two potential variables mu and epsilon are distributed according to two Gaussian distributions: n (0, S)μ) And N (0, S)ε).SμAnd SεRepresents the covariance matrix to be determined;
calculating the similarity between the pedestrian image and a target image according to the combined Bayesian matrix; and
screening satisfies pedestrian's image of predetermineeing the condition, include: screening 20 pedestrian images with the highest similarity to the target pedestrian image;
wherein the position information of the pedestrian further includes: the specific action frame position of the pedestrian image in the video;
the training module is used for executing a preset network model test on the pedestrian image, wherein the preset network model is tested at least comprises an attention branch, and the training module comprises:
the method comprises the steps that a preset network model is a ResNet-50 neural network model, the network model comprises 49 convolutional layers and 1 full-connection layer, a residual error network structure is introduced into a network, and the network model is trained by using a training set;
adding an attention mechanism into the neural network model, processing the pedestrian image based on the attention mechanism network, and obtaining a characteristic representation of the pedestrian image;
in the attention mechanism network, after a network structure obtains a feature map, processing two routes on the feature map, directly transmitting the feature map to Scale of a next layer in one route, performing full connection and Softmax operation on the other route, performing secondary classification on the feature map to obtain the probability of each region in the feature map, performing Scale operation on the two routes, namely allocating different weights to each region in the feature map, and finally performing global pooling to obtain feature representation of a pedestrian picture;
training a combined Bayesian matrix through the preset network model training result, and further comprising:
training a combined Bayesian matrix according to the high-dimensional characteristics of the pedestrian image extracted by the depth residual error network model;
the image processing method for pedestrian re-identification comprises the following steps:
collecting a video image;
detecting the pedestrians by the pictures containing the pedestrians, and cutting out the pictures only containing the single pedestrians;
training a CNN;
training a network fusing the CNN and the attention mechanism;
extracting pedestrian picture features by using a neural network fused with a CNN (central nervous system) and an attention mechanism, and storing a pedestrian feature library;
training a combined Bayesian matrix by using the obtained pedestrian characteristics;
calculating the similarity between pedestrian features by using a combined Bayesian method;
selecting 20 pieces with highest similarity to be displayed through an interface;
and repeating the image processing process of pedestrian re-identification to obtain a pedestrian identification result.
6. The image processing apparatus according to claim 5, wherein the segmentation module comprises:
and the segmentation and storage unit is used for segmenting the pedestrian from the image to be recognized according to the trained model for extracting the position of the pedestrian image and storing the segmented pedestrian image.
7. The image processing apparatus of claim 5, wherein the training module comprises:
the second training unit is used for training a depth residual error network model through the pedestrian image; and
and the extraction unit is used for extracting the high-dimensional characteristics of the pedestrian image according to the depth residual error network model.
8. The image processing apparatus according to claim 5, wherein the identification module comprises:
the third training unit is used for training a combined Bayesian matrix according to the preset network model training result;
the calculating unit is used for calculating the similarity between the pedestrian image and the target image according to the combined Bayesian matrix;
and the screening unit is used for screening the pedestrian images meeting the preset conditions.
CN201810691348.8A 2018-06-28 2018-06-28 Image processing method and device for pedestrian re-identification Active CN108960124B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810691348.8A CN108960124B (en) 2018-06-28 2018-06-28 Image processing method and device for pedestrian re-identification

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810691348.8A CN108960124B (en) 2018-06-28 2018-06-28 Image processing method and device for pedestrian re-identification

Publications (2)

Publication Number Publication Date
CN108960124A CN108960124A (en) 2018-12-07
CN108960124B true CN108960124B (en) 2021-10-01

Family

ID=64487945

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810691348.8A Active CN108960124B (en) 2018-06-28 2018-06-28 Image processing method and device for pedestrian re-identification

Country Status (1)

Country Link
CN (1) CN108960124B (en)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109977823B (en) * 2019-03-15 2021-05-14 百度在线网络技术(北京)有限公司 Pedestrian recognition and tracking method and device, computer equipment and storage medium
CN110096947A (en) * 2019-03-15 2019-08-06 昆明理工大学 A kind of pedestrian based on deep learning recognizer again
CN109948709B (en) * 2019-03-21 2020-06-23 南京博雅集智智能技术有限公司 Multitask attribute identification system of target object
CN110188611A (en) * 2019-04-26 2019-08-30 华中科技大学 A kind of pedestrian recognition methods and system again introducing visual attention mechanism
CN110070073A (en) * 2019-05-07 2019-07-30 国家广播电视总局广播电视科学研究院 Pedestrian's recognition methods again of global characteristics and local feature based on attention mechanism
CN110458004B (en) * 2019-07-02 2022-12-27 浙江吉利控股集团有限公司 Target object identification method, device, equipment and storage medium
CN111507213A (en) * 2020-04-03 2020-08-07 北京三快在线科技有限公司 Image recognition method, image recognition device, storage medium and electronic equipment
CN112488061B (en) * 2020-12-18 2022-04-29 电子科技大学 Multi-aircraft detection and tracking method combined with ADS-B information

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106803055A (en) * 2015-11-26 2017-06-06 腾讯科技(深圳)有限公司 Face identification method and device

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2410467A1 (en) * 2010-07-23 2012-01-25 Nederlandse Organisatie voor toegepast -natuurwetenschappelijk onderzoek TNO System and method for identifying image locations showing the same person in different images
CN106529442B (en) * 2016-10-26 2019-10-18 清华大学 A kind of pedestrian recognition method and device
US10380741B2 (en) * 2016-12-07 2019-08-13 Samsung Electronics Co., Ltd System and method for a deep learning machine for object detection
CN107301246A (en) * 2017-07-14 2017-10-27 河北工业大学 Chinese Text Categorization based on ultra-deep convolutional neural networks structural model
CN108198200B (en) * 2018-01-26 2022-03-08 福州大学 Method for tracking specified pedestrian on line under cross-camera scene

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106803055A (en) * 2015-11-26 2017-06-06 腾讯科技(深圳)有限公司 Face identification method and device

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
《基于局部深度匹配的行人再识别》;李邵梅,等;《计算机应用研究》;20160622;1235-1238 *
《跨视域摄像头网络下的监控视频结构化与检索》;朱烽;《中国博士学位论文全文数据库 信息科技辑》;20170915(第9期);摘要、正文第1-97页 *

Also Published As

Publication number Publication date
CN108960124A (en) 2018-12-07

Similar Documents

Publication Publication Date Title
CN108960124B (en) Image processing method and device for pedestrian re-identification
CN111931684B (en) Weak and small target detection method based on video satellite data identification features
Salimi et al. Visual-based trash detection and classification system for smart trash bin robot
US9008365B2 (en) Systems and methods for pedestrian detection in images
US20150054824A1 (en) Object detection method, object detection device, and image pickup device
CN111813997B (en) Intrusion analysis method, device, equipment and storage medium
CN104615986B (en) The method that pedestrian detection is carried out to the video image of scene changes using multi-detector
CN108009466B (en) Pedestrian detection method and device
CN105574550A (en) Vehicle identification method and device
CN104166841A (en) Rapid detection identification method for specified pedestrian or vehicle in video monitoring network
CN102804208A (en) Automatically mining person models of celebrities for visual search applications
JP2019074945A (en) Apparatus and method for online recognition and setting screen used therefor
CN110674680B (en) Living body identification method, living body identification device and storage medium
CN112365586B (en) 3D face modeling and stereo judging method and binocular 3D face modeling and stereo judging method of embedded platform
CN104463232A (en) Density crowd counting method based on HOG characteristic and color histogram characteristic
EP2860661A1 (en) Mean shift tracking method
CN116647644B (en) Campus interactive monitoring method and system based on digital twin technology
CN111507353B (en) Chinese field detection method and system based on character recognition
CN107610224B (en) 3D automobile object class representation algorithm based on weak supervision and definite block modeling
US20170053172A1 (en) Image processing apparatus, and image processing method
CN113780145A (en) Sperm morphology detection method, sperm morphology detection device, computer equipment and storage medium
CN116597411A (en) Method and system for identifying traffic sign by unmanned vehicle in extreme weather
CN115862113A (en) Stranger abnormity identification method, device, equipment and storage medium
CN111753601B (en) Image processing method, device and storage medium
CN111126112B (en) Candidate region determination method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
EE01 Entry into force of recordation of patent licensing contract
EE01 Entry into force of recordation of patent licensing contract

Application publication date: 20181207

Assignee: Apple R&D (Beijing) Co., Ltd.

Assignor: BEIJING MOSHANGHUA TECHNOLOGY CO., LTD.

Contract record no.: 2019990000054

Denomination of invention: The image processing method and device identified again for pedestrian

License type: Exclusive License

Record date: 20190211

GR01 Patent grant
GR01 Patent grant