CN111461181A - Vehicle fine-grained classification method and device - Google Patents

Vehicle fine-grained classification method and device Download PDF

Info

Publication number
CN111461181A
CN111461181A CN202010183058.XA CN202010183058A CN111461181A CN 111461181 A CN111461181 A CN 111461181A CN 202010183058 A CN202010183058 A CN 202010183058A CN 111461181 A CN111461181 A CN 111461181A
Authority
CN
China
Prior art keywords
feature map
global
classification
classification method
neural network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010183058.XA
Other languages
Chinese (zh)
Other versions
CN111461181B (en
Inventor
傅慧源
马华东
王川铭
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing University of Posts and Telecommunications
Original Assignee
Beijing University of Posts and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Posts and Telecommunications filed Critical Beijing University of Posts and Telecommunications
Priority to CN202010183058.XA priority Critical patent/CN111461181B/en
Publication of CN111461181A publication Critical patent/CN111461181A/en
Application granted granted Critical
Publication of CN111461181B publication Critical patent/CN111461181B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/08Detecting or categorising vehicles

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • Computing Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Probability & Statistics with Applications (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)
  • Traffic Control Systems (AREA)

Abstract

The invention provides a vehicle fine-grained classification method and a vehicle fine-grained classification device, wherein the classification method comprises the following steps: extracting a basic characteristic diagram of an input picture through a convolutional neural network; adaptively constructing a global structure diagram according to the basic feature diagram; adopting a graph convolutional neural network to carry out incidence relation reasoning on the global structure graph to form global guidance information; applying the global guidance information to the basic feature map to obtain an enhanced feature map; and inputting the enhanced feature map into a classifier for classification. The classification method of the invention leads the obtained characteristic diagram to have stronger expression capability by reasoning the incidence relation among different components of the vehicle and applying the inference result of the incidence relation to the basic characteristic diagram, thus improving the accuracy of classification of fine-grained images of the vehicle; in addition, the classification method does not need to label the key area, is rapid in feature extraction, wide in application scene and low in calculation consumption.

Description

Vehicle fine-grained classification method and device
Technical Field
The invention relates to the technical field of vehicle classification, in particular to a vehicle fine-grained classification method and device.
Background
The vehicle is used as an important vehicle for people to go out and is closely related to the life of people, and the development of the modern vehicle production technology enables more and more types of automobiles to appear in the daily life of people; vehicles generally have similar appearance structures, different types of vehicles only have slight differences in positions of lamps, doors and the like, and related experts are often needed for identifying the fine granularity of the vehicles.
The fine-grained classification of vehicles is one of the most important tasks in computer vision, and not only plays an important role in other vision tasks such as vehicle re-identification and vehicle tracking, but also can provide assistance for the brain and intelligent traffic construction of cities; at the same time, this is a very challenging problem, since identifying models requires capturing subtle visual differences between classes or instances that are easily masked by other factors (e.g., viewpoint, lighting, or scene).
The basic flow of the existing vehicle fine-grained classification method can be summarized as follows:
basic feature extraction: usually, the basic feature extraction is performed through the CNNs to obtain a feature map of input picture data;
finding the position corresponding to the key area of the vehicle on the characteristic diagram by a supervision or unsupervised method; the supervision means that each image containing the vehicle to be recognized is marked with corresponding key area positions, usually four coordinate values, which represent the positions of the upper left corner point and the lower right corner point of the area; the unsupervised method is to obtain the position of a key area through an attention mechanism or other unsupervised learning methods;
further extracting fine-grained characteristics of the key area; for the found key area, utilizing a neural network to extract features;
splicing the independent fine-grained characteristics of each key area to obtain the characteristic representation of the vehicle, wherein the splicing method is generally the splicing of arrays or numerical matrixes;
and classifying the spliced vehicle characteristic representation through a classifier to obtain a final classification result.
However, the existing classification methods have the following problems:
when searching for a key area in the existing classification method, the supervised method usually needs additional marking information to mark positions of a car lamp, a car window, a car door and the like of a car, so that the corresponding area on the feature map can be found, the marking cost of data is greatly increased, the number of data is limited, and the applicable scene of the model is also limited;
the existing classification method needs to extract the characteristics of key areas respectively, and the calculation consumption of the classification method is increased by multiple times of calculation.
Disclosure of Invention
In view of the above, the present invention provides a vehicle fine-grained classification method to solve the problems that the existing classification method needs to label the key area information, the key area is not associated, and the calculation consumption is large.
Based on the above object, the present invention provides a vehicle fine-grained classification method, comprising:
the method comprises the following steps: extracting a basic characteristic diagram of an input picture through a convolutional neural network;
step two: adaptively constructing a global structure diagram according to the basic feature diagram;
step three: adopting a graph convolutional neural network to carry out incidence relation reasoning on the global structure graph to form global guidance information;
step four: applying the global guidance information to the basic feature map to obtain an enhanced feature map;
step five: and inputting the enhanced feature map into a classifier for classification.
Optionally, the method for extracting the basic feature map includes:
the method comprises the steps of inputting an input picture into a residual error neural network, taking the output of a layer3 structure of the residual error neural network as the basic feature map, adjusting the size of the input picture to a test size by a linear interpolation method before the input picture is input into the residual error neural network, for example, adjusting the size of the input picture to 448 × 448, and preferably selecting the number of layers of the residual error neural network according to actual needs, for example, selecting a residual error neural network with 34 layers, 50 layers or 101-layer network layers, preferably a residual error neural network with 101-layer network layers.
Optionally, the constructing the global structure diagram includes constructing node information and side information of the global structure diagram;
the construction of the node information comprises the following steps:
carrying out region segmentation on the basic feature graph on the length dimension and the width dimension by adopting the size of a predefined region to obtain a multi-scale region, then respectively aggregating the features of each region through pooling operation in a neural network, and aggregating the features of each region into a one-dimensional vector with the length being the dimension of the channel of the basic feature graph to obtain the node information;
the construction of the side information comprises:
calculating the incidence relation between the node information by adopting matrix multiplication to obtain the side information;
the correlation is shown in formula I:
V=η(A,Wη)*(A,W) Formula I;
in formula I, V is side information, A is node information, η sum is two convolution layers respectively, W represents parameters of the convolution layers, and represents matrix multiplication.
Optionally, the performing association relationship inference on the global structure diagram by using the graph convolutional neural network includes:
inputting the global structure chart into a graph convolution neural network with a layer of graph convolution layer, and carrying out association relationship reasoning on different region characteristics by utilizing the side information.
Optionally, the relationship reasoning is as shown in formula II:
G=V*(A*Wg) Formula II;
in formula II, G is global guidance information; v is side information; a is node information; wg is a parameter of the graph convolution neural network; denotes matrix multiplication.
Optionally, the applying the global guidance information on the basic feature map includes: applying the global guidance information on the channel dimension and the space dimension of the basic feature map;
the applying the global guidance information on the channel dimension of the basic feature map comprises:
pooling operation is carried out according to the space direction to pool the global structure chart into a one-dimensional vector, and then the one-dimensional vector is multiplied by the basic characteristic chart according to the dimension;
the applying global guidance information on the spatial dimension of the basic feature map comprises:
generating a local attention map from the basic feature map through a convolution layer;
pooling the global guidance information into a global feature map with only one dimension according to the channel direction through pooling operation;
the base feature map is multiplied by the local attention map and the global feature map by pixels.
Optionally, the inputting the enhanced feature map into a classifier for classification includes:
aggregating the enhanced feature maps into a feature vector with the channel dimension of the basic feature map through a spatial pooling layer;
and inputting the feature vectors into a layer of full-link layer for classification.
Optionally, the classification method further includes:
taking a Stanford Cars Dataset (Stanford Cars Dataset) as a training Dataset, processing images in the training Dataset according to the steps from one step to five to obtain a prediction classification result, carrying out loss calculation on the prediction classification result and a standard result given in the training Dataset according to a cross entropy loss method, then, reversely transmitting the calculated loss to a corresponding position of a parameter to be learned through a reverse propagation method, and modifying the parameter according to a random gradient descent method to obtain the parameter of the final model.
Optionally, the loss calculation formula is shown in formula III:
Figure BDA0002413233270000041
in formula III, L represents the loss, N represents the total number of images, M represents the number of classes, Pic represents the probability that the current prediction classification result is class c, and yic represents whether the prediction classification result class is the same as the given class in the training data set, and is 1 when the same class is used, and is not 0 when the same class is used.
A second aspect of the embodiments of the present invention also provides a vehicle fine-grained classification device, where the classification device includes: the classification method comprises a memory, a processor connected with the memory, and a computer program stored on the memory and capable of running on the processor, wherein the processor executes the computer program to execute the classification method.
From the above, it can be seen that the method and the device for classifying the fine granularity of the vehicle provided by the invention at least comprise the following technical effects:
according to the classification method, the incidence relation between different components of the vehicle is inferred, and the inference result of the incidence relation is applied to the basic characteristic diagram, so that the obtained characteristic diagram has stronger expression capability, and the accuracy of classification of fine-grained images of the vehicle is improved; in addition, the classification method does not need to label the key area, is rapid in feature extraction, wide in application scene and low in calculation consumption.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the drawings without creative efforts.
FIG. 1 is a flowchart of a fine-grained classification method for vehicles according to an embodiment of the present invention;
FIG. 2 is a flow chart of a construction of a global structure diagram according to an embodiment of the present invention;
FIG. 3 is a flowchart of applying global guidance information provided by an embodiment of the present invention to a basic feature diagram;
fig. 4 is a diagram illustrating a variation of the vehicle classification accuracy in the training process according to the embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is described in further detail below with reference to specific embodiments and the accompanying drawings.
It is to be noted that technical terms or scientific terms used in the embodiments of the present invention should have the ordinary meanings as understood by those having ordinary skill in the art to which the present disclosure belongs, unless otherwise defined.
The fine-grained classification of vehicles is one of the most important tasks in computer vision, and not only plays an important role in other vision tasks such as vehicle re-identification and vehicle tracking, but also can provide assistance for the brain and intelligent traffic construction of cities; at the same time, this is a very challenging problem, since identifying models requires capturing subtle visual differences between classes or instances that are easily masked by other factors (e.g., viewpoint, lighting, or scene).
The basic flow of the existing vehicle fine-grained classification method can be summarized as follows: extracting basic features; finding the position corresponding to the key area of the vehicle on the characteristic diagram by a supervision or unsupervised method; further extracting fine-grained characteristics of the key area; splicing the independent fine-grained characteristics of each key area to obtain the characteristic representation of the vehicle; and classifying the spliced vehicle characteristic representation through a classifier to obtain a final classification result.
However, the existing classification methods have the following problems:
when searching for a key area in the existing classification method, the supervised method usually needs additional marking information to mark positions of a car lamp, a car window, a car door and the like of a car, so that the corresponding area on the feature map can be found, the marking cost of data is greatly increased, the number of data is limited, and the applicable scene of the model is also limited; in addition, the existing classification method needs to extract the features of the key areas respectively, and the calculation consumption of the classification method is increased by multiple times of calculation.
In view of the above technical problem, the present invention provides a vehicle fine-grained classification method, as shown in fig. 1, the method includes:
the method comprises the following steps: extracting a basic feature map of an input picture through a convolutional neural network, wherein the basic feature map is actually an array with multiple dimensions obtained by the picture data through the convolutional neural network;
step two: adaptively constructing a global structure diagram according to the basic feature diagram;
step three: adopting a graph convolutional neural network to carry out incidence relation reasoning on the global structure graph to form global guidance information;
step four: applying the global guidance information to the basic feature map to obtain an enhanced feature map;
step five: and inputting the enhanced feature map into a classifier for classification.
The classification method considers the incidence relation among different components of the vehicle by carrying out incidence relation reasoning, applies the incidence relation reasoning result to the basic characteristic diagram, enables the obtained characteristic diagram to have stronger expression capability, and improves the accuracy of classification of fine-grained images of the vehicle.
The method for extracting the basic feature map is not strictly limited in the invention, for example, in an embodiment, a residual error neural network with a 50-layer network layer can be used for extracting, specifically, an input picture is input into the residual error neural network with the 50-layer network layer, the output of a layer3 structure of a residual error through the network is used as the basic feature map, in an embodiment, before the input picture is input into the residual error neural network with the 50-layer network layer, the size of the input picture can be adjusted to 448 × 448 through a linear interpolation method, under the condition that the input picture is 448 × 448, the size of the basic feature map is 28 × 28, and each pixel point is a vector with 1024 numerical values.
Fig. 2 is a flow chart of the construction of the global structure diagram of the present invention, and as shown in fig. 2, the construction of the global structure diagram includes the construction of node information and side information of the global structure diagram;
the construction of the node information comprises the following steps:
carrying out region segmentation on the basic feature graph on the length dimension and the width dimension by adopting the size of a predefined region to obtain a multi-scale region, then respectively aggregating the features of each region through pooling operation in a neural network, and aggregating the features of each region into a one-dimensional vector with the length being the dimension of the channel of the basic feature graph to obtain node information;
the construction of the side information comprises the following steps:
and calculating the incidence relation between the node information by adopting matrix multiplication to obtain the side information.
The above relationship is shown in formula I:
V=η(A,Wη)*(A,W) Formula I;
in formula I, V is side information, A is node information, η sum is two convolution layers respectively, W represents parameters of the convolution layers, and represents matrix multiplication.
By calculating the incidence relation among the node information, the side information of the global structure chart is constructed in a self-adaptive manner; the constructed global structure diagram is adaptively changed for different input pictures in the presence of side information, so that the constructed global structure diagram can more robustly respond to the change of factors such as vehicle postures in the image.
In the present invention, the method for reasoning the association relationship of the global structure diagram by using the convolutional neural network is not strictly limited, for example, in an embodiment, the method for reasoning the association relationship of the global structure diagram by using the convolutional neural network may specifically be performed by the following method:
inputting the global structure chart into a graph convolution neural network with a layer of graph convolution layer, and carrying out incidence relation reasoning on different region characteristics by utilizing side information of the global structure chart;
the above correlation reasoning is performed by using a formula shown in formula II:
G=V*(A*Wg) Formula II;
in formula II, G is global guidance information; v is side information; a is node information; wg is a parameter of the graph convolution neural network; denotes matrix multiplication.
The node information in the global structure diagram obtained in the invention is still isolated, and although the node information is connected through the side information, the node information and the side information are not exchanged, so that the incidence relation reasoning is carried out on the global structure diagram by adopting the graph convolution neural network, and the expression capability of the vehicle global structure diagram is improved.
FIG. 3 is a flow chart of the application of global guidance information to a base feature map in accordance with the present invention; as shown in fig. 3, applying the global guide information on the basic feature map includes: applying global guidance information on channel dimensions and space dimensions of a basic feature map (which may be abbreviated as a feature map);
specifically, applying the global guidance information on the channel dimension of the basic feature map includes:
pooling the global structure graph into a one-dimensional vector (g) by a spatial direction pooling operationc) Multiplying the one-dimensional vector by the basic characteristic diagram according to dimensions, wherein the multiplication simultaneously follows an array propagation rule;
applying the global guidance information in the spatial dimension of the base feature map comprises:
generating a local attention map (g) from the base feature map by a convolutional layers1);
Pooling the global guide information in channel direction into a global feature map (g) with only one dimension by a pooling operations2);
The base feature map is multiplied by the local attention map and the global feature map by pixels.
The classification method in the present invention is not strictly limited, and for example, the classification can be performed by using a conventional classification method in the art according to actual needs; in one embodiment, the enhanced feature maps may be aggregated into a feature vector having the dimensions of the channel of the basic feature map by a spatial pooling layer;
and inputting the feature vectors into a full-link layer for classification.
In one embodiment, the classification method further comprises:
taking Stanford Cars Dataset as a training Dataset, processing images in the training Dataset according to the steps from one to five to obtain a prediction classification result, performing loss calculation on the prediction classification result and a standard result given in the training Dataset according to a cross entropy loss method, then, transmitting the calculated loss back to a corresponding position of a parameter to be learned through a back propagation method, and modifying the parameter according to a random gradient descent method to obtain the parameter of the final model.
Specifically, in one embodiment, images in the Stanford Cars Dataset training Dataset are normalized to a size of 448x448, the normalized images are input into the above-described steps one-five of the present invention to obtain a predicted classification result, the loss of the predicted classification result and the standard result is calculated according to the formula shown in formula III below,
Figure BDA0002413233270000081
in the formula III, L is loss, N is the total number of images, M is the number of categories, Pic represents the probability when the current prediction classification result is the category c, yic represents whether the prediction classification result category is the same as the given category of the training data set, and is 1 when the prediction classification result category is the same and is not 0 when the prediction classification result category is not the same;
then, a random gradient descent is used as an optimizer, and the parameters thereof are updated according to formula IV:
Figure BDA0002413233270000082
in formula IV, θ is a parameter to be updated, and α represents the learning rate;
setting the learning rate to be 0.1, reducing the learning rate by 10 times after every 30 rounds of training, and obtaining the final parameters after 100 rounds of training; the change of the accuracy in the training process is shown in fig. 4, wherein the ordinate is the accuracy and the abscissa is the number of training rounds;
as can be seen from FIG. 4, the accuracy of the vehicle fine-grained classification method can reach 92.4%.
The invention also provides a vehicle fine-grained classification device, which comprises: the classification method comprises a memory, a processor connected with the memory, and a computer program stored on the memory and capable of running on the processor, wherein the processor executes the computer program to execute the classification method.
According to the classification method, the incidence relation between different components of the vehicle is inferred, and the inference result of the incidence relation is applied to the basic characteristic diagram, so that the obtained characteristic diagram has stronger expression capability, and the accuracy of classification of fine-grained images of the vehicle is improved; in addition, the classification method does not need to label key areas and extract features, has wide application scenes and low calculation consumption.
Those of ordinary skill in the art will understand that: the discussion of any embodiment above is meant to be exemplary only, and is not intended to intimate that the scope of the disclosure, including the claims, is limited to these examples; within the idea of the invention, also features in the above embodiments or in different embodiments may be combined, steps may be implemented in any order, and there are many other variations of the different aspects of the invention as described above, which are not provided in detail for the sake of brevity.
While the present invention has been described in conjunction with specific embodiments thereof, many alternatives, modifications, and variations of these embodiments will be apparent to those of ordinary skill in the art in light of the foregoing description.
The embodiments of the invention are intended to embrace all such alternatives, modifications and variances that fall within the broad scope of the appended claims. Therefore, any omissions, modifications, substitutions, improvements and the like that may be made without departing from the spirit and principles of the invention are intended to be included within the scope of the invention.

Claims (10)

1. A vehicle fine-grained classification method is characterized by comprising the following steps:
the method comprises the following steps: extracting a basic characteristic diagram of an input picture through a convolutional neural network;
step two: adaptively constructing a global structure diagram according to the basic feature diagram;
step three: adopting a graph convolutional neural network to carry out incidence relation reasoning on the global structure graph to form global guidance information;
step four: applying the global guidance information to the basic feature map to obtain an enhanced feature map;
step five: and inputting the enhanced feature map into a classifier for classification.
2. The classification method according to claim 1, wherein the basic feature map extraction method comprises:
and inputting the input picture into the residual error neural network, and taking the output of the layer3 structure of the residual error neural network as the basic feature map.
3. The classification method according to claim 1, wherein the constructing the global structure diagram includes constructing node information and side information of the global structure diagram;
the construction of the node information comprises the following steps:
carrying out region segmentation on the basic feature graph on the length dimension and the width dimension by adopting the size of a predefined region to obtain a multi-scale region, then respectively aggregating the features of each region through pooling operation in a neural network, and aggregating the features of each region into a one-dimensional vector with the length being the dimension of the channel of the basic feature graph to obtain the node information;
the construction of the side information comprises:
calculating the incidence relation between the node information by adopting matrix multiplication to obtain the side information;
the correlation is shown in formula I:
V=η(A,Wη)*(A,W) Formula I;
in formula I, V is side information, A is node information, η sum is two convolution layers respectively, W represents parameters of the convolution layers, and represents matrix multiplication.
4. The classification method according to claim 3, wherein the performing association relationship inference on the global structure graph by using the graph convolutional neural network comprises:
inputting the global structure chart into a graph convolution neural network with a layer of graph convolution layer, and carrying out association relationship reasoning on different region characteristics by utilizing the side information.
5. The classification method according to claim 4, wherein the correlation reasoning is as shown in formula II:
G=V*(A*Wg) Formula II;
in formula II, G is global guidance information; v is side information; a is node information; wg is a parameter of the graph convolution neural network; denotes matrix multiplication.
6. The classification method according to claim 1, wherein the applying global guidance information on the basic feature map comprises: applying the global guidance information on the channel dimension and the space dimension of the basic feature map;
the applying the global guidance information on the channel dimension of the basic feature map comprises:
pooling operation is carried out according to the space direction to pool the global structure chart into a one-dimensional vector, and then the one-dimensional vector is multiplied by the basic characteristic chart according to the dimension;
the applying global guidance information on the spatial dimension of the basic feature map comprises:
generating a local attention map from the basic feature map through a convolution layer;
pooling the global guidance information into a global feature map with only one dimension according to the channel direction through pooling operation;
the base feature map is multiplied by the local attention map and the global feature map by pixels.
7. The classification method according to claim 1, wherein the inputting the enhanced feature map into a classifier for classification comprises:
aggregating the enhanced feature maps into a feature vector with the channel dimension of the basic feature map through a spatial pooling layer;
and inputting the feature vectors into a layer of full-link layer for classification.
8. The classification method according to claim 1, further comprising:
taking Stanford Cars Dataset as a training Dataset, processing images in the training Dataset according to the steps from one to five in the claim 1 to obtain a prediction classification result, carrying out loss calculation on the prediction classification result and a given standard result in the training Dataset according to a cross entropy loss method, then, transmitting the calculated loss back to a corresponding position of a parameter to be learned through a back propagation method, and modifying the parameter according to a random gradient descent method to obtain the parameter of a final model.
9. The classification method according to claim 8, wherein the loss calculation formula is shown in formula III:
Figure FDA0002413233260000021
in formula III, L represents the loss, N represents the total number of images, M represents the number of classes, Pic represents the probability that the current prediction classification result is class c, and yic represents whether the prediction classification result class is the same as the given class in the training data set, and is 1 when the same class is used, and is not 0 when the same class is used.
10. A vehicle fine grain classification apparatus, characterized by comprising: a memory, a processor connected to the memory, and a computer program stored on the memory and executable on the processor, the processor executing the computer program to perform the classification method of any one of claims 1 to 9.
CN202010183058.XA 2020-03-16 2020-03-16 Vehicle fine-grained classification method and device Active CN111461181B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010183058.XA CN111461181B (en) 2020-03-16 2020-03-16 Vehicle fine-grained classification method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010183058.XA CN111461181B (en) 2020-03-16 2020-03-16 Vehicle fine-grained classification method and device

Publications (2)

Publication Number Publication Date
CN111461181A true CN111461181A (en) 2020-07-28
CN111461181B CN111461181B (en) 2021-09-07

Family

ID=71684316

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010183058.XA Active CN111461181B (en) 2020-03-16 2020-03-16 Vehicle fine-grained classification method and device

Country Status (1)

Country Link
CN (1) CN111461181B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111814920A (en) * 2020-09-04 2020-10-23 中国科学院自动化研究所 Fine classification method and system for multi-granularity feature learning based on graph network
CN112150821A (en) * 2020-10-14 2020-12-29 清华大学 Lightweight vehicle detection model construction method, system and device
CN114373080A (en) * 2022-03-22 2022-04-19 中国石油大学(华东) Hyperspectral classification method of lightweight hybrid convolution model based on global reasoning

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160307072A1 (en) * 2015-04-17 2016-10-20 Nec Laboratories America, Inc. Fine-grained Image Classification by Exploring Bipartite-Graph Labels
US20180096457A1 (en) * 2016-09-08 2018-04-05 Carnegie Mellon University Methods and Software For Detecting Objects in Images Using a Multiscale Fast Region-Based Convolutional Neural Network
CN108830823A (en) * 2018-03-14 2018-11-16 西安理工大学 The full-reference image quality evaluating method of frequency-domain analysis is combined based on airspace
CN109359696A (en) * 2018-10-29 2019-02-19 重庆中科云丛科技有限公司 A kind of vehicle money recognition methods, system and storage medium
US20190122111A1 (en) * 2017-10-24 2019-04-25 Nec Laboratories America, Inc. Adaptive Convolutional Neural Knowledge Graph Learning System Leveraging Entity Descriptions
CN110111337A (en) * 2019-04-16 2019-08-09 中山大学 A kind of general human body analytical framework and its analytic method based on figure transfer learning
CN110309732A (en) * 2019-06-13 2019-10-08 浙江大学 Activity recognition method based on skeleton video
CN110443261A (en) * 2019-08-15 2019-11-12 南京邮电大学 A kind of more figure matching process restored based on low-rank tensor
CN110619369A (en) * 2019-09-23 2019-12-27 常熟理工学院 Fine-grained image classification method based on feature pyramid and global average pooling
CN110751212A (en) * 2019-10-21 2020-02-04 南京大学 Efficient fine-grained image identification method on mobile equipment

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160307072A1 (en) * 2015-04-17 2016-10-20 Nec Laboratories America, Inc. Fine-grained Image Classification by Exploring Bipartite-Graph Labels
US20180096457A1 (en) * 2016-09-08 2018-04-05 Carnegie Mellon University Methods and Software For Detecting Objects in Images Using a Multiscale Fast Region-Based Convolutional Neural Network
US20190122111A1 (en) * 2017-10-24 2019-04-25 Nec Laboratories America, Inc. Adaptive Convolutional Neural Knowledge Graph Learning System Leveraging Entity Descriptions
CN108830823A (en) * 2018-03-14 2018-11-16 西安理工大学 The full-reference image quality evaluating method of frequency-domain analysis is combined based on airspace
CN109359696A (en) * 2018-10-29 2019-02-19 重庆中科云丛科技有限公司 A kind of vehicle money recognition methods, system and storage medium
CN110111337A (en) * 2019-04-16 2019-08-09 中山大学 A kind of general human body analytical framework and its analytic method based on figure transfer learning
CN110309732A (en) * 2019-06-13 2019-10-08 浙江大学 Activity recognition method based on skeleton video
CN110443261A (en) * 2019-08-15 2019-11-12 南京邮电大学 A kind of more figure matching process restored based on low-rank tensor
CN110619369A (en) * 2019-09-23 2019-12-27 常熟理工学院 Fine-grained image classification method based on feature pyramid and global average pooling
CN110751212A (en) * 2019-10-21 2020-02-04 南京大学 Efficient fine-grained image identification method on mobile equipment

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
C. WANG等: ""Global Structure Graph Guided Fine-Grained Vehicle Recognition"", 《ICASSP 2020 - 2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP)》 *
J. KRAUSE等: ""Fine-grained recognition without part annotations"", 《2015 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR)》 *
TIANSHUI CHEN等: ""Knowledge-Embedded Representation Learning for Fine-Grained Image Recognition"", 《ARXIV:1807.00505V1》 *
YIN LI等: ""Beyond grids: learning graph representations for visual recognition"", 《NIPS"18: PROCEEDINGS OF THE 32ND INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING SYSTEMS》 *
冯语姗等: ""自上而下注意图分割的细粒度图像分类"", 《中国图象图形学报》 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111814920A (en) * 2020-09-04 2020-10-23 中国科学院自动化研究所 Fine classification method and system for multi-granularity feature learning based on graph network
CN112150821A (en) * 2020-10-14 2020-12-29 清华大学 Lightweight vehicle detection model construction method, system and device
CN114373080A (en) * 2022-03-22 2022-04-19 中国石油大学(华东) Hyperspectral classification method of lightweight hybrid convolution model based on global reasoning
CN114373080B (en) * 2022-03-22 2022-07-29 中国石油大学(华东) Hyperspectral classification method of lightweight hybrid convolution model based on global reasoning

Also Published As

Publication number Publication date
CN111461181B (en) 2021-09-07

Similar Documents

Publication Publication Date Title
Sakkos et al. End-to-end video background subtraction with 3d convolutional neural networks
Li et al. Deep neural network for structural prediction and lane detection in traffic scene
WO2021093435A1 (en) Semantic segmentation network structure generation method and apparatus, device, and storage medium
CN106845487B (en) End-to-end license plate identification method
CN111461181B (en) Vehicle fine-grained classification method and device
CN110119148B (en) Six-degree-of-freedom attitude estimation method and device and computer readable storage medium
CN110163286B (en) Hybrid pooling-based domain adaptive image classification method
CN110633632A (en) Weak supervision combined target detection and semantic segmentation method based on loop guidance
CN113468978B (en) Fine granularity car body color classification method, device and equipment based on deep learning
CN113869138A (en) Multi-scale target detection method and device and computer readable storage medium
Grigorev et al. Depth estimation from single monocular images using deep hybrid network
CN112070174A (en) Text detection method in natural scene based on deep learning
CN116863194A (en) Foot ulcer image classification method, system, equipment and medium
CN113435370B (en) Method and device for acquiring vehicle queuing length based on image feature fusion
Panda et al. Modified ResNet-152 network with hybrid pyramidal pooling for local change detection
Du et al. Adaptive visual interaction based multi-target future state prediction for autonomous driving vehicles
CN114596548A (en) Target detection method, target detection device, computer equipment and computer-readable storage medium
Nguyen et al. Robust stereo data cost with a learning strategy
CN112991239A (en) Image reverse recovery method based on deep learning
CN116452472A (en) Low-illumination image enhancement method based on semantic knowledge guidance
CN115240163A (en) Traffic sign detection method and system based on one-stage detection network
Xu et al. Deep Neural Network‐Based Sports Marketing Video Detection Research
CN111539420A (en) Panoramic image saliency prediction method and system based on attention perception features
Jain et al. A Quarter Century Journey: Evolution of Object Detection Methods
CN114998601B (en) On-line update target tracking method and system based on Transformer

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant