CN115082801A - Airplane model identification system and method based on remote sensing image - Google Patents

Airplane model identification system and method based on remote sensing image Download PDF

Info

Publication number
CN115082801A
CN115082801A CN202210888641.XA CN202210888641A CN115082801A CN 115082801 A CN115082801 A CN 115082801A CN 202210888641 A CN202210888641 A CN 202210888641A CN 115082801 A CN115082801 A CN 115082801A
Authority
CN
China
Prior art keywords
layer
target
fine
airplane
bilinear
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202210888641.XA
Other languages
Chinese (zh)
Other versions
CN115082801B (en
Inventor
杨澜
杨晓冬
刘建明
严华
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Daoda Tianji Technology Co ltd
Original Assignee
Beijing Daoda Tianji Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Daoda Tianji Technology Co ltd filed Critical Beijing Daoda Tianji Technology Co ltd
Priority to CN202210888641.XA priority Critical patent/CN115082801B/en
Publication of CN115082801A publication Critical patent/CN115082801A/en
Application granted granted Critical
Publication of CN115082801B publication Critical patent/CN115082801B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • G06V10/765Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects using rules for classification or partitioning the feature space
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/08Detecting or categorising vehicles

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Multimedia (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Medical Informatics (AREA)
  • Databases & Information Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • General Engineering & Computer Science (AREA)
  • Molecular Biology (AREA)
  • Data Mining & Analysis (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to an airplane model identification system and method based on remote sensing images, which comprises the following steps: the target detection module is used for detecting whether an airplane target exists in the remote sensing image, and if the airplane target exists, the detected airplane target is sent to the fine-grained classification module; and the model identification module is used for predicting the model of the airplane target based on the fine-grained classification network. The invention constructs a combined framework of a target detection module and a bilinear pooling fine-grained classification network based on the void convolution, and the framework is suitable for identifying the type of an airplane target in a remote sensing image and can also be suitable for identifying the targets with the same size and similar appearance characteristics in images in other fields.

Description

Airplane model identification system and method based on remote sensing image
Technical Field
The invention relates to the technical field of target identification, in particular to an airplane model identification system and method based on remote sensing images.
Background
The currently mainstream neural network architecture based on the deep learning target detection algorithm is divided into a one-stage network model and a two-stage network model. The first-stage network model directly regresses the class probability and the position coordinates of the object target, and compared with the second-stage model, the speed is higher, but the precision is lower than that of the second-stage network model. The two-stage network model mainly completes the target detection process through a convolutional neural network, CNN convolutional characteristics are extracted, when the network is trained, the network is mainly trained in two parts, the first step is to train the RPN network, the second step is to train the network detected in a target area, the accuracy of the network is high, but the speed is slow compared with that of the network model in the first stage.
Most remote sensing image airplane target identification methods are in a stage of identifying whether a certain target is an airplane target or not, and further classification and identification of airplane target models are lacked. Limited by the influence of high-resolution, more complex background and interference factors which are difficult to distinguish compared with the airplane target appear in the remote sensing image, and particularly, although some airplane models have certain differences, the common features of the airplane models are obvious and very similar, especially in the image with lower resolution. Therefore, no matter the structure of the one-stage network model or the two-stage network model can not accurately and finely identify two different types of airplanes with similar spatial structures, and the practical precision can not be achieved.
Disclosure of Invention
The invention aims to improve the identification precision of the specific model of the airplane target through a combined framework of a target detection module and a fine-grained classification module, and provides an airplane model identification system and method based on a remote sensing image.
In order to achieve the above object, the embodiments of the present invention provide the following technical solutions:
an aircraft model identification system based on remote sensing images, comprising:
the target detection module is used for detecting whether an airplane target exists in the remote sensing image, and if the airplane target exists, the detected airplane target is sent to the model identification module;
and the model identification module is used for predicting the model of the airplane target based on the fine-grained classification network.
The fine-grained classification network comprises 3 neural network units and 1 bilinear output unit which are connected in sequence, wherein the input end of the 1 st neural network unit is connected with the output end of the target detection module, and the output end of the 3 rd neural network unit is connected with the input end of the bilinear output unit;
each neural network unit comprises a first residual cavity rolling block, a second residual cavity rolling block, a first rolling layer, a second rolling layer, a full connection layer, a BN layer, a third rolling layer and a first activation function layer; the first residual cavity rolling block, the second residual cavity rolling block and the first rolling layer are sequentially connected, the output end of the first rolling layer and the output end of the second rolling layer are respectively connected with the input end of the full-connection layer, and the full-connection layer, the BN layer, the third rolling layer and the first activation function layer are sequentially connected.
The first residual cavity convolution block or the second residual cavity convolution block comprises a cavity convolution layer, a BN layer and a second activation function layer which are sequentially connected.
The bilinear output unit is used for performing transposition operation on the high-dimensional feature matrix output by the 3 rd neural network unit, and performing outer product on the high-dimensional feature matrix before the transposition operation and the high-dimensional feature matrix after the transposition operation to obtain the fused bilinear feature.
The classification loss function of the fine-grained classification network is as follows:
Figure 228543DEST_PATH_IMAGE002
wherein x is i Representing input samples, namely the ith aircraft target of the fine-grained classification network, wherein N represents the total number of the input samples, and i belongs to N; y is i Representing a label truth value, namely inputting a model type real label of the ith aircraft target of the fine-grained classification network; j represents the jth model type, M represents the total number of the model types, and j belongs to M; y is ci Is thin and thinThe prediction output of the granularity classification network indicates that the ith aircraft target belongs to the c model class, and c belongs to M; m represents the distance interval between airplane targets of different types and classes input into the fine-grained classification network;
Figure 693153DEST_PATH_IMAGE003
denotes y i The cosine angle of (d);
Figure DEST_PATH_IMAGE004
prediction output y representing a fine-grained classification network ci And input sample x i The included angle therebetween.
An airplane model identification method based on remote sensing images comprises the following steps:
step S1, the target detection module detects whether an airplane target exists in the remote sensing image, and if the airplane target exists, the detected airplane target is sent to the model identification module;
and step S2, the model identification module predicts the model of the airplane target based on the fine-grained classification network.
The model identification module predicts the model of the airplane target based on the fine-grained classification network, and comprises the following steps:
the fine-grained classification network carries out feature extraction on the airplane target, outputs bilinear features of the airplane target, and predicts the model of the airplane target according to the bilinear features;
the fine-grained classification network comprises 3 neural network units and 1 bilinear output unit which are connected in sequence; the input end of the 1 st neural network unit is connected with the output end of the target detection module, and the output end of the 3 rd neural network unit is connected with the input end of the bilinear output unit; the 3 neural network units output a high-dimensional characteristic matrix of the airplane target, and the bilinear output unit outputs bilinear characteristics of the airplane target.
Each neural network unit comprises a first residual cavity rolling block, a second residual cavity rolling block, a first rolling layer, a second rolling layer, a full connection layer, a BN layer, a third rolling layer and a first activation function layer; the first residual cavity rolling block, the second residual cavity rolling block and the first rolling layer are sequentially connected, the output end of the first rolling layer and the output end of the second rolling layer are respectively connected with the input end of the full-connection layer, and the full-connection layer, the BN layer, the third rolling layer and the first activation function layer are sequentially connected.
The bilinear output unit outputs bilinear characteristics of the aircraft target, and the bilinear output unit comprises the following steps: and the bilinear output unit transposes the high-dimensional characteristic matrix output by the 3 rd neural network unit, and then performs outer product on the high-dimensional characteristic matrix before the transposing operation and the high-dimensional characteristic matrix after the transposing operation to obtain the fused bilinear characteristic.
The classification loss function of the fine-grained classification network is as follows:
Figure 846792DEST_PATH_IMAGE002
wherein x is i Representing input samples, namely the ith aircraft target of the fine-grained classification network, wherein N represents the total number of the input samples, and i belongs to N; y is i Representing a label truth value, namely inputting a model type real label of the ith aircraft target of the fine-grained classification network; j represents the jth model type, M represents the total number of the model types, and j belongs to M; y is ci For the prediction output of the fine-grained classification network, the ith aircraft target belongs to the category of the type c, and c belongs to M; m represents the distance interval between airplane targets of different types and classes input into the fine-grained classification network;
Figure 569897DEST_PATH_IMAGE003
denotes y i The cosine angle of (d);
Figure 400581DEST_PATH_IMAGE004
prediction output y representing a fine-grained classification network ci And input sample x i The included angle therebetween.
Compared with the prior art, the invention has the beneficial effects that:
(1) the invention constructs a combined framework of a target detection module and a bilinear pooling fine-grained classification network based on the void convolution, and the framework is suitable for identifying the type of an airplane target in a remote sensing image and can also be suitable for identifying the targets with the same size and similar appearance characteristics in images in other fields.
(2) The invention extracts a fine-grained classification network structure based on the bilinear feature of the cavity convolution, uses a neural network unit for feature sampling formed by combining common convolution and the cavity convolution and uses residual connection, thereby not only reducing the parameter quantity, but also avoiding the loss of key data, and an output end transposes an extracted high-dimensional feature matrix and outputs a feature calculation result after outer products are fused.
(3) The invention adds cosine angle to enlarge classification boundary, increase class spacing and aggregate class inner distance aiming at the loss function with small class distance of fine-grained classification network, thereby achieving the identification function of airplane target model, rather than simple classification.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings needed to be used in the embodiments will be briefly described below, it should be understood that the following drawings only illustrate some embodiments of the present invention and therefore should not be considered as limiting the scope, and for those skilled in the art, other related drawings can be obtained according to the drawings without inventive efforts.
FIG. 1 is a block diagram of an aircraft model identification system according to the present invention;
fig. 2 is a schematic structural diagram of a fine-grained classification network according to an embodiment of the present invention;
FIG. 3 is a schematic structural diagram of a neural network unit according to an embodiment of the present invention;
fig. 4 is a schematic structural diagram of a residual hole convolution block according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. The components of embodiments of the present invention generally described and illustrated in the figures herein may be arranged and designed in a wide variety of different configurations. Thus, the following detailed description of the embodiments of the present invention, presented in the figures, is not intended to limit the scope of the invention, as claimed, but is merely representative of selected embodiments of the invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments of the present invention without making any creative effort, shall fall within the protection scope of the present invention.
It should be noted that: like reference numbers and letters refer to like items in the following figures, and thus, once an item is defined in one figure, it need not be further defined and explained in subsequent figures. Also, in the description of the present invention, the terms "first", "second", and the like are used for distinguishing between descriptions and not necessarily for describing a relative importance or implying any actual relationship or order between such entities or operations.
Example (b):
the invention is realized by the following technical scheme, as shown in figure 1, the airplane model identification system based on the remote sensing image comprises a target detection module and a model identification module. The target detection module is used for detecting whether an airplane target exists in the input remote sensing image, and if the airplane target exists, the detected airplane target is sent to the model identification module. The model identification module is used for predicting the model of the airplane target based on the fine-grained classification network.
The remote sensing image outputs single-class target detection in the field of a deep learning image visual algorithm, a target detection module of the scheme is based on a Yolov5 algorithm, the Yolov5 algorithm belongs to a one-stage network model (one-stage), the performance is excellent, the precision is high, and the small target detection precision is particularly improved, so that the method is very suitable for remote sensing image aircraft target detection. Referring to fig. 1, the output of the object detection module includes coordinate information [ xmin, ymin, xmax, ymax ] and category information [ class, conf ] (class represents classification, conf represents confidence), and since the object detection module is only used for detecting whether an airplane object exists in the remote sensing image, only the coordinate information [ xmin, ymin, xmax, ymax ] is focused here. Wherein xmin and ymin represent the coordinates of the left lower vertex of the frame where the airplane target is located, and xmax and ymax represent the coordinates of the right upper vertex of the frame where the airplane target is located.
According to the obtained coordinate information, cutting the original remote sensing image to obtain the area where the airplane target is located, and sending the airplane target after cutting to a fine-grained classification network.
In the remote sensing image, because the difference existing between the appearances of different types of airplanes is small, the airplane appearance is very similar due to the characteristic of obvious commonality and is particularly easily influenced by the spatial resolution. The normal convolution reduces the image resolution and the spatial hierarchy information is lost during the downsampling (pooling) process. The size of an airplane target in a remote sensing image is usually small, and if ordinary convolution is used for sampling learning, the feature effective information extracted after multilayer convolution can be greatly compressed. And the space between the airplane target classes is small, and the general image classification task is difficult to finely identify the specific model of the airplane target.
Therefore, the fine-grained classification network used by the model identification module is a bilinear pooling fine-grained classification network based on the hole convolution. The hole convolution is a method of expanding the field of view by injecting a hole in the normal convolution, and has a parameter expansion rate (ratio) that defines the distance between values when the convolution kernel processes data, compared to the normal convolution.
Referring to fig. 2, the structure of the fine-grained classification network includes 3 neural network units and 1 bilinear output unit, where the 3 neural network units are a first neural network unit, a second neural network unit, and a third neural network unit, and the structures of the neural network units are the same. The first neural network unit, the second neural network unit and the third neural network unit are sequentially connected, and the output end of the third neural network unit is connected with the input end of the bilinear output unit.
Please refer to fig. 3, which shows a structure of a neural network unit, including a first residual void volume block, a second residual void volume block, a first volume layer, a second volume layer, a full link layer, a BN layer, a third volume layer, and a first activation function layer, where the first residual void volume block and the second residual void volume block have the same structure. The first residual cavity rolling block, the second residual cavity rolling block and the first rolling layer are sequentially connected, the output end of the first rolling layer and the output end of the second rolling layer are respectively connected with the input end of the full-connection layer, and the full-connection layer, the BN layer, the third rolling layer and the first activation function layer are sequentially connected.
Referring to fig. 4, the structure of the residual void volume block includes a void volume layer, a BN layer, and a second activation function layer, which are connected in sequence. The cavity convolution layer is used for down sampling, sliding compensation is set to be 2, and texture features are better reserved instead of Pooling.
Referring to fig. 2, the bilinear output unit is configured to perform a transpose operation on the high-dimensional feature matrix output by the 3 rd neural network unit, and perform an outer product on the high-dimensional feature matrix before the transpose operation and the high-dimensional feature matrix after the transpose operation to obtain a fused bilinear feature: x (I) = A (I) T Wherein X (I) is bilinear feature, A (I) is high-dimensional feature matrix before transposition operation, A (I) T And the high-dimensional feature matrix after the transposition operation is performed.
And high-order feature representation is obtained after square root and two-norm normalization operation, and bilinear pooling provides stronger feature representation than a linear model, so that the classification accuracy of the high-order feature representation is far higher than that of a common one-stage network model.
Before the fine-grained classification network is used, the fine-grained classification network needs to be trained, and the classification loss function of the fine-grained classification network is as follows:
Figure 285360DEST_PATH_IMAGE002
wherein x is i Representing input samples, namely the ith aircraft target of the fine-grained classification network, wherein N represents the total number of the input samples, and i belongs to N; y is i Representing a label truth value, namely inputting a model type real label of the ith aircraft target of the fine-grained classification network; j represents the jth model type, M represents the total number of the model types, and j belongs to M; y is ci For the prediction output of the fine-grained classification network, the ith aircraft target belongs to the category of the type c, and c belongs to M; m representsInputting distance intervals among airplane targets of different types and classes of the fine-grained classification network;
Figure 457453DEST_PATH_IMAGE003
denotes y i The cosine angle of (d);
Figure 797299DEST_PATH_IMAGE004
prediction output y representing a fine-grained classification network ci And input sample x i The included angle therebetween.
Based on the system, please refer to fig. 1, the scheme also provides an airplane model identification method based on remote sensing images, which comprises the following steps:
and step S1, the target detection module detects whether the remote sensing image has an airplane target, and if so, the detected airplane target is sent to the model identification module.
The size of the remote sensing image is usually far larger than that of the natural scene image, if the remote sensing image is directly sent to a target detection module, the airplane target can be seriously compressed, and particularly, the network of the target detection module is difficult to learn the characteristic information of the small target such as the airplane. Therefore, the cutting pretreatment needs to be carried out on the remote sensing image with the super-large width, and in consideration of the problem that one target is cut off, the scheme sets a certain size and carries out sliding cutting at a certain step length to obtain N small pictures. And setting an overlap area between two adjacent small images, and inputting the cut remote sensing image small images into a target detection module.
The method comprises the steps that after image preprocessing is carried out on a remote sensing image, N small images are generated and sequentially input into a target detection module, coordinate information [ xmin, ymin, xmax, ymax ] of an airplane target is obtained through prediction of the small images through the target detection module, and the airplane target is obtained according to a coordinate information slice.
And step S2, the model identification module predicts the model of the airplane target based on the fine-grained classification network.
The input image size of the fine-grained classification network is set to be 112 x 112, and in order to ensure that the airplane target is not stretched and deformed after being zoomed, the airplane target is subjected to filling zooming processing before being input into the fine-grained classification network. Specifically, a base image with a pixel value of 128 is generated by taking the maximum side W/H (length or width) of an image of an airplane target, then the image of the airplane target is attached to the base image, an equilateral (W = H) square image is generated, and when the size is scaled to 112 × 112, the airplane shape is guaranteed to be scaled in an equal ratio.
And the fine-grained classification network extracts the characteristics of the airplane target, outputs bilinear characteristics of the airplane target, and predicts the model of the airplane target according to the bilinear characteristics.
The fine-grained classification network comprises 3 neural network units and 1 bilinear output unit which are connected in sequence; the input end of the 1 st neural network unit is connected with the output end of the target detection module, and the output end of the 3 rd neural network unit is connected with the input end of the bilinear output unit; the 3 neural network units output a high-dimensional characteristic matrix of the airplane target, and the bilinear output unit outputs bilinear characteristics of the airplane target.
Each neural network unit comprises a first residual cavity rolling block, a second residual cavity rolling block, a first rolling layer, a second rolling layer, a full connection layer, a BN layer, a third rolling layer and a first activation function layer; the first residual cavity rolling block, the second residual cavity rolling block and the first rolling layer are sequentially connected, the output end of the first rolling layer and the output end of the second rolling layer are respectively connected with the input end of the full-connection layer, and the full-connection layer, the BN layer, the third rolling layer and the first activation function layer are sequentially connected.
The step of outputting bilinear characteristics of the aircraft target by the bilinear output unit comprises the following steps: and the bilinear output unit transposes the high-dimensional feature matrix output by the 3 rd neural network unit, and then performs outer product on the high-dimensional feature matrix before the transposing operation and the high-dimensional feature matrix after the transposing operation to obtain the fused bilinear feature.
The classification loss function of the fine-grained classification network is as follows:
Figure 574500DEST_PATH_IMAGE002
wherein x is i Representing input samples, namely the ith aircraft target of the fine-grained classification network, wherein N represents the total number of the input samples, and i belongs to N; y is i Representing a label truth value, namely inputting a model type real label of the ith aircraft target of the fine-grained classification network; j represents the jth model type, M represents the total number of the model types, and j belongs to M; y is ci For the prediction output of the fine-grained classification network, the ith aircraft target belongs to the category of the type c, and c belongs to M; m represents the distance interval between airplane targets of different types and classes input into the fine-grained classification network;
Figure 115334DEST_PATH_IMAGE003
denotes y i The cosine angle of (d);
Figure 666401DEST_PATH_IMAGE004
prediction output y representing a fine-grained classification network ci And input sample x i The included angle therebetween.
The above description is only for the specific embodiments of the present invention, but the scope of the present invention is not limited thereto, and any person skilled in the art can easily conceive of the changes or substitutions within the technical scope of the present invention, and all the changes or substitutions should be covered within the scope of the present invention. Therefore, the protection scope of the present invention shall be subject to the protection scope of the claims.

Claims (10)

1. An aircraft model identification system based on remote sensing image which characterized in that: the method comprises the following steps:
the target detection module is used for detecting whether an airplane target exists in the remote sensing image, and if the airplane target exists, the detected airplane target is sent to the model identification module;
and the model identification module is used for predicting the model of the airplane target based on the fine-grained classification network.
2. An aircraft model identification system based on remote sensing images according to claim 1, characterized in that: the fine-grained classification network comprises 3 neural network units and 1 bilinear output unit which are connected in sequence, wherein the input end of the 1 st neural network unit is connected with the output end of the target detection module, and the output end of the 3 rd neural network unit is connected with the input end of the bilinear output unit;
each neural network unit comprises a first residual cavity convolution block, a second residual cavity convolution block, a first convolution layer, a second convolution layer, a full connection layer, a BN layer, a third convolution layer and a first activation function layer; the first residual cavity rolling block, the second residual cavity rolling block and the first rolling layer are sequentially connected, the output end of the first rolling layer and the output end of the second rolling layer are respectively connected with the input end of the full-connection layer, and the full-connection layer, the BN layer, the third rolling layer and the first activation function layer are sequentially connected.
3. An aircraft model identification system based on remote sensing images according to claim 2, characterized in that: the first residual cavity convolution block or the second residual cavity convolution block comprises a cavity convolution layer, a BN layer and a second activation function layer which are connected in sequence.
4. An aircraft model identification system based on remote sensing images according to claim 2, characterized in that: the bilinear output unit is used for performing transposition operation on the high-dimensional feature matrix output by the 3 rd neural network unit, and performing outer product on the high-dimensional feature matrix before the transposition operation and the high-dimensional feature matrix after the transposition operation to obtain the fused bilinear feature.
5. An aircraft model identification system based on remote sensing images according to claim 1, characterized in that: the classification loss function of the fine-grained classification network is as follows:
Figure 155732DEST_PATH_IMAGE002
wherein x is i Representing input samples, namely the ith aircraft target of the fine-grained classification network, wherein N represents the total number of the input samples, and i belongs to N; y is i Representing a label truth value, namely inputting a model type real label of the ith aircraft target of the fine-grained classification network; j represents the jth model type, M represents the total number of the model types, and j belongs to M; y is ci For the prediction output of the fine-grained classification network, the ith aircraft target belongs to the category of the type c, and c belongs to M; m represents the distance interval between airplane targets of different types and classes input into the fine-grained classification network;
Figure 757746DEST_PATH_IMAGE003
denotes y i The cosine angle of (d);
Figure 984328DEST_PATH_IMAGE004
prediction output y representing a fine-grained classification network ci And input sample x i The included angle therebetween.
6. An airplane model identification method based on remote sensing images is characterized in that: the method comprises the following steps:
step S1, the target detection module detects whether an airplane target exists in the remote sensing image, and if the airplane target exists, the detected airplane target is sent to the model identification module;
and step S2, the model identification module predicts the model of the airplane target based on the fine-grained classification network.
7. The aircraft model identification method based on the remote sensing image as claimed in claim 6, wherein: the model identification module predicts the model of the airplane target based on the fine-grained classification network, and comprises the following steps:
the fine-grained classification network carries out feature extraction on the airplane target, outputs bilinear features of the airplane target, and predicts the model of the airplane target according to the bilinear features;
the fine-grained classification network comprises 3 neural network units and 1 bilinear output unit which are connected in sequence; the input end of the 1 st neural network unit is connected with the output end of the target detection module, and the output end of the 3 rd neural network unit is connected with the input end of the bilinear output unit; the 3 neural network units output a high-dimensional characteristic matrix of the airplane target, and the bilinear output unit outputs bilinear characteristics of the airplane target.
8. The aircraft model identification method based on the remote sensing image as claimed in claim 7, wherein: each neural network unit comprises a first residual cavity rolling block, a second residual cavity rolling block, a first rolling layer, a second rolling layer, a full connection layer, a BN layer, a third rolling layer and a first activation function layer; the first residual cavity convolution block, the second residual cavity convolution block and the first convolution layer are connected in sequence, the output end of the first convolution layer and the output end of the second convolution layer are connected with the input end of the full connection layer respectively, and the full connection layer, the BN layer, the third convolution layer and the first activation function layer are connected in sequence.
9. The aircraft model identification method based on the remote sensing image as claimed in claim 7, characterized in that: the step of outputting bilinear characteristics of the aircraft target by the bilinear output unit comprises the following steps: and the bilinear output unit transposes the high-dimensional feature matrix output by the 3 rd neural network unit, and then performs outer product on the high-dimensional feature matrix before the transposing operation and the high-dimensional feature matrix after the transposing operation to obtain the fused bilinear feature.
10. The aircraft model identification method based on the remote sensing image as claimed in claim 6, wherein: the classification loss function of the fine-grained classification network is as follows:
Figure 560975DEST_PATH_IMAGE002
wherein x is i Representing input samples, i.e. inputEntering an ith airplane target of the fine-grained classification network, wherein N represents the total number of input samples, and i belongs to N; y is i Representing a label truth value, namely inputting a model type real label of the ith aircraft target of the fine-grained classification network; j represents the jth model type, M represents the total number of the model types, and j belongs to M; y is ci For the prediction output of the fine-grained classification network, the ith aircraft target belongs to the category of the type c, and c belongs to M; m represents the distance interval between airplane targets of different types and classes input into the fine-grained classification network;
Figure 695153DEST_PATH_IMAGE003
denotes y i The cosine angle of (d);
Figure 886094DEST_PATH_IMAGE004
prediction output y representing a fine-grained classification network ci And input sample x i The included angle therebetween.
CN202210888641.XA 2022-07-27 2022-07-27 Airplane model identification system and method based on remote sensing image Active CN115082801B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210888641.XA CN115082801B (en) 2022-07-27 2022-07-27 Airplane model identification system and method based on remote sensing image

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210888641.XA CN115082801B (en) 2022-07-27 2022-07-27 Airplane model identification system and method based on remote sensing image

Publications (2)

Publication Number Publication Date
CN115082801A true CN115082801A (en) 2022-09-20
CN115082801B CN115082801B (en) 2022-10-25

Family

ID=83242402

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210888641.XA Active CN115082801B (en) 2022-07-27 2022-07-27 Airplane model identification system and method based on remote sensing image

Country Status (1)

Country Link
CN (1) CN115082801B (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109086792A (en) * 2018-06-26 2018-12-25 上海理工大学 Based on the fine granularity image classification method for detecting and identifying the network architecture
WO2021129691A1 (en) * 2019-12-23 2021-07-01 长沙智能驾驶研究院有限公司 Target detection method and corresponding device
CN113989662A (en) * 2021-10-18 2022-01-28 中国电子科技集团公司第五十二研究所 Remote sensing image fine-grained target identification method based on self-supervision mechanism
CN114140683A (en) * 2020-08-12 2022-03-04 天津大学 Aerial image target detection method, equipment and medium

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109086792A (en) * 2018-06-26 2018-12-25 上海理工大学 Based on the fine granularity image classification method for detecting and identifying the network architecture
WO2021129691A1 (en) * 2019-12-23 2021-07-01 长沙智能驾驶研究院有限公司 Target detection method and corresponding device
CN114140683A (en) * 2020-08-12 2022-03-04 天津大学 Aerial image target detection method, equipment and medium
CN113989662A (en) * 2021-10-18 2022-01-28 中国电子科技集团公司第五十二研究所 Remote sensing image fine-grained target identification method based on self-supervision mechanism

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
WEI LIANG 等: "FGATR-Net: Automatic Network Architecture Design for Fine-Grained Aircraft Type Recognition in Remote Sensing Images", 《REMOTE SENSING》 *
曾妍庆: "基于遥感图像飞行器目标的细分类方法研究", 《中国优秀硕士学位论文全文数据库 工程科技Ⅱ辑》 *

Also Published As

Publication number Publication date
CN115082801B (en) 2022-10-25

Similar Documents

Publication Publication Date Title
CN112381097A (en) Scene semantic segmentation method based on deep learning
CN112381004B (en) Dual-flow self-adaptive graph rolling network behavior recognition method based on framework
CN113673510B (en) Target detection method combining feature point and anchor frame joint prediction and regression
Li et al. Sign language recognition based on computer vision
CN112087443B (en) Sensing data anomaly detection method under physical attack of industrial sensing network information
CN112766229B (en) Human face point cloud image intelligent identification system and method based on attention mechanism
CN111652171B (en) Construction method of facial expression recognition model based on double branch network
CN111401149B (en) Lightweight video behavior identification method based on long-short-term time domain modeling algorithm
CN114332473A (en) Object detection method, object detection device, computer equipment, storage medium and program product
Yang et al. HCNN-PSI: A hybrid CNN with partial semantic information for space target recognition
CN113298817A (en) High-accuracy semantic segmentation method for remote sensing image
CN113283409A (en) Airplane detection method in aerial image based on EfficientDet and Transformer
Bin et al. Tensor-based approach for liquefied natural gas leakage detection from surveillance thermal cameras: A feasibility study in rural areas
CN117237547B (en) Image reconstruction method, reconstruction model processing method and device
CN114881286A (en) Short-time rainfall prediction method based on deep learning
CN115082801B (en) Airplane model identification system and method based on remote sensing image
CN111626197B (en) Recognition method based on human behavior recognition network model
CN110738123B (en) Method and device for identifying densely displayed commodities
CN117368862A (en) High-efficiency weather radar data quality evaluation system
CN117036266A (en) Industrial image anomaly detection method and system based on knowledge distillation
CN115048873B (en) Residual service life prediction system for aircraft engine
Fang et al. STUNNER: Radar echo extrapolation model based on spatiotemporal fusion neural network
CN116403286A (en) Social grouping method for large-scene video
CN114462491A (en) Behavior analysis model training method, behavior analysis method and equipment thereof
Lin et al. SOPNet method for the fine-grained measurement and prediction of precipitation intensity using outdoor surveillance cameras

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant