CN111563414B - SAR image ship target detection method based on non-local feature enhancement - Google Patents

SAR image ship target detection method based on non-local feature enhancement Download PDF

Info

Publication number
CN111563414B
CN111563414B CN202010267019.8A CN202010267019A CN111563414B CN 111563414 B CN111563414 B CN 111563414B CN 202010267019 A CN202010267019 A CN 202010267019A CN 111563414 B CN111563414 B CN 111563414B
Authority
CN
China
Prior art keywords
feature
network
sar image
feature map
ship
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010267019.8A
Other languages
Chinese (zh)
Other versions
CN111563414A (en
Inventor
耿杰
徐哲
蒋雯
邓鑫洋
黄凯
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Northwestern Polytechnical University
Original Assignee
Northwestern Polytechnical University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Northwestern Polytechnical University filed Critical Northwestern Polytechnical University
Priority to CN202010267019.8A priority Critical patent/CN111563414B/en
Publication of CN111563414A publication Critical patent/CN111563414A/en
Application granted granted Critical
Publication of CN111563414B publication Critical patent/CN111563414B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • G06V20/13Satellite images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/08Detecting or categorising vehicles

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Astronomy & Astrophysics (AREA)
  • Remote Sensing (AREA)
  • Multimedia (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses a non-local feature enhancement-based SAR image ship target detection method, which comprises the following steps of: inputting a ship SAR image data set, and converting the data set into a PASCAL VOC format; constructing a convolutional neural network ResNet, and training the ResNet network by using the manufactured ship SAR image data set; fusing the feature graphs output by ResNet by using a feature pyramid network to obtain a fused feature graph; grouping, scaling and averaging the fused feature graphs to obtain an average feature graph; performing feature enhancement on the average feature map by using a non-local feature enhancement model; carrying out scale transformation on the enhanced feature map; and performing regression prediction on the enhanced feature map by using a full convolution network to obtain a final detection result. The invention provides a non-local feature enhancement method, which not only effectively solves the problem of detecting the SAR image ship small target, but also can reduce the influence of the SAR image noise of the ship target on the detection.

Description

SAR image ship target detection method based on non-local feature enhancement
Technical Field
The invention belongs to the field of intelligent interpretation of remote sensing images, and particularly relates to a non-local feature enhancement-based SAR image ship target detection method.
Background
Synthetic Aperture Radar (SAR) is an active microwave remote sensor, can perform high-resolution imaging on various targets, can work all day long and all weather, has the advantages of strong penetration capability, strong anti-interference performance, long acting distance and the like, and is widely applied to the fields of military reconnaissance, marine application, agriculture and forestry monitoring and the like. With the development of synthetic aperture radar imaging technology, the SAR image target detection has been widely applied to military and civil fields, and the SAR image ship target detection is one of the hot spot problems.
The Constant False Alarm Rate (CFAR) detection algorithm is a traditional SAR image ship target detection method, and the ship target is detected by carrying out statistical distribution modeling on background clutter. But due to the complexity of modeling and long calculation time, ideal results are often difficult to obtain for ship detection with complex background. With the development of artificial intelligence technology, deep learning models are gradually used for target detection tasks. Target detection algorithms based on a convolutional neural network can be divided into two categories, one category is a double-stage target detection algorithm, such as fast R-CNN, Mask R-CNN and the like, detection is divided into two stages of candidate region extraction, candidate target category and position prediction by the algorithm, the other category is a single-stage target detection algorithm, such as You Only Look one (Yolov2), Ret inaNet and the like, the algorithm directly predicts the category and position information of a target in a regression mode, compared with the double-stage target detection algorithm, the detection efficiency is greatly improved, and an end-to-end target detection task is realized.
However, existing convolutional neural network-based object detection algorithms are mainly applied to optical image data. The SAR image ship target detection technology based on the convolutional neural network has made great progress in the last two years, but still has some problems, which are mainly shown in the following: (1) the size of a ship target of the low-resolution SAR image is small, and the position of the ship target in the deep characteristic map based on the convolutional neural network is rough, so that the ship target is difficult to locate and even missed to detect; (2) the SAR image of the ship target has the problems of complex background and large noise, the difficulty of distinguishing the target from the background is increased, and the detection effect of the ship target is influenced.
Disclosure of Invention
Aiming at the technical problems, the invention provides the SAR image ship target detection method based on non-local feature enhancement, which solves the problem of unbalance among feature layers, enhances the robustness of features, improves the detection capability of small-size ship targets and improves the precision of ship target detection.
The technical method adopted by the invention is as follows: a non-local feature enhancement-based SAR image ship target detection method is characterized by comprising the following steps:
step one, inputting a ship SAR image data set, and converting the data set into a PASCAL VOC format:
step 101, labeling a ship SAR image data set by using open source software labelImg;
102, respectively putting the annotation file and the image file into different folders to manufacture a standard Pascal VOC data set format;
103, dividing the manufactured ship SAR image data set into a training set, a testing set and a verification set according to the proportion of 7:2:1, and inputting the training set, the testing set and the verification set into an SAR image ship target detection model;
step two, constructing a convolutional neural network ResNet, and training the ResNet network by using the manufactured ship SAR image data set:
step 201, constructing a ResNet-50 network as a basic network, wherein the ResNet-50 network comprises 5 convolution modules which are Conv1, Conv2_ x, Conv3_ x, Conv4_ x and Conv5_ x respectively;
step 202, adopting a ResNet-50 network to extract the features of the SAR image, and outputting feature maps of the last 4 convolution modules, wherein the feature maps are respectively C2、C3、C4、C5
Thirdly, fusing the feature graphs output by ResNet by using the feature pyramid network to obtain fused feature graphs:
step 301, for the characteristic diagram C output in step 2022、C3、C4、C5Convolution operations are respectively carried out by adopting convolution kernels with the number being half of the number of the characteristic diagram channels and the size being 1 multiplied by 1, so that characteristic dimension reduction is realized, and a dimension-reduced characteristic diagram C is obtained2′、C3′、C4′、P5′;
Step 302, interpolate P bilinearly5' size extension to and C4' same size, then with C4' element by element addition to P4', same principle, P4Extended to and C by bilinear interpolation3' same size, then with C3' Pixel-by-pixel addition to P3′,P3Extended to and C by bilinear interpolation2' same size, then with C2' Pixel-by-pixel addition to P2′;
Step 303, checking P by respectively adopting convolution of 3 × 32′、P3′、P4′、P5Performing convolution operation to eliminate aliasing effect caused by bilinear interpolation to obtain a fused characteristic diagram P2、P3、P4、P5
Step four, grouping, scaling and averaging the fused feature graphs to obtain an average feature graph:
step 401, fusing the feature map P5Performing bilinear interpolation expansion to sum P4Same size, to give P5", mixing P with2Performing a spatial pyramid pooling reduction to sum P3Same size, to give P2″;
Step 402, averaging the two groups of feature maps with the re-scaled sizes to obtain an average feature map:
Figure GDA0003458321360000031
Figure GDA0003458321360000032
wherein, Pa、PbRespectively representing the average features of the two groups of feature mapsFigure representation;
fifthly, performing feature enhancement on the average feature map by using a non-local feature enhancement model:
step 501, average feature map PaPerforming three groups of convolution operations by adopting convolution kernels with the number being half of the average feature map channel number and the size being 1 multiplied by 1 respectively to realize feature dimensionality reduction, and expanding the dimensionality of the feature kernels respectively to obtain three feature matrixes:
θ(pi)=Wθpi
φ(pj)=Wφpj
g(pj)=Wgpj
wherein p isiAnd pjRespectively representing the information of the ith and jth positions on the average profile, Wθ、WφAnd WgRespectively representing the weight matrixes of the corresponding convolution kernels;
step 502, calculating a feature matrix theta (p)i) And phi (p)j) To obtain a similarity matrix f (p)i,pj) Comprises the following steps:
Figure GDA0003458321360000041
step 503, match the similarity matrix f (p)i,pj) And a feature matrix g (p)j) Carrying out matrix multiplication to obtain an enhanced feature matrix yiComprises the following steps:
Figure GDA0003458321360000042
step 504, the enhanced feature matrix y is checked by using the convolution with the number of average feature map channels and the size of 1 multiplied by 1iPerforming convolution operation to realize feature dimension increasing, and then performing convolution operation with the average feature map PaAdding element by element to enhance original characteristics to obtain enhanced characteristic diagram D3
Step 505, the average feature map P is processed according to steps 501 to 504bIs treated to obtainStrong feature map D4
Step six, carrying out scale transformation on the enhanced feature graph:
step 601, enhancing the feature map D3Obtaining D by size expansion after bilinear interpolation2
Step 602, enhance feature map D4The size is reduced after the spatial pyramid pooling to obtain D5
Step 603, enhanced feature map D of scale transformation2、D3、D4、D5Respectively replacing the feature map P of step 3032、P3、P4、P5
Step 604, for the enhanced feature map D5The size is reduced after the space pyramid pooling is carried out, and an enhanced feature map D is obtained6
Seventhly, carrying out regression prediction on the enhanced feature graph by using a full convolution network to obtain a final detection result:
step 701, for the enhanced feature map D2、D3、D4、D5、D6Respectively inputting the data into a frame regression sub-network and a classification sub-network;
loss function L of bounding box regression subnetwork trainingregThe formula is as follows:
Figure GDA0003458321360000051
wherein d represents the difference between the predicted bounding box vector and the true label box vector; loss function L for classification subnetwork trainingclsThe formula is as follows:
Lcls=-αt(1-qt)γlog(qt)
wherein q istProbability value, alpha, representing class predictiontAnd gamma is an over-parameter for adjusting the balance of the difficult and easy samples; thus, the Loss function Loss formula of the whole network training is as follows:
Figure GDA0003458321360000052
wherein N isposRepresenting the number of positive samples, mu and lambda representing balance factors of classification loss and bounding box regression loss, respectively;
step 702, enhance the feature map D2、D3、D4、D5、D6And after respectively passing through the frame regression sub-network and the classification sub-network, obtaining a final prediction frame by using a non-maximum suppression algorithm, and outputting a ship SAR image target detection result.
Compared with the prior art, the invention mainly has the following advantages:
firstly, the invention provides a non-local feature enhancement-based SAR image ship target detection method, which balances semantic information among different feature layers through feature layer scaling, integration, refining and enhanced non-local feature enhancement processing, improves the utilization rate of feature layer information, can reduce the influence of ship target SAR image noise on detection, and further improves the robustness of ship target detection;
secondly, the SAR image feature learning is carried out by adopting the feature pyramid network, the high resolution of the low-level features and the semantic information of the high-level features are fused, so that each layer of features of the network can have rich semantic information, and the detection effect of the SAR image ship small target can be further improved.
The technical solution of the present invention is further described in detail by the accompanying drawings and embodiments.
Drawings
FIG. 1 is a flow chart of the method of the present invention.
FIG. 2 is a schematic structural diagram of a non-local feature enhancement model according to the present invention.
Detailed Description
The method of the present invention will be described in further detail below with reference to the accompanying drawings and embodiments of the invention.
It should be noted that the embodiments and features of the embodiments in the present application may be combined with each other without conflict. The present invention will be described in detail below with reference to the embodiments with reference to the attached drawings.
It is noted that the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of example embodiments according to the present application. As used herein, the singular forms "a", "an" and "the" are intended to include the plural forms as well, and it should be understood that when the terms "comprises" and/or "comprising" are used in this specification, they specify the presence of stated features, steps, operations, devices, components, and/or combinations thereof, unless the context clearly indicates otherwise.
It should be noted that the terms "first," "second," and the like in the description and claims of this application and in the drawings described above are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the application described herein are, for example, capable of operation in sequences other than those illustrated or otherwise described herein. Furthermore, the terms "comprises," "comprising," and "having," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed, but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
Spatially relative terms, such as "above … …," "above … …," "above … …," "above," and the like, may be used herein for ease of description to describe one device or feature's spatial relationship to another device or feature as illustrated in the figures. It will be understood that the spatially relative terms are intended to encompass different orientations of the device in use or operation in addition to the orientation depicted in the figures. For example, if a device in the figures is turned over, devices described as "above" or "on" other devices or configurations would then be oriented "below" or "under" the other devices or configurations. Thus, the exemplary term "above … …" can include both an orientation of "above … …" and "below … …". The device may be otherwise variously oriented (rotated 90 degrees or at other orientations) and the spatially relative descriptors used herein interpreted accordingly.
As shown in fig. 1, the disclosed SSDD ship SAR image data set is taken as an example to illustrate the rationality and effectiveness of the present invention, and includes the following steps:
step one, inputting a ship SAR image data set, and converting the data set into a PASCAL VOC format:
101, marking the ship SAR image data set by adopting an open SSDD ship SAR image data set, totaling 1160 images and adopting open source software labelImg;
102, respectively putting the annotation file and the image file into different folders to manufacture a standard Pascal VOC data set format;
103, dividing the manufactured ship SAR image data set into a training set, a testing set and a verification set according to the proportion of 7:2:1, and inputting the training set, the testing set and the verification set into an SAR image ship target detection model;
step two, constructing a convolutional neural network ResNet, and training the ResNet network by using the manufactured ship SAR image data set:
step 201, constructing a ResNet-50 network as a basic network, which comprises 5 convolution modules, namely Conv1, Conv2_ x, Conv3_ x, Conv4_ x and Conv5_ x, wherein the ship SAR image is subjected to bilinear interpolation, is scaled to 400 × 512 in size and is input into Conv1 after being normalized;
step 202, adopting a ResNet-50 network to extract the features of the SAR image, and outputting feature maps of the last 4 convolution modules, wherein the feature maps are respectively C2、C3、C4、C5In which C is2256 feature maps, C, of 100X 128 dimensions are shown3Denotes a 512-piece feature map, C, of dimensions 50X 644Representing 1024 signatures of size 25X 32, C52048 feature maps of size 13 × 16 are shown;
thirdly, fusing the feature graphs output by ResNet by using the feature pyramid network to obtain fused feature graphs:
step 301, for the characteristic diagram C output in step 2022、C3、C4、C5Respectively adopting 256 convolution kernels with the size of 1 multiplied by 1 to carry out convolution operation, realizing feature dimension reduction and obtaining a dimension-reduced feature map C2′、C3′、C4′、P5', wherein C2' denotes 256 feature maps, C, of size 100X 1283' denotes 256 feature maps, C, of size 50X 644' denotes 256 feature maps of size 25X 32, P5' 256 feature maps with dimensions 13 × 16 are shown;
step 302, interpolate P bilinearly5' size extension to and C4' same size, then with C4' element by element addition to P4', same principle, P4Extended to and C by bilinear interpolation3' same size, then with C3' Pixel-by-pixel addition to P3′,P3Extended to and C by bilinear interpolation2' same size, then with C2' Pixel-by-pixel addition to P2′;
Step 303, checking P by respectively adopting convolution of 3 × 32′、P3′、P4′、P5Performing convolution operation to eliminate aliasing effect caused by bilinear interpolation to obtain a fused characteristic diagram P2、P3、P4、P5In which P is2Representing 256 features of size 100X 128, P3Representing 256 features of size 50X 64, P4256 feature maps, P, of 25X 32 size are shown5256 feature maps of size 13 × 16 are shown;
step four, grouping, scaling and averaging the fused feature graphs to obtain an average feature graph:
step 401, fusing the feature map P5Performing bilinear interpolation expansion to sum P4Same size, to give P5", mixing P with2Performing a spatial pyramid pooling reduction to sum P3Same size, to give P2″;
Step 402, averaging the two groups of feature maps with the re-scaled sizes to obtain an average feature map:
Figure GDA0003458321360000091
Figure GDA0003458321360000092
wherein, Pa、PbRespectively representing the feature maps of two groups of feature maps after averaging, wherein PaRepresenting 256 features of size 50X 64, Pb256 feature maps of 25 × 32 size are shown;
fifthly, performing feature enhancement on the average feature map by using a non-local feature enhancement model:
step 501, average feature map PaRespectively adopting 128 convolution kernels with the size of 1 multiplied by 1 to carry out three groups of convolution operations, realizing feature dimensionality reduction, and respectively expanding the dimensionality of the three groups of convolution kernels to obtain three feature matrixes:
θ(pi)=Wθpi
φ(pj)=Wφpj
g(pj)=Wgpj
wherein p isiAnd pjRespectively representing the information of the ith and jth positions on the average profile, Wθ、WφAnd WgRespectively representing the weight matrixes of the corresponding convolution kernels;
step 502, calculating a feature matrix theta (p)i) And phi (p)j) To obtain a similarity matrix f (p)i,pj) Comprises the following steps:
Figure GDA0003458321360000093
step 503, match the similarity matrix f (p)i,pj) And a feature matrix g (p)j) Carrying out matrix multiplication to obtain an enhanced feature matrix yiComprises the following steps:
Figure GDA0003458321360000094
step 504, check the enhanced feature matrix y with 256 convolution of size 1 × 1iPerforming convolution operation to realize feature dimension increasing, and then performing convolution operation with the average feature map PaAdding element by element to enhance original characteristics to obtain enhanced characteristic diagram D3
Step 505, the average feature map P is processed according to steps 501 to 504bProcessing to obtain enhanced feature map D4
Step six, carrying out scale transformation on the enhanced feature graph:
step 601, enhancing the feature map D3Obtaining D by size expansion after bilinear interpolation2
Step 602, enhance feature map D4The size is reduced after the spatial pyramid pooling to obtain D5
Step 603, enhanced feature map D of scale transformation2、D3、D4、D5Respectively replacing the feature map P of step 3032、P3、P4、P5Wherein D is2256 feature maps, D, of 100X 128 size are shown3256 feature maps of size 50X 64, D4256 feature maps of 25X 32 size, D5256 feature maps of size 13 × 16 are shown;
step 604, for the enhanced feature map D5The size is reduced after the space pyramid pooling is carried out, and an enhanced feature map D is obtained6Wherein D is6256 feature maps of size 7 × 8 are shown;
seventhly, carrying out regression prediction on the enhanced feature graph by using a full convolution network to obtain a final detection result:
step 701, enhancing the feature map D2、D3、D4、D5、D6Respectively inputting the data into a frame regression sub-network and a classification sub-network;
bounding box regression subnetworkLoss function of training LregThe formula is as follows:
Figure GDA0003458321360000101
wherein d represents the difference between the predicted bounding box vector and the true label box vector; loss function L for classification subnetwork trainingclsThe formula is as follows:
Lcls=-αt(1-qt)γlog(qt)
wherein q istProbability value, alpha, representing class predictiontAnd gamma is an over-parameter for adjusting the balance of the difficult and easy samples; thus, the Loss function Loss formula of the whole network training is as follows:
Figure GDA0003458321360000102
wherein N isposRepresenting the number of positive samples, mu and lambda representing balance factors of classification loss and bounding box regression loss, respectively;
step 702, enhance the feature map D2、D3、D4、D5、D6And respectively obtaining a final prediction frame by using a non-maximum suppression algorithm through a frame regression sub-network and a classification sub-network, and outputting a ship SAR image target detection result.
The above embodiments are only examples of the present invention, and are not intended to limit the present invention, and all simple modifications, changes and equivalent structural changes made to the above embodiments according to the technical spirit of the present invention still fall within the protection scope of the technical solution of the present invention.

Claims (1)

1. A non-local feature enhancement-based SAR image ship target detection method is characterized by comprising the following steps:
step one, inputting a ship SAR image data set, and converting the data set into a PASCAL VOC format:
step 101, labeling a ship SAR image data set by using open source software labelImg;
102, respectively putting the annotation file and the image file into different folders to manufacture a standard Pascal VOC data set format;
103, dividing the manufactured ship SAR image data set into a training set, a testing set and a verification set according to the proportion of 7:2:1, and inputting the training set, the testing set and the verification set into an SAR image ship target detection model;
step two, constructing a convolutional neural network ResNet, and training the ResNet network by using the manufactured ship SAR image data set:
step 201, constructing a ResNet-50 network as a basic network, wherein the ResNet-50 network comprises 5 convolution modules which are Conv1, Conv2_ x, Conv3_ x, Conv4_ x and Conv5_ x respectively;
step 202, adopting a ResNet-50 network to extract the features of the SAR image, and outputting feature maps of the last 4 convolution modules, wherein the feature maps are respectively C2、C3、C4、C5
Thirdly, fusing the feature graphs output by ResNet by using the feature pyramid network to obtain fused feature graphs:
step 301, for the characteristic diagram C output in step 2022、C3、C4、C5Convolution operations are respectively carried out by adopting convolution kernels with the number being half of the number of the characteristic diagram channels and the size being 1 multiplied by 1, so that characteristic dimension reduction is realized, and a dimension-reduced characteristic diagram C is obtained2′、C3′、C4′、P5′;
Step 302, interpolate P bilinearly5' size extension to and C4' same size, then with C4' element by element addition to P4', same principle, P4Extended to and C by bilinear interpolation3' same size, then with C3' Pixel-by-pixel addition to P3′,P3Extended to and C by bilinear interpolation2' same size, then with C2' Pixel-by-pixel addition to P2′;
Step 303, checking P by respectively adopting convolution of 3 × 32′、P3′、P4′、P5Performing convolution operation to eliminate aliasing effect caused by bilinear interpolation to obtain a fused characteristic diagram P2、P3、P4、P5
Step four, grouping, scaling and averaging the fused feature graphs to obtain an average feature graph:
step 401, fusing the feature map P5Performing bilinear interpolation expansion to sum P4Same size, to give P5", mixing P with2Performing a spatial pyramid pooling reduction to sum P3Same size, to give P2″;
Step 402, averaging the two groups of feature maps with the re-scaled sizes to obtain an average feature map:
Figure FDA0003458321350000021
Figure FDA0003458321350000022
wherein, Pa、PbRespectively representing the feature maps after the two groups of feature maps are averaged;
fifthly, performing feature enhancement on the average feature map by using a non-local feature enhancement model:
step 501, average feature map PaPerforming three groups of convolution operations by adopting convolution kernels with the number being half of the average feature map channel number and the size being 1 multiplied by 1 respectively to realize feature dimensionality reduction, and expanding the dimensionality of the feature kernels respectively to obtain three feature matrixes:
θ(pi)=Wθpi
φ(pj)=Wφpj
g(pj)=Wgpj
wherein p isiAnd pjRespectively representing the information of the ith and jth positions on the average profile, Wθ、WφAnd WgRespectively representing the weight matrixes of the corresponding convolution kernels;
step 502, calculating a feature matrix theta (p)i) And phi (p)j) To obtain a similarity matrix f (p)i,pj) Comprises the following steps:
Figure FDA0003458321350000023
step 503, match the similarity matrix f (p)i,pj) And a feature matrix g (p)j) Carrying out matrix multiplication to obtain an enhanced feature matrix yiComprises the following steps:
Figure FDA0003458321350000031
step 504, the enhanced feature matrix y is checked by using the convolution with the number of average feature map channels and the size of 1 multiplied by 1iPerforming convolution operation to realize feature dimension increasing, and then performing convolution operation with the average feature map PaAdding element by element to enhance original characteristics to obtain enhanced characteristic diagram D3
Step 505, the average feature map P is processed according to steps 501 to 504bProcessing to obtain enhanced feature map D4
Step six, carrying out scale transformation on the enhanced feature graph:
step 601, enhancing the feature map D3Obtaining D by size expansion after bilinear interpolation2
Step 602, enhance feature map D4The size is reduced after the spatial pyramid pooling to obtain D5
Step 603, enhanced feature map D of scale transformation2、D3、D4、D5Respectively replacing the feature map P of step 3032、P3、P4、P5
Step 604, for the enhanced feature map D5The size is reduced after the space pyramid pooling is carried out, and an enhanced feature map D is obtained6
Seventhly, carrying out regression prediction on the enhanced feature graph by using a full convolution network to obtain a final detection result:
step 701, enhancing the feature map D2、D3、D4、D5、D6Respectively inputting the data into a frame regression sub-network and a classification sub-network;
loss function L of bounding box regression subnetwork trainingregThe formula is as follows:
Figure FDA0003458321350000032
wherein d represents the difference between the predicted bounding box vector and the true label box vector; loss function L for classification subnetwork trainingclsThe formula is as follows:
Lcls=-αt(1-qt)γlog(qt)
wherein q istProbability value, alpha, representing class predictiontAnd gamma is an over-parameter for adjusting the balance of the difficult and easy samples; thus, the Loss function Loss formula of the whole network training is as follows:
Figure FDA0003458321350000041
wherein N isposRepresenting the number of positive samples, mu and lambda representing balance factors of classification loss and bounding box regression loss, respectively;
step 702, enhance the feature map D2、D3、D4、D5、D6And after respectively passing through the frame regression sub-network and the classification sub-network, obtaining a final prediction frame by using a non-maximum suppression algorithm, and outputting a ship SAR image target detection result.
CN202010267019.8A 2020-04-08 2020-04-08 SAR image ship target detection method based on non-local feature enhancement Active CN111563414B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010267019.8A CN111563414B (en) 2020-04-08 2020-04-08 SAR image ship target detection method based on non-local feature enhancement

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010267019.8A CN111563414B (en) 2020-04-08 2020-04-08 SAR image ship target detection method based on non-local feature enhancement

Publications (2)

Publication Number Publication Date
CN111563414A CN111563414A (en) 2020-08-21
CN111563414B true CN111563414B (en) 2022-03-01

Family

ID=72074155

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010267019.8A Active CN111563414B (en) 2020-04-08 2020-04-08 SAR image ship target detection method based on non-local feature enhancement

Country Status (1)

Country Link
CN (1) CN111563414B (en)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112215199B (en) * 2020-10-28 2023-09-15 中国人民解放军战略支援部队航天工程大学 SAR image ship detection method based on multi-receptive field and dense feature aggregation network
CN112487912B (en) * 2020-11-24 2024-02-06 杭州电子科技大学 Arbitrary direction ship detection method based on improved YOLOv3
CN112749734B (en) * 2020-12-29 2024-01-05 北京环境特性研究所 Domain-adaptive target detection method based on movable attention mechanism
CN113191406B (en) * 2021-04-19 2023-10-27 金科智融科技(珠海)有限公司 Cartoon image classification method based on gram matrix
CN112949779A (en) * 2021-04-20 2021-06-11 中国人民解放军国防科技大学 Global feature enhanced small target feature extraction method and device
CN113408340B (en) * 2021-05-12 2024-03-29 北京化工大学 Dual-polarization SAR small ship detection method based on enhanced feature pyramid
CN113657181B (en) * 2021-07-23 2024-01-23 西北工业大学 SAR image rotation target detection method based on smooth tag coding and feature enhancement
CN113780422B (en) * 2021-09-13 2023-06-27 北京环境特性研究所 Background clutter similarity evaluation method and device

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110084210A (en) * 2019-04-30 2019-08-02 电子科技大学 The multiple dimensioned Ship Detection of SAR image based on attention pyramid network
CN110097129A (en) * 2019-05-05 2019-08-06 西安电子科技大学 Remote sensing target detection method based on profile wave grouping feature pyramid convolution
CN110414414A (en) * 2019-07-25 2019-11-05 合肥工业大学 SAR image Ship Target discrimination method based on the fusion of multi-layer depths of features
CN110728192A (en) * 2019-09-16 2020-01-24 河海大学 High-resolution remote sensing image classification method based on novel characteristic pyramid depth network
CN110782420A (en) * 2019-09-19 2020-02-11 杭州电子科技大学 Small target feature representation enhancement method based on deep learning
CN110796037A (en) * 2019-10-15 2020-02-14 武汉大学 Satellite-borne optical remote sensing image ship target detection method based on lightweight receptive field pyramid

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016054778A1 (en) * 2014-10-09 2016-04-14 Microsoft Technology Licensing, Llc Generic object detection in images

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110084210A (en) * 2019-04-30 2019-08-02 电子科技大学 The multiple dimensioned Ship Detection of SAR image based on attention pyramid network
CN110097129A (en) * 2019-05-05 2019-08-06 西安电子科技大学 Remote sensing target detection method based on profile wave grouping feature pyramid convolution
CN110414414A (en) * 2019-07-25 2019-11-05 合肥工业大学 SAR image Ship Target discrimination method based on the fusion of multi-layer depths of features
CN110728192A (en) * 2019-09-16 2020-01-24 河海大学 High-resolution remote sensing image classification method based on novel characteristic pyramid depth network
CN110782420A (en) * 2019-09-19 2020-02-11 杭州电子科技大学 Small target feature representation enhancement method based on deep learning
CN110796037A (en) * 2019-10-15 2020-02-14 武汉大学 Satellite-borne optical remote sensing image ship target detection method based on lightweight receptive field pyramid

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
Enhanced Feature Extraction for Ship Detection from Multi-Resolution and Multi-Scene Synthetic Aperture Radar (SAR) Images;Fei Gao等;《Remote Sensing》;20191231;第11卷(第22期);1-22 *
Multi-Scale Rotation-Invariant Haar-Like Feature Integrated CNN-Based Ship Detection Algorithm of Multiple-Target Environment in SAR Imagery;Jiaqiu Ai等;《IEEE Transactions on Geoscience and Remote Sensing》;20190826;第57卷(第12期);10071-10087 *
基于深度卷积神经网络的SAR图像舰船小目标检测;胡昌华等;《中国惯性技术学报》;20190630;第27卷(第3期);397-405 *
基于深度学习的SAR舰船目标识别方法研究;丛龙剑;《中国优秀博硕士学位论文全文数据库(硕士)》;20190215;第2019年卷(第2期);C032-1 *

Also Published As

Publication number Publication date
CN111563414A (en) 2020-08-21

Similar Documents

Publication Publication Date Title
CN111563414B (en) SAR image ship target detection method based on non-local feature enhancement
CN112818903B (en) Small sample remote sensing image target detection method based on meta-learning and cooperative attention
US10984289B2 (en) License plate recognition method, device thereof, and user equipment
CN109584248B (en) Infrared target instance segmentation method based on feature fusion and dense connection network
CN108764063B (en) Remote sensing image time-sensitive target identification system and method based on characteristic pyramid
Liu et al. Fg-net: A fast and accurate framework for large-scale lidar point cloud understanding
CN113657181B (en) SAR image rotation target detection method based on smooth tag coding and feature enhancement
Neagoe et al. Concurrent self-organizing maps for supervised/unsupervised change detection in remote sensing images
CN115631344B (en) Target detection method based on feature self-adaptive aggregation
CN116228702A (en) Camouflage target detection method based on attention mechanism and convolutional neural network
CN112348116A (en) Target detection method and device using spatial context and computer equipment
Lin et al. SAN: Scale-aware network for semantic segmentation of high-resolution aerial images
CN115995042A (en) Video SAR moving target detection method and device
CN110399868B (en) Coastal wetland bird detection method
Kim Infrared variation reduction by simultaneous background suppression and target contrast enhancement for deep convolutional neural network-based automatic target recognition
Peng et al. CourtNet: Dynamically balance the precision and recall rates in infrared small target detection
Wang et al. Multi-size object detection in large scene remote sensing images under dual attention mechanism
Bouzerdoum et al. Improved deep learning-based classification of mine-like contacts in sonar images from autonomous underwater vehicles
Yang et al. A modified YOLOv5 for object detection in UAV-captured scenarios
CN116486238B (en) Target fine granularity identification method combining point set representation and graph classification
Lu et al. Neural network post-processing of grayscale optical correlator
CN115984951A (en) Method and device for judging shielding type of fisheye image
CN116310323A (en) Aircraft target instance segmentation method, system and readable storage medium
CN115100136A (en) Workpiece category and pose estimation method based on YOLOv4-tiny model
CN114821777A (en) Gesture detection method, device, equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant