CN110796009A - Method and system for detecting marine vessel based on multi-scale convolution neural network model - Google Patents

Method and system for detecting marine vessel based on multi-scale convolution neural network model Download PDF

Info

Publication number
CN110796009A
CN110796009A CN201910930804.4A CN201910930804A CN110796009A CN 110796009 A CN110796009 A CN 110796009A CN 201910930804 A CN201910930804 A CN 201910930804A CN 110796009 A CN110796009 A CN 110796009A
Authority
CN
China
Prior art keywords
layer
ship
image
convolution
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910930804.4A
Other languages
Chinese (zh)
Inventor
王平
李明
雷建胜
赵光辉
安玉拴
金明磊
李超
陈浩
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Aerospace Star Technology Co Ltd
Space Star Technology Co Ltd
Original Assignee
Aerospace Star Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Aerospace Star Technology Co Ltd filed Critical Aerospace Star Technology Co Ltd
Priority to CN201910930804.4A priority Critical patent/CN110796009A/en
Publication of CN110796009A publication Critical patent/CN110796009A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/46Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Multimedia (AREA)
  • Software Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention provides a method and a system for detecting a marine ship based on a multi-scale convolutional neural network model, wherein the method comprises the following steps: constructing a ship image sample library, acquiring ship video data of a coastal area under visible light based on an unmanned aerial vehicle platform, extracting each frame of image, and obtaining a true value and a length and a width of a ship position; then, enhancing the data through digital image processing algorithms such as inversion and scaling; constructing a multilayer convolutional neural network as a ship target detector, inputting the obtained processed image serving as sample data into a deep learning network, and obtaining a feature map after convolution; and constructing a multi-scale convolution neural unit, performing feature fusion on the multilayer convolution feature maps based on the feature maps after convolution, and training according to the obtained real position of the ship to obtain a training model. The invention adopts a multi-scale fusion method, thereby well ensuring the detection accuracy and reducing the training difficulty.

Description

Method and system for detecting marine vessel based on multi-scale convolution neural network model
Technical Field
The invention belongs to the technical field of ship digital image processing, and particularly relates to a method and a system for detecting a marine ship based on a multi-scale convolutional neural network model.
Background
In the modern society, monitoring cameras are everywhere, and if the monitoring cameras are only observed and detected by human eyes, abnormal events in videos are easily missed. With the rapid development of computer networks, communication and semiconductor technologies, people are more and more interested in analyzing video images by using computer vision instead of human eyes to obtain useful information in the video images. Target detection is a key point of computer vision research, and the main function of the target detection is to extract the position of a target which is interested in people in an image and other information. The target detection is the basis of many video applications, more is the necessity of applications such as traffic monitoring, intelligent robot and man-machine interaction, and the like, has an important role in intelligent city management, illegal crime fighting and safe city and smart city construction, and is the key and difficult point of current video processing research. For the ship target, the ship management, supervision and scheduling of coastal cities play a crucial role.
The inventor provides a ship detection method and system based on scene multi-dimensional features in a patent application named as a ship detection method and system based on scene multi-dimensional features (patent application No. 201711311822.1 publication No. CN 107818326B), wherein all edges of each frame of image are extracted as the fourth dimension of the image; extracting to obtain a coastline, and enabling a sea surface area to be a ship appearance range area; constructing a Fast Region-based functional Neural Networks (RCNN) convolution network as a deep learning network, and inputting sample data into the deep learning network; constructing an RPN (region pro-social network) network, generating region suggestion frames with different sizes in a ship appearance range region by using a sliding window, combining the region suggestion frames with the obtained deep learning network, and training a model according to the real position of the ship; and carrying out ship detection on the part between the coastlines on the detection image based on the model obtained by training. The method has the disadvantages that Hough only has a good segmentation effect on the coastline which is in a straight line, and the segmentation robustness is poor; the fast RCNN-like network needs to manually specify the default back box size of the algorithm in advance, has high training difficulty and is difficult to adapt to ships with various dimensions.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and provide a method and a system for detecting a marine vessel based on a multi-scale convolutional neural network model.
In order to achieve the purpose, the technical scheme of the invention is as follows:
a marine vessel detection method based on a multi-scale convolutional neural network model comprises the following steps:
step 1, constructing a ship image sample library, collecting ship video data of a coastal area under visible light, extracting each frame of image, obtaining a true value and a length and a width of a ship position, and further performing enhancement processing on the ship video data through a digital image processing algorithm;
step 2, constructing a multilayer convolutional neural network as a ship target detector, inputting the processed image obtained in the step 1 into a deep learning network as sample data, and obtaining a feature map after convolution;
and 3, constructing a multi-scale convolution neural unit, performing feature fusion on the multilayer convolution feature maps based on the feature maps obtained in the step 2 after convolution, and training according to the real position of the ship to further obtain a training model.
As a further preferable solution of the method for detecting a marine vessel based on the multi-scale convolutional neural network model of the present invention, in step 1, the scaling operation of the data enhancement process is implemented as follows:
step 1.1, scaling the image according to the original length-width ratio until the longest edge is equal to 500;
step 1.2, then, the short edges with the length less than 500 pixels are filled with gray to form a picture with the side length of 500 pixels;
and step 1.3, if the longest edge of the original picture is less than 500 pixels, performing gray compensation operation.
As a further preferable aspect of the method for detecting a marine vessel based on the multi-scale convolutional neural network model of the present invention, in step 2,
the network structure of the constructed multilayer convolutional neural network consists of 62 convolutional layers and 5 pooling layers;
for the common convolution layer, a convolution core with a modifiable numerical value is used for convolving the feature layer of the previous layer, and an output feature layer can be obtained by activating a function;
Figure BDA0002220237450000031
where Mj represents the set of input layers selected, i is the index value of the input layer cell, j is the index value of the output layer cell,representing the weight between the i-th layer input layer and the j-th layer output layer,
Figure BDA0002220237450000033
represents the activation bias for the output layer of layer j, f () represents the activation function for that output layer,
Figure BDA0002220237450000034
the jth output layer representing the l layer,
Figure BDA0002220237450000035
the ith input layer of the l-1 layer is represented, and the pooling layer comprises N input layers and N output layers;
wherein d isown () represents a down-sampling function, typically summing all pixels in different n × n regions of the input image, so that the output image is reduced by a factor of n in both dimensions, each output layer corresponding to a respective multiplicative offsetAnd an additive bias
Figure BDA0002220237450000042
Figure BDA0002220237450000043
The jth output layer representing the l layer,
Figure BDA0002220237450000044
the jth input layer representing the l-1 layer;
for the output full connection layer, convolution is input into multiple characteristic layers, and the convolution values are summed to obtain the output layer, using aijRepresenting the weight or contribution of the ith input layer in the jth output feature layer, the jth output layer can be represented as:
and the weight aijThe constraints need to be satisfied:
Figure BDA0002220237450000046
wherein N isinIndicates the number of layers of the input features,
Figure BDA0002220237450000047
representing the weight between the input layer and the output layer,
Figure BDA0002220237450000048
indicating the activation bias between the various layers,the jth output layer representing the l layer,
Figure BDA00022202374500000410
the jth input layer representing the l-1 layer;
as a further preferable scheme of the method for detecting the marine vessel based on the multi-scale convolutional neural network model of the present invention, in step 3, the multi-scale feature fusion module specifically comprises:
step 3.1, extracting feature maps output by the convolution layers of layers 9, 12, 15, 17 and 19 in the whole neural network, and then splicing the 5 feature maps to form a new feature map; then, the new feature map is fed into two multi-scale loss function modules
Figure BDA00022202374500000411
And
Figure BDA00022202374500000412
calculating a multi-scale loss function and updating network parameters; the multiscale loss function during training is defined as follows:
Figure BDA0002220237450000051
the multi-scale loss function in training consists of two multi-scale loss function modules
Figure BDA0002220237450000052
And
Figure BDA0002220237450000053
a composition in which i represents an element number in a feature map; p is a radical ofiIndicating the probability that the ith element in the feature map contains the target at the position corresponding to the original image,
Figure BDA0002220237450000054
the probability of whether the original image position in the real label data contains the target or not is shown, and if the real value of the label is judged to be the real valueThe ith position in the feature map is a target, and if the ith position is 0, the ith position is not the target; t is tiIndicating the coordinate offset of the ith element in the feature map corresponding to the frame in the original image
Figure BDA0002220237450000056
Then represents the offset of the real frame coordinates; n is a radical ofclsAnd NregThe total number of target classes and the total number of frame coordinate offsets contained in the feature map, and the classification loss functionUsing a conventional softmax function, a regression loss functionThe used method is smoothL1loss:
Figure BDA0002220237450000059
Wherein the content of the first and second substances,
Figure BDA00022202374500000510
which means that the regression loss function is calculated only for the candidate box with the target, lambda is an adjustable parameter, and is set to 3 by default, to balance the unbalanced influence of the ratio between positive and negative samples in the data on the final loss function.
A ship detection system for building a deep learning network model based on scene multi-dimensional features comprises the following steps:
the image acquisition module is used for constructing a ship image sample library, and comprises the steps of acquiring video data acquired by the unmanned aerial vehicle in the coastal region under visible light, extracting each frame of image and obtaining a ship position true value and length and width;
the data enhancement processing module is used for carrying out data enhancement processing on the data set, and comprises classic digital image processing algorithms such as inversion and scaling on the image;
the characteristic extraction module is used for constructing a multilayer convolutional neural network as a ship target detector, and inputting an image serving as sample data into the multilayer convolutional neural network after data enhancement to obtain a characteristic diagram;
and the training module is used for constructing a multi-scale convolution nerve unit, performing feature fusion on the plurality of feature maps, and training according to the real position of the ship in the data set to obtain a training result.
The technical scheme provided by the invention has the beneficial effects that:
(1) according to the actual data condition, various small ships and large ships coexist on the sea, and the large difference of the same kind of targets is the main reason of low detection recall rate. The invention integrates the characteristics of a plurality of characteristic graphs of the multilayer convolutional neural network, integrates the image detail information of the network shallow layer and the macroscopic semantic information of the network high layer, and improves the recall rate of small target detection while ensuring the original large target detection quality.
(2) According to the invention, an image data enhancement strategy is added in target detection, so that the robustness detection of the algorithm is improved. The method still has a good detection result for complex scenes such as cloud and fog, cloudy days, raining and the like. The method can be used for providing the ocean supervision work efficiency, saving the supervision cost, providing scientific basis for the formulation of ocean management decisions, and having important market value.
Drawings
FIG. 1 is a flow chart of an embodiment of the present invention;
fig. 2 is a schematic diagram of a network structure according to an embodiment of the present invention.
Detailed Description
The invention provides a ship detection method of a multi-scale feature fusion convolutional neural network. Firstly, an image sample library is constructed, and a ship image is subjected to sample marking to obtain a calibrated image sample. And then enhancing the data through digital image processing algorithms such as inversion and scaling, and constructing a deep learning network to carry out convolution on the image. And then constructing a multilayer convolutional neural network as a ship target detector, inputting the processed image as sample data into a deep learning network to obtain a feature map after convolution, constructing a multi-scale convolutional neural unit, performing feature fusion on the multilayer convolutional feature map, finally obtaining a loss function of a suggestion box by using a ship position truth value, training the whole network, and outputting a trained model. And finally, carrying out ship detection on the test data by using the trained model. The method mainly comprises five processes of sample library construction, image data enhancement, multi-scale feature fusion, deep learning network training and ship detection.
To illustrate the specific embodiments in detail, as shown in fig. 1, the flow of the example is as follows:
step a, constructing a ship image sample library; the data is enhanced through digital image processing algorithms such as inversion and scaling, and the number of the image data is expanded.
Firstly, ship images are prepared, and data required to be collected by the method are mainly video data shot by unmanned aerial vehicles in coastal areas under visible light. For the acquired video data, each frame of image needs to be obtained through a decoder or a code, and a ship image sample library with sufficient diversity is obtained for a plurality of videos. And then, obtaining a true value and a length and a width of the ship position through a preselected mark for each frame of image in the ship image sample library. And then, enhancing the image data through inversion and scaling, adding a new image formed after inversion and scaling into a sample library, and carrying out pre-selection marking and training together.
Wherein implementation details of scaling include:
(1) scaling the image to have its original aspect ratio with the longest side equal to 500 pixels;
(2) then, the short edges with the length less than 500 pixels are filled with gray, and finally a picture with the side length of 500 pixels is formed;
(3) if the longest edge of the original picture is less than 500 pixels, the gray padding operation is directly performed.
And b, constructing a multilayer convolutional neural network as a ship target detector, and inputting the processed image obtained in the step a into the multilayer convolutional neural network as sample data to obtain a characteristic diagram.
The network structure of the built multilayer convolutional neural network consists of 62 convolutional layers and 5 maximum pooling layers.
For the common convolution layer, a convolution core with a value being modified is used to convolute the feature layer of the previous layer, and then the output feature layer can be obtained by activating the function. Each output layer may be a combination of convolving the values of multiple input layers:
Figure BDA0002220237450000081
where Mj represents the set of input layers selected, i is the index value of the input layer cell, j is the index value of the output layer cell,
Figure BDA0002220237450000082
representing the weight between the i-th layer input layer and the j-th layer output layer,
Figure BDA0002220237450000083
represents the activation bias for the output layer of layer j, f () represents the activation function for that output layer,
Figure BDA0002220237450000084
the jth output layer representing the l layer,
Figure BDA0002220237450000085
the ith input layer representing the l-1 layer. For the pooling layer, there are N input layers and N output layers, except that each output layer is smaller.
Figure BDA0002220237450000086
down () represents a downsampling function. Typically, all pixels in different n × n regions of the input image are summed. So that the output image is reduced by a factor of n in both dimensions. Each output layer corresponds to a respective multiplicative bias
Figure BDA0002220237450000087
And an additive bias
Figure BDA0002220237450000088
Figure BDA0002220237450000089
The jth output layer representing the l layer,the jth input layer representing the l-1 layer.
For the output fully connected layer, it is often better to convolve the input multiple feature layers and then sum the convolved values to obtain the output layer. Example of the invention usesijIndicating the weight or contribution of the ith input layer in obtaining the jth output feature layer. Thus, the jth output layer can be represented as:
Figure BDA0002220237450000092
and the weight aijThe constraints need to be satisfied:
Figure BDA0002220237450000093
wherein N isinIndicates the number of layers of the input features,
Figure BDA0002220237450000094
representing the weight between the input layer and the output layer,
Figure BDA0002220237450000095
indicating the activation bias between the various layers,
Figure BDA0002220237450000096
the jth output layer representing the l layer,
Figure BDA0002220237450000097
the jth input layer representing the l-1 layer.
And c, constructing a multi-scale convolution nerve training unit, performing feature fusion on a plurality of feature maps based on the plurality of feature maps obtained in the step b, and finally training according to the real position of the ship obtained in the step a to obtain a training model.
As shown in fig. 2, the workflow of the multi-scale feature fusion module is as follows:
firstly, extracting feature maps output by the convolution layers of layers 9, 12, 15, 17 and 19 in the whole neural network, and then splicing the 5 feature maps to form a new feature map; then, the new feature map is fed into two multi-scale loss function modules
Figure BDA0002220237450000098
And
Figure BDA0002220237450000099
and calculating a multi-scale loss function and updating network parameters. The multiscale loss function during training is defined as follows:
Figure BDA0002220237450000101
the function is composed of two multi-scale loss function modules
Figure BDA0002220237450000102
Andand (4) forming. Wherein i represents the element number in the feature map; p is a radical ofiIndicating the probability that the ith element in the feature map contains the target at the position corresponding to the original image,
Figure BDA0002220237450000104
indicating the probability of whether the original image position in the real label data contains the target or not, if the label real value
Figure BDA0002220237450000105
It means that the ith position in the feature map is the target, and if it is 0Then it is not; t is tiIndicating the coordinate offset of the ith element in the feature map corresponding to the frame in the original image
Figure BDA0002220237450000106
An offset of the real frame coordinates is indicated. N is a radical ofclsAnd NregThe total number of object classes and the total number of frame coordinate offsets included in the feature map are provided. Classification loss function
Figure BDA0002220237450000107
The conventional softmax function is used. Function of regression lossThe used method is smoothL1loss:
Figure BDA0002220237450000109
Wherein
Figure BDA00022202374500001010
Indicating that the regression loss function is calculated only for the candidate box with the target. λ is an adjustable parameter, set to 3 by default, to balance the unbalanced effect of the ratio between positive and negative samples in the data on the final loss function.
So far, the detailed implementation process of the marine vessel detection method based on the multi-scale feature fusion convolutional neural network used in the embodiment of the application is introduced. In specific implementation, the method provided by the invention can realize automatic operation flow based on software technology, and can also realize a corresponding system in a modularized mode.
The embodiment of the invention also provides a ship detection system for constructing a deep learning network model based on scene multi-dimensional features, which comprises the following modules:
the image acquisition module is used for constructing a ship image sample library, and comprises the steps of acquiring video data acquired by the unmanned aerial vehicle in the coastal region under visible light, extracting each frame of image and obtaining a ship position true value and length and width;
the data enhancement processing module is used for carrying out data enhancement processing on the data set, and comprises classic digital image processing algorithms such as inversion and scaling on the image;
the characteristic extraction module is used for constructing a multilayer convolutional neural network as a ship target detector, and inputting an image serving as sample data into the multilayer convolutional neural network after data enhancement to obtain a characteristic diagram;
and the training module is used for constructing a multi-scale convolution nerve unit, performing feature fusion on the plurality of feature maps, and training according to the real position of the ship in the data set to obtain a training result. The specific implementation of each module can refer to the corresponding step, and the detailed description of the invention is omitted. The specific examples described herein are merely illustrative of the spirit of the invention. Various modifications or additions may be made or substituted in a similar manner to the specific embodiments described herein by those skilled in the art without departing from the spirit of the invention or exceeding the scope thereof as defined in the appended claims.

Claims (5)

1. A method for detecting a marine vessel based on a multi-scale convolutional neural network model is characterized by comprising the following steps:
step 1, constructing a ship image sample library, collecting ship video data of a coastal area under visible light, extracting each frame of image, obtaining a true value and a length and a width of a ship position, and performing enhancement processing on the ship video data through a digital image processing algorithm;
step 2, constructing a multilayer convolutional neural network as a ship target detector, inputting the processed image obtained in the step 1 into a deep learning network as sample data, and obtaining a feature map after convolution;
and 3, constructing a multi-scale convolution neural unit, performing feature fusion on the multilayer convolution feature maps based on the feature maps obtained in the step 2 after convolution, and training according to the real position of the ship to further obtain a training model.
2. The method of claim 1, wherein the method comprises the following steps: in step 1, the data enhancement process includes:
step 1.1, scaling the image according to the original length-width ratio until the longest edge is equal to 500 pixels;
step 1.2, then, the short edges with the length less than 500 pixels are filled with gray to form a picture with the side length of 500 pixels;
and step 1.3, if the longest edge of the original picture is less than 500 pixels, performing gray compensation.
3. The method of claim 1, wherein the method comprises the following steps: in the step 2, the process is carried out,
the network structure of the constructed multilayer convolutional neural network consists of 62 convolutional layers and 5 pooling layers;
for the common convolutional layer, carrying out convolution on the feature layer of the previous layer by using a convolution core with a modifiable numerical value, and obtaining an output feature layer through an activation function;
Figure FDA0002220237440000021
where Mj represents the set of input layers selected, i is the index value of the input layer cell, j is the index value of the output layer cell,
Figure FDA0002220237440000022
representing the weight between the i-th layer input layer and the j-th layer output layer,
Figure FDA0002220237440000023
represents the activation bias for the output layer of layer j, f () represents the activation function for that output layer,
Figure FDA0002220237440000024
the jth output layer representing the l layer,the ith input layer of the l-1 layer is represented, and the pooling layer comprises N input layers and N output layers;
Figure FDA0002220237440000026
wherein down () represents a down-sampling function, which is the summation of all pixels in different n × n regions of the input image, such that the output image is reduced by n times in both dimensions, and each output layer corresponds to a respective multiplicative biasAnd an additive bias
Figure FDA0002220237440000028
Figure FDA00022202374400000211
The jth output layer representing the l layer,
Figure FDA0002220237440000029
the jth input layer representing the l-1 layer;
for the output full-connection layer, convolution is input into a plurality of characteristic layers, and the convolution values are summed to obtain the output layer if aijRepresenting the weight or contribution of the ith input layer in the jth output feature layer, the jth output layer can be represented as:
and a isijThe constraints need to be satisfied:
Figure FDA0002220237440000031
wherein N isinIndicates the number of layers of the input features,
Figure FDA0002220237440000032
representing the weight between the input layer and the output layer,
Figure FDA0002220237440000033
indicating the activation bias between the various layers,the jth output layer representing the l layer,
Figure FDA0002220237440000035
the jth input layer representing the l-1 layer.
4. The method of claim 1, wherein the method comprises the following steps: in step 3, feature fusion comprises:
step 3.1, extracting feature maps output by the convolution layers of layers 9, 12, 15, 17 and 19 in the whole neural network, and then splicing the 5 feature maps to form a new feature map; then, the new characteristic graphs are respectively input into two multi-scale loss function modules
Figure FDA0002220237440000036
And
Figure FDA0002220237440000037
calculating a multi-scale loss function and updating network parameters; the multiscale loss function during training is defined as follows:
Figure FDA0002220237440000038
the multi-scale loss function in training consists of two multi-scale loss function modules
Figure FDA0002220237440000039
And
Figure FDA00022202374400000310
a composition in which i represents an element number in a feature map; p is a radical ofiIndicating the probability that the ith element in the feature map contains the target at the position corresponding to the original image,
Figure FDA00022202374400000311
the probability of whether the original image position in the real label data contains the target or not is shown, and if the real value of the label is judged to be the real value
Figure FDA00022202374400000312
The ith position in the feature map is a target, and if the ith position is 0, the ith position is not the target; t is tiIndicating the coordinate offset of the ith element in the feature map corresponding to the frame in the original image
Figure FDA00022202374400000313
Then represents the offset of the real frame coordinates; n is a radical ofclsAnd NregThe total number of target classes and the total number of frame coordinate offsets contained in the feature map, and the classification loss functionUsing a conventional softmax function, a regression loss function
Figure FDA00022202374400000315
The used method is smoothL1loss:
Figure FDA0002220237440000041
Wherein the content of the first and second substances,
Figure FDA0002220237440000042
meaning that the regression loss function is calculated only for the candidate box with the target, λ is an adjustable parameter, and is set to 3 by default, to balance the ratio between positive and negative samples in the dataUnbalanced effects on the final loss function.
5. A ship detection system for building a deep learning network model based on scene multi-dimensional features according to any one of claims 1 to 4, the system comprising:
the image acquisition module is used for constructing a ship image sample library, and comprises the steps of acquiring video data acquired by the unmanned aerial vehicle in the coastal region under visible light, extracting each frame of image and obtaining a ship position true value and length and width;
the data enhancement processing module is used for carrying out data enhancement processing on the data set, and comprises classic digital image processing algorithms such as inversion and scaling on the image;
the characteristic extraction module is used for constructing a multilayer convolutional neural network as a ship target detector, and inputting an image serving as sample data into the multilayer convolutional neural network after data enhancement to obtain a characteristic diagram;
and the training module is used for constructing a multi-scale convolution nerve unit, performing feature fusion on the plurality of feature maps, and training according to the real position of the ship in the data set to obtain a training result.
CN201910930804.4A 2019-09-29 2019-09-29 Method and system for detecting marine vessel based on multi-scale convolution neural network model Pending CN110796009A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910930804.4A CN110796009A (en) 2019-09-29 2019-09-29 Method and system for detecting marine vessel based on multi-scale convolution neural network model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910930804.4A CN110796009A (en) 2019-09-29 2019-09-29 Method and system for detecting marine vessel based on multi-scale convolution neural network model

Publications (1)

Publication Number Publication Date
CN110796009A true CN110796009A (en) 2020-02-14

Family

ID=69438678

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910930804.4A Pending CN110796009A (en) 2019-09-29 2019-09-29 Method and system for detecting marine vessel based on multi-scale convolution neural network model

Country Status (1)

Country Link
CN (1) CN110796009A (en)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111814536A (en) * 2020-05-21 2020-10-23 闽江学院 Breeding monitoring method and device
CN111967305A (en) * 2020-07-01 2020-11-20 华南理工大学 Real-time multi-scale target detection method based on lightweight convolutional neural network
CN112016542A (en) * 2020-05-08 2020-12-01 珠海欧比特宇航科技股份有限公司 Urban waterlogging intelligent detection method and system
CN112036404A (en) * 2020-08-31 2020-12-04 上海大学 Target detection method and system for offshore ship
CN112085001A (en) * 2020-09-23 2020-12-15 清华大学苏州汽车研究院(相城) Tunnel recognition model and method based on multi-scale edge feature detection
CN112183463A (en) * 2020-10-23 2021-01-05 珠海大横琴科技发展有限公司 Ship identification model verification method and device based on radar image
CN112183232A (en) * 2020-09-09 2021-01-05 上海鹰觉科技有限公司 Ship board number position positioning method and system based on deep learning
CN112285712A (en) * 2020-10-15 2021-01-29 电子科技大学 Method for improving detection precision of ship on shore in SAR image
CN112464765A (en) * 2020-09-10 2021-03-09 天津师范大学 Safety helmet detection algorithm based on single-pixel characteristic amplification and application thereof
CN112986210A (en) * 2021-02-10 2021-06-18 四川大学 Scale-adaptive microbial Raman spectrum detection method and system
CN113807386A (en) * 2021-07-21 2021-12-17 广东工业大学 Target detection method and system fusing multi-scale information and computer equipment
CN114720957A (en) * 2022-06-08 2022-07-08 中国人民解放军空军预警学院 Radar target detection method and system and storable medium
CN114943888A (en) * 2022-03-24 2022-08-26 中国人民解放军海军大连舰艇学院 Sea surface small target detection method based on multi-scale information fusion, electronic equipment and computer readable medium
CN116051548A (en) * 2023-03-14 2023-05-02 中国铁塔股份有限公司 Positioning method and device
CN116883913A (en) * 2023-09-05 2023-10-13 长江信达软件技术(武汉)有限责任公司 Ship identification method and system based on video stream adjacent frames

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107818326A (en) * 2017-12-11 2018-03-20 珠海大横琴科技发展有限公司 A kind of ship detection method and system based on scene multidimensional characteristic
CN109345508A (en) * 2018-08-31 2019-02-15 北京航空航天大学 A kind of Assessing Standards For Skeletal method based on two stages neural network
CN109522963A (en) * 2018-11-26 2019-03-26 北京电子工程总体研究所 A kind of the feature building object detection method and system of single-unit operation
CN109598290A (en) * 2018-11-22 2019-04-09 上海交通大学 A kind of image small target detecting method combined based on hierarchical detection
CN109740515A (en) * 2018-12-29 2019-05-10 科大讯飞股份有限公司 One kind reading and appraising method and device
CN110110783A (en) * 2019-04-30 2019-08-09 天津大学 A kind of deep learning object detection method based on the connection of multilayer feature figure

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107818326A (en) * 2017-12-11 2018-03-20 珠海大横琴科技发展有限公司 A kind of ship detection method and system based on scene multidimensional characteristic
CN109345508A (en) * 2018-08-31 2019-02-15 北京航空航天大学 A kind of Assessing Standards For Skeletal method based on two stages neural network
CN109598290A (en) * 2018-11-22 2019-04-09 上海交通大学 A kind of image small target detecting method combined based on hierarchical detection
CN109522963A (en) * 2018-11-26 2019-03-26 北京电子工程总体研究所 A kind of the feature building object detection method and system of single-unit operation
CN109740515A (en) * 2018-12-29 2019-05-10 科大讯飞股份有限公司 One kind reading and appraising method and device
CN110110783A (en) * 2019-04-30 2019-08-09 天津大学 A kind of deep learning object detection method based on the connection of multilayer feature figure

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
刘云等: ""深度学习的多尺度多人目标检测方法研究"", 《计算机工程与应用》 *
顾亚风: ""面向图像内容检索的卷积神经网络"", 《中国优秀博硕士学位论文全文数据库(硕士)信息科技辑》 *

Cited By (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112016542A (en) * 2020-05-08 2020-12-01 珠海欧比特宇航科技股份有限公司 Urban waterlogging intelligent detection method and system
CN111814536B (en) * 2020-05-21 2023-11-28 闽江学院 Culture monitoring method and device
CN111814536A (en) * 2020-05-21 2020-10-23 闽江学院 Breeding monitoring method and device
CN111967305B (en) * 2020-07-01 2022-03-18 华南理工大学 Real-time multi-scale target detection method based on lightweight convolutional neural network
CN111967305A (en) * 2020-07-01 2020-11-20 华南理工大学 Real-time multi-scale target detection method based on lightweight convolutional neural network
CN112036404A (en) * 2020-08-31 2020-12-04 上海大学 Target detection method and system for offshore ship
CN112036404B (en) * 2020-08-31 2024-01-02 上海大学 Marine ship target detection method and system
CN112183232A (en) * 2020-09-09 2021-01-05 上海鹰觉科技有限公司 Ship board number position positioning method and system based on deep learning
CN112464765B (en) * 2020-09-10 2022-09-23 天津师范大学 Safety helmet detection method based on single-pixel characteristic amplification and application thereof
CN112464765A (en) * 2020-09-10 2021-03-09 天津师范大学 Safety helmet detection algorithm based on single-pixel characteristic amplification and application thereof
CN112085001A (en) * 2020-09-23 2020-12-15 清华大学苏州汽车研究院(相城) Tunnel recognition model and method based on multi-scale edge feature detection
CN112085001B (en) * 2020-09-23 2024-04-23 清华大学苏州汽车研究院(相城) Tunnel identification model and method based on multi-scale edge feature detection
CN112285712A (en) * 2020-10-15 2021-01-29 电子科技大学 Method for improving detection precision of ship on shore in SAR image
CN112285712B (en) * 2020-10-15 2023-09-15 电子科技大学 Method for improving detection precision of coasting ship in SAR image
CN112183463A (en) * 2020-10-23 2021-01-05 珠海大横琴科技发展有限公司 Ship identification model verification method and device based on radar image
CN112183463B (en) * 2020-10-23 2021-10-15 珠海大横琴科技发展有限公司 Ship identification model verification method and device based on radar image
CN112986210A (en) * 2021-02-10 2021-06-18 四川大学 Scale-adaptive microbial Raman spectrum detection method and system
CN113807386B (en) * 2021-07-21 2023-08-01 广东工业大学 Target detection method, system and computer equipment integrating multi-scale information
CN113807386A (en) * 2021-07-21 2021-12-17 广东工业大学 Target detection method and system fusing multi-scale information and computer equipment
CN114943888A (en) * 2022-03-24 2022-08-26 中国人民解放军海军大连舰艇学院 Sea surface small target detection method based on multi-scale information fusion, electronic equipment and computer readable medium
CN114720957A (en) * 2022-06-08 2022-07-08 中国人民解放军空军预警学院 Radar target detection method and system and storable medium
CN116051548A (en) * 2023-03-14 2023-05-02 中国铁塔股份有限公司 Positioning method and device
CN116051548B (en) * 2023-03-14 2023-08-11 中国铁塔股份有限公司 Positioning method and device
CN116883913B (en) * 2023-09-05 2023-11-21 长江信达软件技术(武汉)有限责任公司 Ship identification method and system based on video stream adjacent frames
CN116883913A (en) * 2023-09-05 2023-10-13 长江信达软件技术(武汉)有限责任公司 Ship identification method and system based on video stream adjacent frames

Similar Documents

Publication Publication Date Title
CN110796009A (en) Method and system for detecting marine vessel based on multi-scale convolution neural network model
CN113065558B (en) Lightweight small target detection method combined with attention mechanism
CN109584248B (en) Infrared target instance segmentation method based on feature fusion and dense connection network
CN110310241B (en) Method for defogging traffic image with large air-light value by fusing depth region segmentation
CN109871798A (en) A kind of remote sensing image building extracting method based on convolutional neural networks
CN111046880A (en) Infrared target image segmentation method and system, electronic device and storage medium
CN111832443B (en) Construction method and application of construction violation detection model
CN117078943B (en) Remote sensing image road segmentation method integrating multi-scale features and double-attention mechanism
CN109919223B (en) Target detection method and device based on deep neural network
CN115205264A (en) High-resolution remote sensing ship detection method based on improved YOLOv4
CN110490155B (en) Method for detecting unmanned aerial vehicle in no-fly airspace
CN113160062A (en) Infrared image target detection method, device, equipment and storage medium
CN116311254B (en) Image target detection method, system and equipment under severe weather condition
CN115035295B (en) Remote sensing image semantic segmentation method based on shared convolution kernel and boundary loss function
CN113420794B (en) Binaryzation Faster R-CNN citrus disease and pest identification method based on deep learning
CN111353396A (en) Concrete crack segmentation method based on SCSEOCUnet
CN114359245A (en) Method for detecting surface defects of products in industrial scene
CN116071676A (en) Infrared small target detection method based on attention-directed pyramid fusion
CN115527096A (en) Small target detection method based on improved YOLOv5
CN116469020A (en) Unmanned aerial vehicle image target detection method based on multiscale and Gaussian Wasserstein distance
CN112785610B (en) Lane line semantic segmentation method integrating low-level features
CN117372829A (en) Marine vessel target identification method, device, electronic equipment and readable medium
CN111950476A (en) Deep learning-based automatic river channel ship identification method in complex environment
CN116977866A (en) Lightweight landslide detection method
CN116452900A (en) Target detection method based on lightweight neural network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20200214