CN110796009A - Method and system for detecting marine vessel based on multi-scale convolution neural network model - Google Patents
Method and system for detecting marine vessel based on multi-scale convolution neural network model Download PDFInfo
- Publication number
- CN110796009A CN110796009A CN201910930804.4A CN201910930804A CN110796009A CN 110796009 A CN110796009 A CN 110796009A CN 201910930804 A CN201910930804 A CN 201910930804A CN 110796009 A CN110796009 A CN 110796009A
- Authority
- CN
- China
- Prior art keywords
- layer
- ship
- image
- convolution
- feature
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 30
- 238000003062 neural network model Methods 0.000 title description 2
- 238000012549 training Methods 0.000 claims abstract description 30
- 238000013527 convolutional neural network Methods 0.000 claims abstract description 26
- 238000001514 detection method Methods 0.000 claims abstract description 22
- 238000012545 processing Methods 0.000 claims abstract description 18
- 230000004927 fusion Effects 0.000 claims abstract description 14
- 238000013135 deep learning Methods 0.000 claims abstract description 12
- 230000001537 neural effect Effects 0.000 claims abstract description 4
- 230000006870 function Effects 0.000 claims description 36
- 230000004913 activation Effects 0.000 claims description 10
- 238000011176 pooling Methods 0.000 claims description 6
- 238000010586 diagram Methods 0.000 claims description 5
- 230000008569 process Effects 0.000 claims description 5
- 238000013528 artificial neural network Methods 0.000 claims description 4
- 210000005036 nerve Anatomy 0.000 claims description 4
- 239000000654 additive Substances 0.000 claims description 3
- 230000000996 additive effect Effects 0.000 claims description 3
- 230000000694 effects Effects 0.000 claims description 3
- 238000000605 extraction Methods 0.000 claims description 3
- 239000000203 mixture Substances 0.000 claims description 3
- 238000005070 sampling Methods 0.000 claims description 2
- 239000000126 substance Substances 0.000 claims description 2
- 230000002708 enhancing effect Effects 0.000 abstract description 3
- 238000007500 overflow downdraw method Methods 0.000 abstract 1
- 238000012544 monitoring process Methods 0.000 description 3
- 230000003213 activating effect Effects 0.000 description 2
- 238000010276 construction Methods 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 230000011218 segmentation Effects 0.000 description 2
- 230000002159 abnormal effect Effects 0.000 description 1
- 238000007792 addition Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000009472 formulation Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
- G06V20/41—Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
- G06V20/46—Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- Multimedia (AREA)
- Software Systems (AREA)
- General Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Evolutionary Computation (AREA)
- Biophysics (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Mathematical Physics (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The invention provides a method and a system for detecting a marine ship based on a multi-scale convolutional neural network model, wherein the method comprises the following steps: constructing a ship image sample library, acquiring ship video data of a coastal area under visible light based on an unmanned aerial vehicle platform, extracting each frame of image, and obtaining a true value and a length and a width of a ship position; then, enhancing the data through digital image processing algorithms such as inversion and scaling; constructing a multilayer convolutional neural network as a ship target detector, inputting the obtained processed image serving as sample data into a deep learning network, and obtaining a feature map after convolution; and constructing a multi-scale convolution neural unit, performing feature fusion on the multilayer convolution feature maps based on the feature maps after convolution, and training according to the obtained real position of the ship to obtain a training model. The invention adopts a multi-scale fusion method, thereby well ensuring the detection accuracy and reducing the training difficulty.
Description
Technical Field
The invention belongs to the technical field of ship digital image processing, and particularly relates to a method and a system for detecting a marine ship based on a multi-scale convolutional neural network model.
Background
In the modern society, monitoring cameras are everywhere, and if the monitoring cameras are only observed and detected by human eyes, abnormal events in videos are easily missed. With the rapid development of computer networks, communication and semiconductor technologies, people are more and more interested in analyzing video images by using computer vision instead of human eyes to obtain useful information in the video images. Target detection is a key point of computer vision research, and the main function of the target detection is to extract the position of a target which is interested in people in an image and other information. The target detection is the basis of many video applications, more is the necessity of applications such as traffic monitoring, intelligent robot and man-machine interaction, and the like, has an important role in intelligent city management, illegal crime fighting and safe city and smart city construction, and is the key and difficult point of current video processing research. For the ship target, the ship management, supervision and scheduling of coastal cities play a crucial role.
The inventor provides a ship detection method and system based on scene multi-dimensional features in a patent application named as a ship detection method and system based on scene multi-dimensional features (patent application No. 201711311822.1 publication No. CN 107818326B), wherein all edges of each frame of image are extracted as the fourth dimension of the image; extracting to obtain a coastline, and enabling a sea surface area to be a ship appearance range area; constructing a Fast Region-based functional Neural Networks (RCNN) convolution network as a deep learning network, and inputting sample data into the deep learning network; constructing an RPN (region pro-social network) network, generating region suggestion frames with different sizes in a ship appearance range region by using a sliding window, combining the region suggestion frames with the obtained deep learning network, and training a model according to the real position of the ship; and carrying out ship detection on the part between the coastlines on the detection image based on the model obtained by training. The method has the disadvantages that Hough only has a good segmentation effect on the coastline which is in a straight line, and the segmentation robustness is poor; the fast RCNN-like network needs to manually specify the default back box size of the algorithm in advance, has high training difficulty and is difficult to adapt to ships with various dimensions.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and provide a method and a system for detecting a marine vessel based on a multi-scale convolutional neural network model.
In order to achieve the purpose, the technical scheme of the invention is as follows:
a marine vessel detection method based on a multi-scale convolutional neural network model comprises the following steps:
step 1, constructing a ship image sample library, collecting ship video data of a coastal area under visible light, extracting each frame of image, obtaining a true value and a length and a width of a ship position, and further performing enhancement processing on the ship video data through a digital image processing algorithm;
step 2, constructing a multilayer convolutional neural network as a ship target detector, inputting the processed image obtained in the step 1 into a deep learning network as sample data, and obtaining a feature map after convolution;
and 3, constructing a multi-scale convolution neural unit, performing feature fusion on the multilayer convolution feature maps based on the feature maps obtained in the step 2 after convolution, and training according to the real position of the ship to further obtain a training model.
As a further preferable solution of the method for detecting a marine vessel based on the multi-scale convolutional neural network model of the present invention, in step 1, the scaling operation of the data enhancement process is implemented as follows:
step 1.1, scaling the image according to the original length-width ratio until the longest edge is equal to 500;
step 1.2, then, the short edges with the length less than 500 pixels are filled with gray to form a picture with the side length of 500 pixels;
and step 1.3, if the longest edge of the original picture is less than 500 pixels, performing gray compensation operation.
As a further preferable aspect of the method for detecting a marine vessel based on the multi-scale convolutional neural network model of the present invention, in step 2,
the network structure of the constructed multilayer convolutional neural network consists of 62 convolutional layers and 5 pooling layers;
for the common convolution layer, a convolution core with a modifiable numerical value is used for convolving the feature layer of the previous layer, and an output feature layer can be obtained by activating a function;
where Mj represents the set of input layers selected, i is the index value of the input layer cell, j is the index value of the output layer cell,representing the weight between the i-th layer input layer and the j-th layer output layer,represents the activation bias for the output layer of layer j, f () represents the activation function for that output layer,the jth output layer representing the l layer,the ith input layer of the l-1 layer is represented, and the pooling layer comprises N input layers and N output layers;
wherein d isown () represents a down-sampling function, typically summing all pixels in different n × n regions of the input image, so that the output image is reduced by a factor of n in both dimensions, each output layer corresponding to a respective multiplicative offsetAnd an additive bias The jth output layer representing the l layer,the jth input layer representing the l-1 layer;
for the output full connection layer, convolution is input into multiple characteristic layers, and the convolution values are summed to obtain the output layer, using aijRepresenting the weight or contribution of the ith input layer in the jth output feature layer, the jth output layer can be represented as:
and the weight aijThe constraints need to be satisfied:
wherein N isinIndicates the number of layers of the input features,representing the weight between the input layer and the output layer,indicating the activation bias between the various layers,the jth output layer representing the l layer,the jth input layer representing the l-1 layer;
as a further preferable scheme of the method for detecting the marine vessel based on the multi-scale convolutional neural network model of the present invention, in step 3, the multi-scale feature fusion module specifically comprises:
step 3.1, extracting feature maps output by the convolution layers of layers 9, 12, 15, 17 and 19 in the whole neural network, and then splicing the 5 feature maps to form a new feature map; then, the new feature map is fed into two multi-scale loss function modulesAndcalculating a multi-scale loss function and updating network parameters; the multiscale loss function during training is defined as follows:
the multi-scale loss function in training consists of two multi-scale loss function modulesAnda composition in which i represents an element number in a feature map; p is a radical ofiIndicating the probability that the ith element in the feature map contains the target at the position corresponding to the original image,the probability of whether the original image position in the real label data contains the target or not is shown, and if the real value of the label is judged to be the real valueThe ith position in the feature map is a target, and if the ith position is 0, the ith position is not the target; t is tiIndicating the coordinate offset of the ith element in the feature map corresponding to the frame in the original imageThen represents the offset of the real frame coordinates; n is a radical ofclsAnd NregThe total number of target classes and the total number of frame coordinate offsets contained in the feature map, and the classification loss functionUsing a conventional softmax function, a regression loss functionThe used method is smoothL1loss:
Wherein the content of the first and second substances,which means that the regression loss function is calculated only for the candidate box with the target, lambda is an adjustable parameter, and is set to 3 by default, to balance the unbalanced influence of the ratio between positive and negative samples in the data on the final loss function.
A ship detection system for building a deep learning network model based on scene multi-dimensional features comprises the following steps:
the image acquisition module is used for constructing a ship image sample library, and comprises the steps of acquiring video data acquired by the unmanned aerial vehicle in the coastal region under visible light, extracting each frame of image and obtaining a ship position true value and length and width;
the data enhancement processing module is used for carrying out data enhancement processing on the data set, and comprises classic digital image processing algorithms such as inversion and scaling on the image;
the characteristic extraction module is used for constructing a multilayer convolutional neural network as a ship target detector, and inputting an image serving as sample data into the multilayer convolutional neural network after data enhancement to obtain a characteristic diagram;
and the training module is used for constructing a multi-scale convolution nerve unit, performing feature fusion on the plurality of feature maps, and training according to the real position of the ship in the data set to obtain a training result.
The technical scheme provided by the invention has the beneficial effects that:
(1) according to the actual data condition, various small ships and large ships coexist on the sea, and the large difference of the same kind of targets is the main reason of low detection recall rate. The invention integrates the characteristics of a plurality of characteristic graphs of the multilayer convolutional neural network, integrates the image detail information of the network shallow layer and the macroscopic semantic information of the network high layer, and improves the recall rate of small target detection while ensuring the original large target detection quality.
(2) According to the invention, an image data enhancement strategy is added in target detection, so that the robustness detection of the algorithm is improved. The method still has a good detection result for complex scenes such as cloud and fog, cloudy days, raining and the like. The method can be used for providing the ocean supervision work efficiency, saving the supervision cost, providing scientific basis for the formulation of ocean management decisions, and having important market value.
Drawings
FIG. 1 is a flow chart of an embodiment of the present invention;
fig. 2 is a schematic diagram of a network structure according to an embodiment of the present invention.
Detailed Description
The invention provides a ship detection method of a multi-scale feature fusion convolutional neural network. Firstly, an image sample library is constructed, and a ship image is subjected to sample marking to obtain a calibrated image sample. And then enhancing the data through digital image processing algorithms such as inversion and scaling, and constructing a deep learning network to carry out convolution on the image. And then constructing a multilayer convolutional neural network as a ship target detector, inputting the processed image as sample data into a deep learning network to obtain a feature map after convolution, constructing a multi-scale convolutional neural unit, performing feature fusion on the multilayer convolutional feature map, finally obtaining a loss function of a suggestion box by using a ship position truth value, training the whole network, and outputting a trained model. And finally, carrying out ship detection on the test data by using the trained model. The method mainly comprises five processes of sample library construction, image data enhancement, multi-scale feature fusion, deep learning network training and ship detection.
To illustrate the specific embodiments in detail, as shown in fig. 1, the flow of the example is as follows:
step a, constructing a ship image sample library; the data is enhanced through digital image processing algorithms such as inversion and scaling, and the number of the image data is expanded.
Firstly, ship images are prepared, and data required to be collected by the method are mainly video data shot by unmanned aerial vehicles in coastal areas under visible light. For the acquired video data, each frame of image needs to be obtained through a decoder or a code, and a ship image sample library with sufficient diversity is obtained for a plurality of videos. And then, obtaining a true value and a length and a width of the ship position through a preselected mark for each frame of image in the ship image sample library. And then, enhancing the image data through inversion and scaling, adding a new image formed after inversion and scaling into a sample library, and carrying out pre-selection marking and training together.
Wherein implementation details of scaling include:
(1) scaling the image to have its original aspect ratio with the longest side equal to 500 pixels;
(2) then, the short edges with the length less than 500 pixels are filled with gray, and finally a picture with the side length of 500 pixels is formed;
(3) if the longest edge of the original picture is less than 500 pixels, the gray padding operation is directly performed.
And b, constructing a multilayer convolutional neural network as a ship target detector, and inputting the processed image obtained in the step a into the multilayer convolutional neural network as sample data to obtain a characteristic diagram.
The network structure of the built multilayer convolutional neural network consists of 62 convolutional layers and 5 maximum pooling layers.
For the common convolution layer, a convolution core with a value being modified is used to convolute the feature layer of the previous layer, and then the output feature layer can be obtained by activating the function. Each output layer may be a combination of convolving the values of multiple input layers:
where Mj represents the set of input layers selected, i is the index value of the input layer cell, j is the index value of the output layer cell,representing the weight between the i-th layer input layer and the j-th layer output layer,represents the activation bias for the output layer of layer j, f () represents the activation function for that output layer,the jth output layer representing the l layer,the ith input layer representing the l-1 layer. For the pooling layer, there are N input layers and N output layers, except that each output layer is smaller.
down () represents a downsampling function. Typically, all pixels in different n × n regions of the input image are summed. So that the output image is reduced by a factor of n in both dimensions. Each output layer corresponds to a respective multiplicative biasAnd an additive bias The jth output layer representing the l layer,the jth input layer representing the l-1 layer.
For the output fully connected layer, it is often better to convolve the input multiple feature layers and then sum the convolved values to obtain the output layer. Example of the invention usesijIndicating the weight or contribution of the ith input layer in obtaining the jth output feature layer. Thus, the jth output layer can be represented as:
and the weight aijThe constraints need to be satisfied:
wherein N isinIndicates the number of layers of the input features,representing the weight between the input layer and the output layer,indicating the activation bias between the various layers,the jth output layer representing the l layer,the jth input layer representing the l-1 layer.
And c, constructing a multi-scale convolution nerve training unit, performing feature fusion on a plurality of feature maps based on the plurality of feature maps obtained in the step b, and finally training according to the real position of the ship obtained in the step a to obtain a training model.
As shown in fig. 2, the workflow of the multi-scale feature fusion module is as follows:
firstly, extracting feature maps output by the convolution layers of layers 9, 12, 15, 17 and 19 in the whole neural network, and then splicing the 5 feature maps to form a new feature map; then, the new feature map is fed into two multi-scale loss function modulesAndand calculating a multi-scale loss function and updating network parameters. The multiscale loss function during training is defined as follows:
the function is composed of two multi-scale loss function modulesAndand (4) forming. Wherein i represents the element number in the feature map; p is a radical ofiIndicating the probability that the ith element in the feature map contains the target at the position corresponding to the original image,indicating the probability of whether the original image position in the real label data contains the target or not, if the label real valueIt means that the ith position in the feature map is the target, and if it is 0Then it is not; t is tiIndicating the coordinate offset of the ith element in the feature map corresponding to the frame in the original imageAn offset of the real frame coordinates is indicated. N is a radical ofclsAnd NregThe total number of object classes and the total number of frame coordinate offsets included in the feature map are provided. Classification loss functionThe conventional softmax function is used. Function of regression lossThe used method is smoothL1loss:
WhereinIndicating that the regression loss function is calculated only for the candidate box with the target. λ is an adjustable parameter, set to 3 by default, to balance the unbalanced effect of the ratio between positive and negative samples in the data on the final loss function.
So far, the detailed implementation process of the marine vessel detection method based on the multi-scale feature fusion convolutional neural network used in the embodiment of the application is introduced. In specific implementation, the method provided by the invention can realize automatic operation flow based on software technology, and can also realize a corresponding system in a modularized mode.
The embodiment of the invention also provides a ship detection system for constructing a deep learning network model based on scene multi-dimensional features, which comprises the following modules:
the image acquisition module is used for constructing a ship image sample library, and comprises the steps of acquiring video data acquired by the unmanned aerial vehicle in the coastal region under visible light, extracting each frame of image and obtaining a ship position true value and length and width;
the data enhancement processing module is used for carrying out data enhancement processing on the data set, and comprises classic digital image processing algorithms such as inversion and scaling on the image;
the characteristic extraction module is used for constructing a multilayer convolutional neural network as a ship target detector, and inputting an image serving as sample data into the multilayer convolutional neural network after data enhancement to obtain a characteristic diagram;
and the training module is used for constructing a multi-scale convolution nerve unit, performing feature fusion on the plurality of feature maps, and training according to the real position of the ship in the data set to obtain a training result. The specific implementation of each module can refer to the corresponding step, and the detailed description of the invention is omitted. The specific examples described herein are merely illustrative of the spirit of the invention. Various modifications or additions may be made or substituted in a similar manner to the specific embodiments described herein by those skilled in the art without departing from the spirit of the invention or exceeding the scope thereof as defined in the appended claims.
Claims (5)
1. A method for detecting a marine vessel based on a multi-scale convolutional neural network model is characterized by comprising the following steps:
step 1, constructing a ship image sample library, collecting ship video data of a coastal area under visible light, extracting each frame of image, obtaining a true value and a length and a width of a ship position, and performing enhancement processing on the ship video data through a digital image processing algorithm;
step 2, constructing a multilayer convolutional neural network as a ship target detector, inputting the processed image obtained in the step 1 into a deep learning network as sample data, and obtaining a feature map after convolution;
and 3, constructing a multi-scale convolution neural unit, performing feature fusion on the multilayer convolution feature maps based on the feature maps obtained in the step 2 after convolution, and training according to the real position of the ship to further obtain a training model.
2. The method of claim 1, wherein the method comprises the following steps: in step 1, the data enhancement process includes:
step 1.1, scaling the image according to the original length-width ratio until the longest edge is equal to 500 pixels;
step 1.2, then, the short edges with the length less than 500 pixels are filled with gray to form a picture with the side length of 500 pixels;
and step 1.3, if the longest edge of the original picture is less than 500 pixels, performing gray compensation.
3. The method of claim 1, wherein the method comprises the following steps: in the step 2, the process is carried out,
the network structure of the constructed multilayer convolutional neural network consists of 62 convolutional layers and 5 pooling layers;
for the common convolutional layer, carrying out convolution on the feature layer of the previous layer by using a convolution core with a modifiable numerical value, and obtaining an output feature layer through an activation function;
where Mj represents the set of input layers selected, i is the index value of the input layer cell, j is the index value of the output layer cell,representing the weight between the i-th layer input layer and the j-th layer output layer,represents the activation bias for the output layer of layer j, f () represents the activation function for that output layer,the jth output layer representing the l layer,the ith input layer of the l-1 layer is represented, and the pooling layer comprises N input layers and N output layers;
wherein down () represents a down-sampling function, which is the summation of all pixels in different n × n regions of the input image, such that the output image is reduced by n times in both dimensions, and each output layer corresponds to a respective multiplicative biasAnd an additive bias The jth output layer representing the l layer,the jth input layer representing the l-1 layer;
for the output full-connection layer, convolution is input into a plurality of characteristic layers, and the convolution values are summed to obtain the output layer if aijRepresenting the weight or contribution of the ith input layer in the jth output feature layer, the jth output layer can be represented as:
and a isijThe constraints need to be satisfied:
4. The method of claim 1, wherein the method comprises the following steps: in step 3, feature fusion comprises:
step 3.1, extracting feature maps output by the convolution layers of layers 9, 12, 15, 17 and 19 in the whole neural network, and then splicing the 5 feature maps to form a new feature map; then, the new characteristic graphs are respectively input into two multi-scale loss function modulesAndcalculating a multi-scale loss function and updating network parameters; the multiscale loss function during training is defined as follows:
the multi-scale loss function in training consists of two multi-scale loss function modulesAnda composition in which i represents an element number in a feature map; p is a radical ofiIndicating the probability that the ith element in the feature map contains the target at the position corresponding to the original image,the probability of whether the original image position in the real label data contains the target or not is shown, and if the real value of the label is judged to be the real valueThe ith position in the feature map is a target, and if the ith position is 0, the ith position is not the target; t is tiIndicating the coordinate offset of the ith element in the feature map corresponding to the frame in the original imageThen represents the offset of the real frame coordinates; n is a radical ofclsAnd NregThe total number of target classes and the total number of frame coordinate offsets contained in the feature map, and the classification loss functionUsing a conventional softmax function, a regression loss functionThe used method is smoothL1loss:
Wherein the content of the first and second substances,meaning that the regression loss function is calculated only for the candidate box with the target, λ is an adjustable parameter, and is set to 3 by default, to balance the ratio between positive and negative samples in the dataUnbalanced effects on the final loss function.
5. A ship detection system for building a deep learning network model based on scene multi-dimensional features according to any one of claims 1 to 4, the system comprising:
the image acquisition module is used for constructing a ship image sample library, and comprises the steps of acquiring video data acquired by the unmanned aerial vehicle in the coastal region under visible light, extracting each frame of image and obtaining a ship position true value and length and width;
the data enhancement processing module is used for carrying out data enhancement processing on the data set, and comprises classic digital image processing algorithms such as inversion and scaling on the image;
the characteristic extraction module is used for constructing a multilayer convolutional neural network as a ship target detector, and inputting an image serving as sample data into the multilayer convolutional neural network after data enhancement to obtain a characteristic diagram;
and the training module is used for constructing a multi-scale convolution nerve unit, performing feature fusion on the plurality of feature maps, and training according to the real position of the ship in the data set to obtain a training result.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910930804.4A CN110796009A (en) | 2019-09-29 | 2019-09-29 | Method and system for detecting marine vessel based on multi-scale convolution neural network model |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910930804.4A CN110796009A (en) | 2019-09-29 | 2019-09-29 | Method and system for detecting marine vessel based on multi-scale convolution neural network model |
Publications (1)
Publication Number | Publication Date |
---|---|
CN110796009A true CN110796009A (en) | 2020-02-14 |
Family
ID=69438678
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910930804.4A Pending CN110796009A (en) | 2019-09-29 | 2019-09-29 | Method and system for detecting marine vessel based on multi-scale convolution neural network model |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110796009A (en) |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111814536A (en) * | 2020-05-21 | 2020-10-23 | 闽江学院 | Breeding monitoring method and device |
CN111967305A (en) * | 2020-07-01 | 2020-11-20 | 华南理工大学 | Real-time multi-scale target detection method based on lightweight convolutional neural network |
CN112016542A (en) * | 2020-05-08 | 2020-12-01 | 珠海欧比特宇航科技股份有限公司 | Urban waterlogging intelligent detection method and system |
CN112036404A (en) * | 2020-08-31 | 2020-12-04 | 上海大学 | Target detection method and system for offshore ship |
CN112085001A (en) * | 2020-09-23 | 2020-12-15 | 清华大学苏州汽车研究院(相城) | Tunnel recognition model and method based on multi-scale edge feature detection |
CN112183463A (en) * | 2020-10-23 | 2021-01-05 | 珠海大横琴科技发展有限公司 | Ship identification model verification method and device based on radar image |
CN112183232A (en) * | 2020-09-09 | 2021-01-05 | 上海鹰觉科技有限公司 | Ship board number position positioning method and system based on deep learning |
CN112285712A (en) * | 2020-10-15 | 2021-01-29 | 电子科技大学 | Method for improving detection precision of ship on shore in SAR image |
CN112464765A (en) * | 2020-09-10 | 2021-03-09 | 天津师范大学 | Safety helmet detection algorithm based on single-pixel characteristic amplification and application thereof |
CN112986210A (en) * | 2021-02-10 | 2021-06-18 | 四川大学 | Scale-adaptive microbial Raman spectrum detection method and system |
CN113807386A (en) * | 2021-07-21 | 2021-12-17 | 广东工业大学 | Target detection method and system fusing multi-scale information and computer equipment |
CN114720957A (en) * | 2022-06-08 | 2022-07-08 | 中国人民解放军空军预警学院 | Radar target detection method and system and storable medium |
CN114943888A (en) * | 2022-03-24 | 2022-08-26 | 中国人民解放军海军大连舰艇学院 | Sea surface small target detection method based on multi-scale information fusion, electronic equipment and computer readable medium |
CN116051548A (en) * | 2023-03-14 | 2023-05-02 | 中国铁塔股份有限公司 | Positioning method and device |
CN116883913A (en) * | 2023-09-05 | 2023-10-13 | 长江信达软件技术(武汉)有限责任公司 | Ship identification method and system based on video stream adjacent frames |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107818326A (en) * | 2017-12-11 | 2018-03-20 | 珠海大横琴科技发展有限公司 | A kind of ship detection method and system based on scene multidimensional characteristic |
CN109345508A (en) * | 2018-08-31 | 2019-02-15 | 北京航空航天大学 | A kind of Assessing Standards For Skeletal method based on two stages neural network |
CN109522963A (en) * | 2018-11-26 | 2019-03-26 | 北京电子工程总体研究所 | A kind of the feature building object detection method and system of single-unit operation |
CN109598290A (en) * | 2018-11-22 | 2019-04-09 | 上海交通大学 | A kind of image small target detecting method combined based on hierarchical detection |
CN109740515A (en) * | 2018-12-29 | 2019-05-10 | 科大讯飞股份有限公司 | One kind reading and appraising method and device |
CN110110783A (en) * | 2019-04-30 | 2019-08-09 | 天津大学 | A kind of deep learning object detection method based on the connection of multilayer feature figure |
-
2019
- 2019-09-29 CN CN201910930804.4A patent/CN110796009A/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107818326A (en) * | 2017-12-11 | 2018-03-20 | 珠海大横琴科技发展有限公司 | A kind of ship detection method and system based on scene multidimensional characteristic |
CN109345508A (en) * | 2018-08-31 | 2019-02-15 | 北京航空航天大学 | A kind of Assessing Standards For Skeletal method based on two stages neural network |
CN109598290A (en) * | 2018-11-22 | 2019-04-09 | 上海交通大学 | A kind of image small target detecting method combined based on hierarchical detection |
CN109522963A (en) * | 2018-11-26 | 2019-03-26 | 北京电子工程总体研究所 | A kind of the feature building object detection method and system of single-unit operation |
CN109740515A (en) * | 2018-12-29 | 2019-05-10 | 科大讯飞股份有限公司 | One kind reading and appraising method and device |
CN110110783A (en) * | 2019-04-30 | 2019-08-09 | 天津大学 | A kind of deep learning object detection method based on the connection of multilayer feature figure |
Non-Patent Citations (2)
Title |
---|
刘云等: ""深度学习的多尺度多人目标检测方法研究"", 《计算机工程与应用》 * |
顾亚风: ""面向图像内容检索的卷积神经网络"", 《中国优秀博硕士学位论文全文数据库(硕士)信息科技辑》 * |
Cited By (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112016542A (en) * | 2020-05-08 | 2020-12-01 | 珠海欧比特宇航科技股份有限公司 | Urban waterlogging intelligent detection method and system |
CN111814536B (en) * | 2020-05-21 | 2023-11-28 | 闽江学院 | Culture monitoring method and device |
CN111814536A (en) * | 2020-05-21 | 2020-10-23 | 闽江学院 | Breeding monitoring method and device |
CN111967305B (en) * | 2020-07-01 | 2022-03-18 | 华南理工大学 | Real-time multi-scale target detection method based on lightweight convolutional neural network |
CN111967305A (en) * | 2020-07-01 | 2020-11-20 | 华南理工大学 | Real-time multi-scale target detection method based on lightweight convolutional neural network |
CN112036404A (en) * | 2020-08-31 | 2020-12-04 | 上海大学 | Target detection method and system for offshore ship |
CN112036404B (en) * | 2020-08-31 | 2024-01-02 | 上海大学 | Marine ship target detection method and system |
CN112183232A (en) * | 2020-09-09 | 2021-01-05 | 上海鹰觉科技有限公司 | Ship board number position positioning method and system based on deep learning |
CN112464765B (en) * | 2020-09-10 | 2022-09-23 | 天津师范大学 | Safety helmet detection method based on single-pixel characteristic amplification and application thereof |
CN112464765A (en) * | 2020-09-10 | 2021-03-09 | 天津师范大学 | Safety helmet detection algorithm based on single-pixel characteristic amplification and application thereof |
CN112085001A (en) * | 2020-09-23 | 2020-12-15 | 清华大学苏州汽车研究院(相城) | Tunnel recognition model and method based on multi-scale edge feature detection |
CN112085001B (en) * | 2020-09-23 | 2024-04-23 | 清华大学苏州汽车研究院(相城) | Tunnel identification model and method based on multi-scale edge feature detection |
CN112285712A (en) * | 2020-10-15 | 2021-01-29 | 电子科技大学 | Method for improving detection precision of ship on shore in SAR image |
CN112285712B (en) * | 2020-10-15 | 2023-09-15 | 电子科技大学 | Method for improving detection precision of coasting ship in SAR image |
CN112183463A (en) * | 2020-10-23 | 2021-01-05 | 珠海大横琴科技发展有限公司 | Ship identification model verification method and device based on radar image |
CN112183463B (en) * | 2020-10-23 | 2021-10-15 | 珠海大横琴科技发展有限公司 | Ship identification model verification method and device based on radar image |
CN112986210A (en) * | 2021-02-10 | 2021-06-18 | 四川大学 | Scale-adaptive microbial Raman spectrum detection method and system |
CN113807386B (en) * | 2021-07-21 | 2023-08-01 | 广东工业大学 | Target detection method, system and computer equipment integrating multi-scale information |
CN113807386A (en) * | 2021-07-21 | 2021-12-17 | 广东工业大学 | Target detection method and system fusing multi-scale information and computer equipment |
CN114943888A (en) * | 2022-03-24 | 2022-08-26 | 中国人民解放军海军大连舰艇学院 | Sea surface small target detection method based on multi-scale information fusion, electronic equipment and computer readable medium |
CN114720957A (en) * | 2022-06-08 | 2022-07-08 | 中国人民解放军空军预警学院 | Radar target detection method and system and storable medium |
CN116051548A (en) * | 2023-03-14 | 2023-05-02 | 中国铁塔股份有限公司 | Positioning method and device |
CN116051548B (en) * | 2023-03-14 | 2023-08-11 | 中国铁塔股份有限公司 | Positioning method and device |
CN116883913B (en) * | 2023-09-05 | 2023-11-21 | 长江信达软件技术(武汉)有限责任公司 | Ship identification method and system based on video stream adjacent frames |
CN116883913A (en) * | 2023-09-05 | 2023-10-13 | 长江信达软件技术(武汉)有限责任公司 | Ship identification method and system based on video stream adjacent frames |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110796009A (en) | Method and system for detecting marine vessel based on multi-scale convolution neural network model | |
CN113065558B (en) | Lightweight small target detection method combined with attention mechanism | |
CN109584248B (en) | Infrared target instance segmentation method based on feature fusion and dense connection network | |
CN110310241B (en) | Method for defogging traffic image with large air-light value by fusing depth region segmentation | |
CN109871798A (en) | A kind of remote sensing image building extracting method based on convolutional neural networks | |
CN111046880A (en) | Infrared target image segmentation method and system, electronic device and storage medium | |
CN111832443B (en) | Construction method and application of construction violation detection model | |
CN117078943B (en) | Remote sensing image road segmentation method integrating multi-scale features and double-attention mechanism | |
CN109919223B (en) | Target detection method and device based on deep neural network | |
CN115205264A (en) | High-resolution remote sensing ship detection method based on improved YOLOv4 | |
CN110490155B (en) | Method for detecting unmanned aerial vehicle in no-fly airspace | |
CN113160062A (en) | Infrared image target detection method, device, equipment and storage medium | |
CN116311254B (en) | Image target detection method, system and equipment under severe weather condition | |
CN115035295B (en) | Remote sensing image semantic segmentation method based on shared convolution kernel and boundary loss function | |
CN113420794B (en) | Binaryzation Faster R-CNN citrus disease and pest identification method based on deep learning | |
CN111353396A (en) | Concrete crack segmentation method based on SCSEOCUnet | |
CN114359245A (en) | Method for detecting surface defects of products in industrial scene | |
CN116071676A (en) | Infrared small target detection method based on attention-directed pyramid fusion | |
CN115527096A (en) | Small target detection method based on improved YOLOv5 | |
CN116469020A (en) | Unmanned aerial vehicle image target detection method based on multiscale and Gaussian Wasserstein distance | |
CN112785610B (en) | Lane line semantic segmentation method integrating low-level features | |
CN117372829A (en) | Marine vessel target identification method, device, electronic equipment and readable medium | |
CN111950476A (en) | Deep learning-based automatic river channel ship identification method in complex environment | |
CN116977866A (en) | Lightweight landslide detection method | |
CN116452900A (en) | Target detection method based on lightweight neural network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20200214 |