CN111898699A - Automatic detection and identification method for hull target - Google Patents
Automatic detection and identification method for hull target Download PDFInfo
- Publication number
- CN111898699A CN111898699A CN202010802211.2A CN202010802211A CN111898699A CN 111898699 A CN111898699 A CN 111898699A CN 202010802211 A CN202010802211 A CN 202010802211A CN 111898699 A CN111898699 A CN 111898699A
- Authority
- CN
- China
- Prior art keywords
- module
- convolution
- data set
- hull
- training data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 46
- 238000000034 method Methods 0.000 title claims abstract description 36
- 238000012549 training Methods 0.000 claims abstract description 74
- 238000003062 neural network model Methods 0.000 claims abstract description 33
- 238000012360 testing method Methods 0.000 claims abstract description 12
- 208000037170 Delayed Emergence from Anesthesia Diseases 0.000 claims description 16
- 230000006870 function Effects 0.000 claims description 16
- 230000004913 activation Effects 0.000 claims description 15
- 230000004927 fusion Effects 0.000 claims description 15
- 238000005070 sampling Methods 0.000 claims description 12
- 238000011176 pooling Methods 0.000 claims description 9
- 230000003044 adaptive effect Effects 0.000 claims description 4
- 238000000137 annealing Methods 0.000 claims description 4
- 230000009466 transformation Effects 0.000 claims description 4
- 230000001629 suppression Effects 0.000 claims 1
- 238000013528 artificial neural network Methods 0.000 description 9
- 238000000605 extraction Methods 0.000 description 7
- 230000008569 process Effects 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000005764 inhibitory process Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/07—Target detection
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computational Linguistics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Evolutionary Biology (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Image Analysis (AREA)
Abstract
The invention relates to a hull target automatic detection and identification method, which comprises the following steps: step 1: acquiring a training data set and a test data set; step 2: constructing a ship body recognition neural network model; and step 3: expanding the training data set to obtain an expanded training data set, and training the neural network model by using the expanded training data set; and 4, step 4: testing the neural network model by using the test data set, judging whether the model precision meets the preset precision, if so, executing the step 5, otherwise, returning to the step 3; and 5: and carrying out automatic detection and identification on the ship body by using the trained ship body identification neural network model. Compared with the prior art, the method has the advantages of high identification precision, high identification speed and the like.
Description
Technical Field
The invention relates to the technical field of automatic identification of hull targets under a navigation condition, in particular to an automatic detection and identification method of hull targets.
Background
With the development of computer vision, especially in the field of target detection in recent years, more and more automatic driving systems are equipped with automatic image recognition systems to ensure the safety of unmanned driving, and similarly, in the field of navigation, the automatic navigation of unmanned ships also needs to automatically recognize the ship body to ensure the safety of the ship-moving task.
The existing target detection algorithms rarely carry out target identification training specially aiming at the ship body, and the algorithms are not high in identification precision and speed of the ship body, so that the ship body identification precision and speed are improved by a characteristic target detection algorithm aiming at the ship.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and provide an automatic detection and identification method for a ship hull target, which is based on yolov3 and has the advantages of improvement, high identification precision and high identification speed on the basis.
The purpose of the invention can be realized by the following technical scheme:
a ship hull target automatic detection and identification method comprises the following steps:
step 1: acquiring a training data set and a test data set;
step 2: constructing a ship body recognition neural network model;
and step 3: expanding the training data set to obtain an expanded training data set, and training the neural network model by using the expanded training data set;
and 4, step 4: testing the neural network model by using the test data set, judging whether the model precision meets the preset precision, if so, executing the step 5, otherwise, returning to the step 3;
and 5: and carrying out automatic detection and identification on the ship body by using the trained ship body identification neural network model.
Preferably, the method for acquiring the training data set comprises:
the hull pictures are labeled using LabelImg, and then processed into txt format files using python, all of which constitute the training data set.
Preferably, the method for expanding the training data set specifically comprises:
and mixing different pictures in the training data set with image pixels by using a Mixup image enhancement mode to expand the data set, then carrying out scale transformation on the expanded training data set to expand the training data set again, and finally taking the training data set subjected to two times of expansion as a final expanded training data set.
Preferably, the training of the hull recognition neural network model in the step 3 adopts a mixed precision training mode.
More preferably, the mixing precision training mode is specifically as follows:
the 16-bit half-precision weight is used for calculating the activation function part, and the 32-bit single-precision weight is used for updating.
Preferably, the hull recognition neural network model comprises an improved CSPDarknet53 framework, a second convolution module, an SPP pooling layer, a feature fusion module, an output layer and an anchor frame detection module which are connected in sequence; the improved CSPDarknet53 framework is also connected with an output layer;
the improved CSPDarknet53 framework comprises a first convolution module, a first residual error module, a second residual error module, a third residual error module, a fourth residual error module and a fifth residual error module which are sequentially connected;
the output layer comprises a first splicing module, a third convolution module, a fourth convolution module, a second splicing module, a fifth convolution module, a sixth convolution module, a first up-sampling module, a seventh convolution module and a second up-sampling module;
the input end of the first splicing module is connected with the third residual error module, and the output end of the first splicing module is sequentially connected with the third convolution module, the fourth convolution module and the anchor frame detection module; the input end of the second splicing module is connected with the fourth residual error module, and the output end of the second splicing module is sequentially connected with the fifth convolution module, the sixth convolution module and the anchor frame detection module; the input end of the seventh convolution module is connected with the characteristic fusion module, and the output end of the seventh convolution module is connected with the anchor frame detection module;
the input end of the first up-sampling module is connected with the output end of the fifth convolution module, and the output end of the first up-sampling module is connected with the input end of the first splicing module; the input end of the second up-sampling module is connected with the output end of the seventh convolution module, and the output end of the second up-sampling module is connected with the input end of a second splicing module (504);
the feature fusion module is three adaptive spatial feature fusion structures (ASFFs) which are sequentially connected.
More preferably, the first convolution module, the second convolution module, the third convolution module, the fourth convolution module, the fifth convolution module and the sixth convolution module are respectively composed of a plurality of convolution blocks; the convolution block comprises a convolution layer and a DropBlock regularization unit; the output end of the convolution layer is connected with the input end of the DropBlock;
the first residual error module, the second residual error module, the third residual error module, the fourth residual error module and the fifth residual error module are respectively composed of a plurality of residual error blocks; the residual blocks are represented as residual blocks n, and the expression of the residual block is as follows:
residual block n input + convolution block + residual unit n
The residual unit is specifically a residual network containing two convolution blocks.
More preferably, the convolutional layer takes the form of a convolution of Mixconv2 d;
the activation function of the convolutional layer is specifically a Mish activation function, and specifically includes:
Mish=x*tanh(ln1+e^x)。
more preferably, the hull recognition neural network model adopts a cosine annealing mode to realize automatic attenuation of the learning rate.
More preferably, the anchor frame detection module comprises three anchor frame modules, and the modules are obtained by clustering a training set; three outputs obtained by the hull recognition neural network model are respectively input into the three anchor frame modules to obtain three outputs related to the coordinate information and the probability of the hull; and then, respectively carrying out non-maximum value inhibition on the three outputs to obtain the same proper output, if the probability of identifying the ship output by the anchor frame detection module is greater than the set basic probability, judging that the ship is detected by the output, and finally outputting the coordinate information output by the anchor frame detection module as the position information of the ship in the image.
Compared with the prior art, the invention has the following advantages:
firstly, the recognition precision is high, and the recognition speed is fast: the target detection and identification method of the invention uses the hull recognition neural network model to automatically recognize the hull target, and adopts the Mixup image enhancement mode to expand the training data set, so that the training data set is more comprehensive, and the recognition precision of the hull recognition neural network model obtained by training is higher; the hull recognition neural network model simplifies the number of convolution layers, and simultaneously, a Mish activation function is selected, so that the recognition speed of the model is improved.
Secondly, perfecting the model structure: the use of the spp structure improves the feature extraction capability of the neural network and reduces the possibility of overfitting, the use of the DenseNet connection method improves the gradient back propagation capability, so that the network is easier to train, the use of the MixConv2d convolution method improves the feature extraction capability of convolution, so that the structure of the hull recognition neural network model is more perfect, and the hull recognition method is more suitable for hull recognition.
Drawings
FIG. 1 is a schematic flow chart of a hull target automatic detection and identification method according to the present invention;
fig. 2 is a schematic structural diagram of a hull recognition neural network model in the invention.
The reference numbers in the figures indicate:
1. the improved CSPDarknet53 framework, 2, a second convolution module, 3, an SPP pooling layer, 4, a feature fusion module, 5, an output layer, 6, an anchor frame detection module, 101, a first convolution module, 102, a first residual module, 103, a second residual module, 104, a third residual module, 105, a fourth residual module, 106, a fifth residual module, 501, a first splicing module, 502, a third convolution module, 503, a fourth convolution module, 504, a second splicing module, 505, a fifth convolution module, 506, a sixth convolution module, 507, a first upsampling module, 508, a seventh convolution module, 509, and a second upsampling module.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, not all, embodiments of the present invention. All other embodiments, which can be obtained by a person skilled in the art without any inventive step based on the embodiments of the present invention, shall fall within the scope of protection of the present invention.
An automatic detection and identification method for a ship body target is structurally shown in fig. 1 and comprises the following steps:
step 1: acquiring a training data set and a test data set;
step 2: constructing a ship body recognition neural network model;
and step 3: expanding the training data set to obtain an expanded training data set, and training the neural network model by using the expanded training data set;
and 4, step 4: testing the neural network model by using the test data set, judging whether the model precision meets the preset precision, if so, executing the step 5, otherwise, returning to the step 3;
and 5: and carrying out automatic detection and identification on the ship body by using the trained ship body identification neural network model.
The method for acquiring the training data set in the embodiment comprises the following steps: the hull pictures are labeled using LabelImg, and then processed into txt format files using python, all of which constitute the training data set.
The method for expanding the training data set in the embodiment comprises the following steps: and mixing different pictures in the training data set with image pixels by using a Mixup image enhancement mode to expand the data set, then carrying out scale transformation on the expanded training data set to expand the training data set again, and finally taking the training data set subjected to two times of expansion as a final expanded training data set.
The training of the hull recognition neural network model in the embodiment adopts a mixed precision training mode, which specifically comprises the following steps: the 16-bit half-precision weight is used for calculating the activation function part, and the 32-bit single-precision weight is used for updating.
The structure of the hull recognition neural network model in this embodiment is shown in fig. 2, and includes an improved CSPDarknet53 framework 1, a second convolution module 2, an SPP pooling layer 3, a feature fusion module 4, an output layer 5, and an anchor frame detection module 6, which are connected in sequence, and the improved CSPDarknet53 framework 1 is further connected to the output layer 5.
The improved CSPDarknet53 framework 1 includes a first convolution module 101, a first residual module 102, a second residual module 103, a third residual module 104, a fourth residual module 105, and a fifth residual module 106, which are connected in sequence.
The output layer 5 comprises a first concatenation module 501, a third convolution module 502, a fourth convolution module 503, a second concatenation module 504, a fifth convolution module 505, a sixth convolution module 506, a first upsampling module 507, a seventh convolution module 508 and a second upsampling module 509.
The input end of the first splicing module 501 is connected with the third residual error module 104, the output end of the first splicing module 501 is sequentially connected with the third convolution module 502, the fourth convolution module 503 and the anchor frame detection module 6, the input end of the second splicing module 505 is connected with the fourth residual error module 105, the output end of the second splicing module 505 is sequentially connected with the fifth convolution module 505, the sixth convolution module 506 and the anchor frame detection module 6, the input end of the seventh convolution module 508 is connected with the feature fusion module 4, and the output end of the seventh convolution module 508 is connected with the anchor frame detection module 6.
The input end of the first upsampling module 507 is connected to the output end of the fifth convolution module 505, the output end of the first upsampling module 507 is connected to the input end of the first splicing module 501, the input end of the second upsampling module 509 is connected to the output end of the seventh convolution module 508, and the output end of the second upsampling module 509 is connected to the input end of the second splicing module 504.
The feature fusion module 4 in this example is three adaptive spatial feature fusion constructs ASFF connected in series.
The convolution modules in this embodiment, that is, the first convolution module 101, the second convolution module 2, the third convolution module 502, the fourth convolution module 503, the fifth convolution module 505, and the sixth convolution module 506, are respectively composed of a plurality of convolution blocks, each convolution block includes a convolution layer and a dryblock regularization unit, and an output end of the convolution layer is connected to an input end of the dryblock;
the first residual module 102, the second residual module 103, the third residual module 104, the fourth residual module 105 and the fifth residual module 106 are respectively composed of a plurality of residual blocks, the plurality of residual blocks are represented as residual blocks x n, and the expression of the residual modules is as follows:
residual block n input + convolution block + residual unit n
The residual unit is specifically a residual network containing two convolution blocks.
The convolutional layer in this embodiment adopts a convolutional form of Mixconv2d, and the activation function is a Mish function, specifically:
Mish=x*tanh(ln1+e^x)
the hull recognition neural network model in the embodiment adopts a cosine annealing mode to realize automatic attenuation of the learning rate.
The following describes the model structure in detail:
one, improve CSPDarknet53 framework: the CSPDarknet53 frame-based target identification structure simplifies the middle neural network structure based on the existing CSPDarknet53 frame, so that the number of convolutional layers is reduced from 75 layers to 43 layers, the training process can be accelerated, and the detection performance cannot be reduced due to the reduction of convolutional layers because only the target detection of a single ship body is performed.
Secondly, SPP pooling layer: spatial Pyramid Pooling, which is a Spatial Pyramid Pooling layer, can better extract image features than a mature Pooling structure.
Thirdly, convolutional layer using the form of a Mixconv2d convolution: the Mixconv2d structure is a mixed receptive field convolution structure, and the capability of feature extraction in the image convolution process is effectively improved.
Fourthly, Mish activation function: the activation function allows better information to be driven down to the neural network, resulting in better accuracy and generalization.
Fifthly, a Mixup image enhancement technology: the image classification accuracy of 1 percent can be stably improved under the condition of almost no extra calculation overhead.
Sixthly, an adaptive spatial feature fusion structure ASFF: inconsistency in the training process is filtered by learning self-adaptive spatial fuzzy weight values, powerful baseline and tiny thrust overhead are obviously improved, and the most advanced speed and precision balance is realized among all single-point detectors.
Seventhly, Mixed Precision Training: the same precision can be achieved under the condition that the memory is reduced by half, and the training efficiency is improved.
Eight, DropBlock regularization: in order to discard local semantic information for more effectiveness, a network is stimulated to learn more robust and effective characteristics, a mode of discarding according to blocks is adopted, and the method is different from a Dropout method in that information of pixel points in a certain area of an image is lost together, so that the reuse rate of the information can be accelerated.
Ninthly, attenuation of cosine annealing learning rate: an automatic learning rate attenuation method can enable an algorithm to have a high learning rate during training and air storage, and a relatively low learning rate at the later stage of training and training, and receive a preset minimum value more highly.
The structural composition of each module is shown in table 1.
TABLE 1 structural composition of the various modules
Module | Structural assembly | Module | Structural assembly |
First convolution module 101 | (608,608,32)*3 | Seventh convolution module 508 | (19,19,1024)*2 |
Second convolution module 2 | (19,19,1024)*2 | First residual module 102 | (304,304,64)*1 |
Third convolution module 502 | (76,76,256)*2 | Second residual module 103 | (152,152,128)*2 |
Fourth convolution module 503 | (76,76,512)*2 | Third residual module 104 | (76,76,256)*2 |
Fifth convolution Module 505 | (38,38,512)*2 | Fourth residual module 105 | (38,38,512)*2 |
Sixth convolution module 506 | (38,38,1024)*2 | Fifth residual module 106 | (19,19,1024)*2 |
In the Training stage of the ship hull recognition neural network model in the embodiment, different Training data set pictures are Mixed on image pixels by using a Mixup image enhancement mode to expand a data set, the expanded Training set is subjected to scale transformation to expand the data set again, the data set after two times of expansion is transmitted to a neural network for Training, the Training adopts a Mixed Precision Training mode Mixed Precision Training, the embodiment is that 32-bit weight during Training is cut into 16 bits to be used for calculating a part of an activation function, but 32-bit weight is adopted during updating, so that the loss of semi-Precision information can be controlled, during Training, the picture firstly passes through 3 convolutional layers (including an extraction layer) to extract features, all convolutional layers adopt a convolutional form of Mixconv2d, in the convolutional layers, firstly, the image features are extracted by convolution, the activation functions of the convolution layers are changed from the traditional relu activation functions into the mish activation functions, the last part of the convolution layers is normalized through a Drop regularization unit, the output is used for feature extraction through the extraction layers of three residual blocks, after the output passes through the two residual blocks, one part of the output is transmitted into the output layer to serve as high-scale output (76), the other part of the output is transmitted downwards along the neural network structure, one part of the result obtained after the output passes through the two residual blocks is transmitted into the output layer to serve as medium-scale output (38) and the other part of the result is transmitted downwards along the neural network structure, one part of the result obtained after the output passes through the two residual blocks is transmitted into the output layer to serve as low-scale output (19), the output passes through the three convolution blocks and is transmitted downwards along the neural network structure to pass through an SPP pooling layer 2 and sequentially pass through 3 ASFF extraction structures The output obtained after the feature fusion module 4 is connected and formed is used as a third output after passing through one convolution block, the third output can be simultaneously used as an input to be spliced with the original 38 × 38 output after being expanded to the size of 38 × 38 along the upper sampling layer, and then is used as a second output after being spliced with the original 76 × 76 output after being expanded to the size of 76 × 76 along the upper sampling layer, and then is used as a first output after being subjected to two convolution blocks, and then the first output, the second output and the third output are multiplied by three anchors box corresponding to the first output respectively to obtain a final output, wherein the formats of all the outputs are the same and comprise: and if the probability c is greater than a basic set value of 0.5, the ships can be judged as the real detected ship results, and x, y, h and w can be output as the coordinate positions of the ship in the image.
In step 3, the number of the types to be detected in the original yolo algorithm is changed into 1 by the network architecture, and the ship body recognition neural network is built according to the method, so that the structure of the neural network and the number of layers of the convolutional layers are reduced.
While the invention has been described with reference to specific embodiments, the invention is not limited thereto, and various equivalent modifications and substitutions can be easily made by those skilled in the art within the technical scope of the invention. Therefore, the protection scope of the present invention shall be subject to the protection scope of the claims.
Claims (10)
1. A ship body target automatic detection and identification method is characterized by comprising the following steps:
step 1: acquiring a training data set and a test data set;
step 2: constructing a ship body recognition neural network model;
and step 3: expanding the training data set to obtain an expanded training data set, and training the neural network model by using the expanded training data set;
and 4, step 4: testing the neural network model by using the test data set, judging whether the model precision meets the preset precision, if so, executing the step 5, otherwise, returning to the step 3;
and 5: and carrying out automatic detection and identification on the ship body by using the trained ship body identification neural network model.
2. The hull target automatic detection and identification method according to claim 1, characterized in that the training data set is obtained by the following method:
the hull pictures are labeled using LabelImg, and then processed into txt format files using python, all of which constitute the training data set.
3. The hull target automatic detection and identification method according to claim 1, characterized in that the method for expanding the training data set specifically comprises:
and mixing different pictures in the training data set with image pixels by using a Mixup image enhancement mode to expand the data set, then carrying out scale transformation on the expanded training data set to expand the training data set again, and finally taking the training data set subjected to two times of expansion as a final expanded training data set.
4. The hull target automatic detection and identification method according to claim 1, characterized in that the training of the hull recognition neural network model in the step 3 adopts a mixed precision training mode.
5. The automatic detection and identification method for the ship hull target according to claim 4, characterized in that the mixed precision training mode is specifically as follows:
the 16-bit half-precision weight is used for calculating the activation function part, and the 32-bit single-precision weight is used for updating.
6. The automatic detection and identification method for the ship hull targets according to claim 1, characterized in that the ship hull recognition neural network model comprises a modified CSPDarknet53 framework (1), a second convolution module (2), an SPP pooling layer (3), a feature fusion module (4), an output layer (5) and an anchor frame detection module (6) which are connected in sequence; the improved CSPDarknet53 framework (1) is also connected with an output layer (5);
the improved CSPDarknet53 framework (1) comprises a first convolution module (101), a first residual error module (102), a second residual error module (103), a third residual error module (104), a fourth residual error module (105) and a fifth residual error module (106) which are connected in sequence;
the output layer (5) comprises a first splicing module (501), a third convolution module (502), a fourth convolution module (503), a second splicing module (504), a fifth convolution module (505), a sixth convolution module (506), a first upsampling module (507), a seventh convolution module (508) and a second upsampling module (509);
the input end of the first splicing module (501) is connected with the third residual error module (104), and the output end of the first splicing module (501) is sequentially connected with the third convolution module (502), the fourth convolution module (503) and the anchor frame detection module (6); the input end of the second splicing module (505) is connected with the fourth residual error module (105), and the output end of the second splicing module (505) is sequentially connected with the fifth convolution module (505), the sixth convolution module (506) and the anchor frame detection module (6); the input end of the seventh convolution module (508) is connected with the feature fusion module (4), and the output end of the seventh convolution module is connected with the anchor frame detection module (6);
the input end of the first up-sampling module (507) is connected with the output end of the fifth convolution module (505), and the output end of the first up-sampling module (507) is connected with the input end of the first splicing module (501); the input end of the second up-sampling module (509) is connected with the output end of the seventh convolution module (508), and the output end of the second up-sampling module (509) is connected with the input end of the second splicing module (504);
the feature fusion module (4) is three adaptive spatial feature fusion structures (ASFF) which are sequentially connected.
7. The automatic detection and identification method for the ship hull targets according to claim 6, characterized in that the first convolution module (101), the second convolution module (2), the third convolution module (502), the fourth convolution module (503), the fifth convolution module (505) and the sixth convolution module (506) are respectively composed of a plurality of convolution blocks; the convolution block comprises a convolution layer and a DropBlock regularization unit; the output end of the convolution layer is connected with the input end of the DropBlock;
the first residual error module (102), the second residual error module (103), the third residual error module (104), the fourth residual error module (105) and the fifth residual error module (106) are respectively composed of a plurality of residual error blocks; the residual blocks are represented as residual blocks n, and the expression of the residual block is as follows:
residual block n input + convolution block + residual unit n
The residual unit is specifically a residual network containing two convolution blocks.
8. The automatic detection and identification method for the ship hull targets according to claim 7, characterized in that the convolution layer adopts a convolution form of Mixconv2 d;
the activation function of the convolutional layer is specifically a Mish activation function, and specifically includes:
Mish=x*tanh(ln|1+e^x|)。
9. the automatic hull object detecting and identifying method according to claim 6, characterized in that the hull identifying neural network model adopts a cosine annealing mode to realize automatic attenuation of learning rate.
10. The hull target automatic detection and identification method according to claim 6, characterized in that the anchor frame detection module (6) comprises three anchor frame modules, which are obtained by training set clustering; three outputs obtained by the hull recognition neural network model are respectively input into the three anchor frame modules to obtain three outputs related to the coordinate information and the probability of the hull; and then, the three outputs are respectively subjected to non-maximum suppression to obtain the same proper output, if the probability of identifying the ship output by the anchor frame detection module (6) is greater than the set basic probability, the output is judged to be the ship detected, and the coordinate information output by the anchor frame detection module (6) is used as the position information of the ship in the image and is finally output.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010802211.2A CN111898699B (en) | 2020-08-11 | 2020-08-11 | Automatic detection and identification method for ship body target |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010802211.2A CN111898699B (en) | 2020-08-11 | 2020-08-11 | Automatic detection and identification method for ship body target |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111898699A true CN111898699A (en) | 2020-11-06 |
CN111898699B CN111898699B (en) | 2024-05-10 |
Family
ID=73228819
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010802211.2A Active CN111898699B (en) | 2020-08-11 | 2020-08-11 | Automatic detection and identification method for ship body target |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111898699B (en) |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112508001A (en) * | 2020-12-03 | 2021-03-16 | 安徽理工大学 | Coal gangue positioning method based on multispectral waveband screening and improved U-Net |
CN112598604A (en) * | 2021-03-04 | 2021-04-02 | 成都东方天呈智能科技有限公司 | Blind face restoration method and system |
CN112700446A (en) * | 2021-03-23 | 2021-04-23 | 常州微亿智造科技有限公司 | Algorithm model training method and device for industrial quality inspection |
CN112818913A (en) * | 2021-02-24 | 2021-05-18 | 西南石油大学 | Real-time smoking calling identification method |
CN113449634A (en) * | 2021-06-28 | 2021-09-28 | 上海翰声信息技术有限公司 | Video detection method and device for processing under strong light environment |
CN114170421A (en) * | 2022-02-10 | 2022-03-11 | 青岛海尔工业智能研究院有限公司 | Image detection method, device, equipment and storage medium |
WO2022142297A1 (en) * | 2021-01-04 | 2022-07-07 | Guangzhou Institute Of Advanced Technology, Chinese Academy Of Sciences | A robot grasping system and method based on few-shot learning |
WO2022147965A1 (en) * | 2021-01-09 | 2022-07-14 | 江苏拓邮信息智能技术研究院有限公司 | Arithmetic question marking system based on mixnet-yolov3 and convolutional recurrent neural network (crnn) |
CN118172787A (en) * | 2024-05-09 | 2024-06-11 | 南昌航空大学 | Lightweight document layout analysis method |
Citations (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170068840A1 (en) * | 2015-09-09 | 2017-03-09 | Accenture Global Services Limited | Predicting accuracy of object recognition in a stitched image |
CN108647648A (en) * | 2018-05-14 | 2018-10-12 | 电子科技大学 | A kind of Ship Recognition system and method under visible light conditions based on convolutional neural networks |
CN108776779A (en) * | 2018-05-25 | 2018-11-09 | 西安电子科技大学 | SAR Target Recognition of Sequential Images methods based on convolution loop network |
CN108875486A (en) * | 2017-09-28 | 2018-11-23 | 北京旷视科技有限公司 | Recongnition of objects method, apparatus, system and computer-readable medium |
CN109165585A (en) * | 2018-06-15 | 2019-01-08 | 沈阳理工大学 | A kind of improved ship target detection method based on YOLO V2 |
CN109271856A (en) * | 2018-08-03 | 2019-01-25 | 西安电子科技大学 | Remote sensing image object detection method based on expansion residual error convolution |
CN109685847A (en) * | 2018-12-26 | 2019-04-26 | 北京因时机器人科技有限公司 | A kind of training method and device of sensation target detection model |
CN110060508A (en) * | 2019-04-08 | 2019-07-26 | 武汉理工大学 | A kind of ship automatic testing method for inland river bridge zone |
CN110070142A (en) * | 2019-04-29 | 2019-07-30 | 上海大学 | A kind of marine vessel object detection method based on YOLO neural network |
US20190303715A1 (en) * | 2018-03-29 | 2019-10-03 | Qualcomm Incorporated | Combining convolution and deconvolution for object detection |
CN110321874A (en) * | 2019-07-12 | 2019-10-11 | 南京航空航天大学 | A kind of light-weighted convolutional neural networks pedestrian recognition method |
CN110348303A (en) * | 2019-06-06 | 2019-10-18 | 武汉理工大学 | A kind of auxiliary water surface patrol system being equipped on unmanned boat and water surface monitoring method |
CN110688955A (en) * | 2019-09-27 | 2020-01-14 | 西安建筑科技大学 | Building construction target detection method based on YOLO neural network |
US20200103498A1 (en) * | 2018-10-02 | 2020-04-02 | Metawave Corporation | Adaptive radar for near-far target identification |
WO2020071839A1 (en) * | 2018-10-04 | 2020-04-09 | 씨드로닉스㈜ | Ship and harbor monitoring device and method |
CN111046793A (en) * | 2019-12-11 | 2020-04-21 | 北京工业大学 | Tomato disease identification method based on deep convolutional neural network |
US20200160061A1 (en) * | 2017-12-11 | 2020-05-21 | Zhuhai Da Hengqin Technology Development Co., Ltd. | Automatic ship tracking method and system based on deep learning network and mean shift |
-
2020
- 2020-08-11 CN CN202010802211.2A patent/CN111898699B/en active Active
Patent Citations (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170068840A1 (en) * | 2015-09-09 | 2017-03-09 | Accenture Global Services Limited | Predicting accuracy of object recognition in a stitched image |
CN108875486A (en) * | 2017-09-28 | 2018-11-23 | 北京旷视科技有限公司 | Recongnition of objects method, apparatus, system and computer-readable medium |
US20200160061A1 (en) * | 2017-12-11 | 2020-05-21 | Zhuhai Da Hengqin Technology Development Co., Ltd. | Automatic ship tracking method and system based on deep learning network and mean shift |
US20190303715A1 (en) * | 2018-03-29 | 2019-10-03 | Qualcomm Incorporated | Combining convolution and deconvolution for object detection |
CN108647648A (en) * | 2018-05-14 | 2018-10-12 | 电子科技大学 | A kind of Ship Recognition system and method under visible light conditions based on convolutional neural networks |
CN108776779A (en) * | 2018-05-25 | 2018-11-09 | 西安电子科技大学 | SAR Target Recognition of Sequential Images methods based on convolution loop network |
CN109165585A (en) * | 2018-06-15 | 2019-01-08 | 沈阳理工大学 | A kind of improved ship target detection method based on YOLO V2 |
CN109271856A (en) * | 2018-08-03 | 2019-01-25 | 西安电子科技大学 | Remote sensing image object detection method based on expansion residual error convolution |
US20200103498A1 (en) * | 2018-10-02 | 2020-04-02 | Metawave Corporation | Adaptive radar for near-far target identification |
WO2020071839A1 (en) * | 2018-10-04 | 2020-04-09 | 씨드로닉스㈜ | Ship and harbor monitoring device and method |
CN109685847A (en) * | 2018-12-26 | 2019-04-26 | 北京因时机器人科技有限公司 | A kind of training method and device of sensation target detection model |
CN110060508A (en) * | 2019-04-08 | 2019-07-26 | 武汉理工大学 | A kind of ship automatic testing method for inland river bridge zone |
CN110070142A (en) * | 2019-04-29 | 2019-07-30 | 上海大学 | A kind of marine vessel object detection method based on YOLO neural network |
CN110348303A (en) * | 2019-06-06 | 2019-10-18 | 武汉理工大学 | A kind of auxiliary water surface patrol system being equipped on unmanned boat and water surface monitoring method |
CN110321874A (en) * | 2019-07-12 | 2019-10-11 | 南京航空航天大学 | A kind of light-weighted convolutional neural networks pedestrian recognition method |
CN110688955A (en) * | 2019-09-27 | 2020-01-14 | 西安建筑科技大学 | Building construction target detection method based on YOLO neural network |
CN111046793A (en) * | 2019-12-11 | 2020-04-21 | 北京工业大学 | Tomato disease identification method based on deep convolutional neural network |
Non-Patent Citations (2)
Title |
---|
王新立;江福才;宁方鑫;马全党;张帆;邹红兵;: "基于改进卷积神经网络的船舶目标检测", 中国航海, no. 02 * |
王新立;江福才;宁方鑫;马全党;张帆;邹红兵;: "基于改进卷积神经网络的船舶目标检测", 中国航海, no. 02, 25 June 2018 (2018-06-25) * |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112508001A (en) * | 2020-12-03 | 2021-03-16 | 安徽理工大学 | Coal gangue positioning method based on multispectral waveband screening and improved U-Net |
WO2022142297A1 (en) * | 2021-01-04 | 2022-07-07 | Guangzhou Institute Of Advanced Technology, Chinese Academy Of Sciences | A robot grasping system and method based on few-shot learning |
WO2022147965A1 (en) * | 2021-01-09 | 2022-07-14 | 江苏拓邮信息智能技术研究院有限公司 | Arithmetic question marking system based on mixnet-yolov3 and convolutional recurrent neural network (crnn) |
CN112818913A (en) * | 2021-02-24 | 2021-05-18 | 西南石油大学 | Real-time smoking calling identification method |
CN112598604A (en) * | 2021-03-04 | 2021-04-02 | 成都东方天呈智能科技有限公司 | Blind face restoration method and system |
CN112700446A (en) * | 2021-03-23 | 2021-04-23 | 常州微亿智造科技有限公司 | Algorithm model training method and device for industrial quality inspection |
CN113449634A (en) * | 2021-06-28 | 2021-09-28 | 上海翰声信息技术有限公司 | Video detection method and device for processing under strong light environment |
CN114170421A (en) * | 2022-02-10 | 2022-03-11 | 青岛海尔工业智能研究院有限公司 | Image detection method, device, equipment and storage medium |
CN114170421B (en) * | 2022-02-10 | 2022-06-17 | 卡奥斯工业智能研究院(青岛)有限公司 | Image detection method, device, equipment and storage medium |
CN118172787A (en) * | 2024-05-09 | 2024-06-11 | 南昌航空大学 | Lightweight document layout analysis method |
Also Published As
Publication number | Publication date |
---|---|
CN111898699B (en) | 2024-05-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111898699A (en) | Automatic detection and identification method for hull target | |
CN110298361B (en) | Semantic segmentation method and system for RGB-D image | |
CN110348376B (en) | Pedestrian real-time detection method based on neural network | |
CN111126359B (en) | High-definition image small target detection method based on self-encoder and YOLO algorithm | |
CN112766087A (en) | Optical remote sensing image ship detection method based on knowledge distillation | |
CN111563508A (en) | Semantic segmentation method based on spatial information fusion | |
CN111079739B (en) | Multi-scale attention feature detection method | |
JP2020123330A (en) | Method for acquiring sample image for label acceptance inspection from among auto-labeled images utilized for neural network learning, and sample image acquisition device utilizing the same | |
CN111489339A (en) | Method for detecting defects of bolt spare nuts of high-speed railway positioner | |
CN111079604A (en) | Method for quickly detecting tiny target facing large-scale remote sensing image | |
CN114495029A (en) | Traffic target detection method and system based on improved YOLOv4 | |
CN113052057A (en) | Traffic sign identification method based on improved convolutional neural network | |
CN110533166A (en) | It is a kind of based on when space fusion feature indoor orientation method | |
CN115620180A (en) | Aerial image target detection method based on improved YOLOv5 | |
CN115346068A (en) | Automatic generation method for bolt loss fault image of railway freight train | |
CN113298817A (en) | High-accuracy semantic segmentation method for remote sensing image | |
CN115797808A (en) | Unmanned aerial vehicle inspection defect image identification method, system, device and medium | |
CN115240259A (en) | Face detection method and face detection system based on YOLO deep network in classroom environment | |
CN113052071B (en) | Method and system for rapidly detecting distraction behavior of driver of hazardous chemical substance transport vehicle | |
CN113963333A (en) | Traffic sign board detection method based on improved YOLOF model | |
CN116778346B (en) | Pipeline identification method and system based on improved self-attention mechanism | |
CN114494893B (en) | Remote sensing image feature extraction method based on semantic reuse context feature pyramid | |
CN115861595A (en) | Multi-scale domain self-adaptive heterogeneous image matching method based on deep learning | |
CN115661657A (en) | Lightweight unmanned ship target detection method | |
CN113569650A (en) | Unmanned aerial vehicle autonomous inspection positioning method based on electric power tower label identification |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |