CN117037132A - Ship water gauge reading detection and identification method based on machine vision - Google Patents
Ship water gauge reading detection and identification method based on machine vision Download PDFInfo
- Publication number
- CN117037132A CN117037132A CN202311003627.8A CN202311003627A CN117037132A CN 117037132 A CN117037132 A CN 117037132A CN 202311003627 A CN202311003627 A CN 202311003627A CN 117037132 A CN117037132 A CN 117037132A
- Authority
- CN
- China
- Prior art keywords
- ship
- water gauge
- water
- area
- predicted
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 title claims abstract description 103
- 238000001514 detection method Methods 0.000 title claims abstract description 45
- 238000000034 method Methods 0.000 title claims abstract description 28
- 238000012549 training Methods 0.000 claims abstract description 12
- 230000011218 segmentation Effects 0.000 claims abstract description 10
- 238000007781 pre-processing Methods 0.000 claims description 4
- 230000004913 activation Effects 0.000 claims description 3
- 238000010586 diagram Methods 0.000 claims description 3
- 230000006870 function Effects 0.000 claims description 3
- 238000012544 monitoring process Methods 0.000 claims description 3
- 238000010606 normalization Methods 0.000 claims description 3
- 238000000605 extraction Methods 0.000 claims description 2
- 238000012805 post-processing Methods 0.000 abstract description 11
- 238000005452 bending Methods 0.000 abstract description 3
- 230000008569 process Effects 0.000 abstract description 3
- 238000012545 processing Methods 0.000 abstract description 2
- 238000012015 optical character recognition Methods 0.000 description 7
- 238000004364 calculation method Methods 0.000 description 5
- 230000008859 change Effects 0.000 description 2
- 238000013135 deep learning Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000003064 k means clustering Methods 0.000 description 2
- 238000005259 measurement Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- PXFBZOLANLWPMH-UHFFFAOYSA-N 16-Epiaffinine Natural products C1C(C2=CC=CC=C2N2)=C2C(=O)CC2C(=CC)CN(C)C1C2CO PXFBZOLANLWPMH-UHFFFAOYSA-N 0.000 description 1
- 238000012935 Averaging Methods 0.000 description 1
- 102100032202 Cornulin Human genes 0.000 description 1
- 101000920981 Homo sapiens Cornulin Proteins 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000013527 convolutional neural network Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 238000009432 framing Methods 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 238000012804 iterative process Methods 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 238000003062 neural network model Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
- 238000005303 weighing Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/62—Text, e.g. of license plates, overlay texts or captions on TV images
- G06V20/63—Scene text, e.g. street names
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/52—Scale-space analysis, e.g. wavelet analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/762—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using clustering, e.g. of similar faces in social networks
- G06V10/763—Non-hierarchical techniques, e.g. based on statistics of modelling distributions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
- G06V10/765—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects using rules for classification or partitioning the feature space
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
- G06V30/14—Image acquisition
- G06V30/148—Segmentation of character regions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
- G06V30/19—Recognition using electronic means
- G06V30/191—Design or setup of recognition systems or techniques; Extraction of features in feature space; Clustering techniques; Blind source separation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
- G06V30/19—Recognition using electronic means
- G06V30/191—Design or setup of recognition systems or techniques; Extraction of features in feature space; Clustering techniques; Blind source separation
- G06V30/19173—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/07—Target detection
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02A—TECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
- Y02A90/00—Technologies having an indirect contribution to adaptation to climate change
- Y02A90/30—Assessment of water resources
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Software Systems (AREA)
- Evolutionary Computation (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Computing Systems (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Databases & Information Systems (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Probability & Statistics with Applications (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a ship water gauge reading detection and identification method based on machine vision, and relates to the technical field of image processing; the method comprises the steps of S1, collecting pictures of the past ship water gauge area, S2, establishing a ship water limit line, S3, predicting and marking the ship water gauge scale, S4, obtaining a precise area to be predicted, and S5, carrying out character prediction on the area to be predicted; according to the method, firstly, a strategy of marking integer parts and decimal parts on the water gauge readings of the ship can avoid complex post-processing algorithm logic, higher accuracy and robustness are still achieved in the face of the condition of inclination of shooting angles or bending of the ship body, in the practical application process, along with continuous expansion of training data, the detection accuracy of the YOLOv5 target detection algorithm on the integer parts and the decimal parts of the water gauge is continuously improved, and secondly, the algorithm is introduced into a Mask RCNN example segmentation algorithm to divide the ship and the water area, and accuracy of a region to be predicted can be guaranteed no matter whether the water body of an actual scene is turbid or not.
Description
Technical Field
The invention relates to the technical field of image processing, in particular to a ship water gauge reading detection and identification method based on machine vision.
Background
The ship water gauge weighing is a method for measuring ship-borne cargoes, and the method is widely applied to the weight measurement of ship transportation bulk cheap bulk cargoes. Whether the accuracy of the ship water gauge reading identification directly influences the accuracy of ship cargo capacity calculation, and further influences settlement of commodity trade, clearance tax counting and the like, so that the accuracy of ship water gauge line detection and water gauge reading identification is guaranteed to have important significance. Currently, the measurement of ship water gauge lines mainly comprises the following methods:
1. manual observation method: after the ship approaches the shore, manually observing, recording and manually checking by a water gauge observer with abundant experience, comparing and averaging the observed results to obtain a final result, wherein the method is greatly influenced by main objective factors, for example, the water gauge reading is difficult to observe when the water surface fluctuates greatly, the observer cannot obtain accurate numerical values due to the limitation of weather and an observation visual angle, and secondly, the efficiency of manually observing the water gauge reading is low and cannot meet the high-efficiency passing of a port;
2. the ship water level line detection method based on the deep learning algorithm comprises the following steps: with the rapid development of a deep learning algorithm, more and more models based on convolutional neural networks are used for detecting and identifying water gauge readings of ships, the method is used for correcting original input images through affine transformation, secondly, an example segmentation algorithm is used for extracting water gauge areas and ship water boundary positions of the ships, water gauge area pictures are sent into an OCR character recognition algorithm to read water gauge values, finally, integer values and decimal values of the water gauges are judged through complex post-processing logic and combined with ship water boundary position estimation to obtain final water gauge readings, the method is higher in accuracy requirements on image preprocessing and example segmentation algorithm, water gauge reading distortion caused by inclination of shooting angles, surface bending of ships and the like is usually encountered in practical application, in this case, the accuracy of the method is not high, and secondly, complex post-processing logic is only required to meet partial specific scenes, the post-processing algorithm is required to be redesigned when ship types or monitoring scenes change, therefore, the method for judging the integer values and the decimal values through the post-processing algorithm is not high in performance, and cannot be deployed on a large scale, and the inventor proposes a water gauge detection method based on machine vision and ship identification method for solving the problems.
Disclosure of Invention
The method aims to solve the problem of low accuracy of ship water level line detection; the invention aims to provide a ship water gauge reading detection and identification method based on machine vision.
In order to solve the technical problems, the invention adopts the following technical scheme: a ship water gauge reading detection and identification method based on machine vision comprises the following steps:
s1, acquiring a picture of a water gauge area of a past ship through an unmanned plane or a wharf fixed monitoring camera, wherein the image acquisition is required to ensure that light is sufficient, and the water gauge graduation and the water content boundary line of the ship are clearly visible;
s2, performing example segmentation on a ship area and a water area by using a Mask-RCNN model to obtain a ship water boundary;
s3, taking M as a unit, using Arabic numerals to represent integer and decimal parts, respectively predicting the integer part and the decimal part of the scale of the ship water gauge by using a YOLOv5 target detection algorithm to obtain the coordinate position of a boundary frame of the water gauge reading and the category of each boundary frame, wherein the integer part is marked as "int" category, and the decimal part is marked as "float" category;
s4, respectively finding out a target boundary box with the largest y-axis coordinate of a center point in an 'int' type and a 'float' type target by taking the upper left corner of an original image as a coordinate origin, namely an effective reading area, wherein the area further obtains a more accurate area to be predicted by combining the ship water boundary line obtained in the S2;
s5, character prediction is carried out on the area to be predicted through an OCR algorithm, accuracy of water gauge reading prediction is guaranteed, interference of water surface fluctuation on prediction is avoided, all water gauge readings detected within 10 seconds are taken in actual pushing, and an average value is taken as a final prediction result.
Preferably, in S3, under such labeling strategy, the target bounding boxes of different classes have different sizes and aspect ratios, so that the target can be predicted in different prediction feature graphs by the subsequent YOLOv5 target detection algorithm.
Preferably, in S3, YOLOv5 belongs to an object detection algorithm of Anchor-Based, that is, regression prediction needs to be performed by using an a priori Anchor Box (Anchor Box), the basic principle is that an input image is scaled to a size required by a model through a preprocessing operation, and on the basis, a series of convolution, batch normalization and activation functions are used for feature extraction and downsampling of the image to 8 times, 16 times and 32 times, and due to different downsampling multiplying power, a large object in a training set is usually predicted in a small-size prediction feature map, and a small object is predicted in a large-size prediction feature map.
Compared with the prior art, the invention has the beneficial effects that:
1. in the invention, the integer and decimal part of the ship water gauge reading are directly distinguished and detected by utilizing the target detection algorithm, so that complex numerical judgment logic in post-processing is avoided, and the accuracy and robustness of an algorithm model are improved;
2. according to the invention, the ship water gauge image is segmented, detected and OCR character recognition is carried out by fusing the multiple models, the result of ship water gauge reading depends on the prediction output of the multiple models, and the detection result has higher reliability. Meanwhile, the invention simplifies the detection flow of the water gauge reading, avoids redundant operation and improves the detection efficiency of the algorithm model;
3. according to the invention, the ship water gauge can adapt to different types of ship water gauges in a mode of loading a pre-training model and fine tuning, can be easily adapted to other similar water gauge reading tasks, and has strong generalization capability;
4. according to the method, firstly, a strategy of marking integer parts and decimal parts on the water gauge readings of the ship can avoid complex post-processing algorithm logic, higher accuracy and robustness are still achieved in the face of the condition of inclination of shooting angles or bending of the ship body, in the practical application process, along with continuous expansion of training data, the detection accuracy of the YOLOv5 target detection algorithm on the integer parts and the decimal parts of the water gauge is continuously improved, and secondly, the algorithm is introduced into a Mask RCNN example segmentation algorithm to divide the ship and the water area, and accuracy of a region to be predicted can be guaranteed no matter whether the water body of an actual scene is turbid or not.
Drawings
In order to more clearly illustrate the embodiments of the invention or the technical solutions in the prior art, the drawings that are required in the embodiments or the description of the prior art will be briefly described, it being obvious that the drawings in the following description are only some embodiments of the invention, and that other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is a schematic diagram of the flow of detection and identification of the readings of the water gauge of the ship according to the present invention.
FIG. 2 is a diagram showing an example of the water gauge reading label according to the present invention.
Fig. 3 is a comparison of an original picture and a binarized image according to the present invention.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
Examples: 1-3, the invention provides a ship water gauge reading detection and identification method based on machine vision, which comprises the following steps:
s1, acquiring a ship water gauge area video by using an unmanned plane or a fixed camera, and framing the video to obtain original image data, wherein in order to ensure the training effect of a subsequent deep neural network model, the acquired data need to cover scenes of different illumination, weather and ship types, and the water gauge scale of the ship in the picture and the water content boundary line of the ship are clearly visible;
s2, marking water gauge readings of original image data by using a LabelImg open source image marking tool, marking integer parts of the water gauge readings as 'int' types and decimal parts as 'float' types for training a subsequent YOLOv5 target detection algorithm, particularly as shown in FIG. 2, marking ship areas and water areas by using a Labelme image marking tool, wherein the marking types are 'polygon', an original picture and a binary image, as shown in FIG. 3;
s3, pre-training a CRNN model on an open source optical character recognition data set ICDAR2013 by an optical character recognition algorithm, and performing fine tuning on the model by using a marked ship water gauge reading data set;
s4, based on the water gauge reading target detection data set in S2, three scales are generated by clustering the water gauge reading target bounding box by using a K-Means algorithm, wherein each group of scales has 3 aspect ratios, and the aspect ratios are respectively 1: 1. 1:2 and 2:1, training a YOLOv5 target detection model by 9 anchors, setting the resolution of model input to 640 x 640, ensuring that part of small target water gauge reading bounding boxes still have characteristic information after downsampling, and training a Mask RCNN model by utilizing a segmentation data set of a water area and a ship in S2;
s5, in the actual reasoning process, a Mask RCNN segmentation model is used for predicting the position of a ship water boundary line in a picture, YOLOv5 is used for detecting an integer region and a decimal region of a water gauge, and an effective reading region of the ship water gauge is obtained through filtering logic;
s6, sending the effective reading area into an OCR character recognition model to identify specific readings, setting the detection time to 10S in a post-processing stage, extracting 80% of video frames in 10S, sending the video frames into the algorithm model, and carrying out weighted summation on the detection results of the extracted video frames in 10S to obtain a final water gauge reading result, wherein in order to avoid the accidental single reasoning and the interference caused by water surface fluctuation, the post-processing mode is not only in line with the scientific mode of conventional manual observation reading, but also the expenditure of calculation force is saved.
In S3, under the marking strategy, different types of target bounding boxes have different sizes and aspect ratios, so that the target can be predicted in different prediction feature graphs by a subsequent YOLOv5 target detection algorithm.
In S3, YOLOv5 belongs to an object detection algorithm of Anchor-Based, that is, regression prediction needs to be performed by using an Anchor Box (Anchor Box), the basic principle is that an input image is scaled to a size required by a model through preprocessing operation, and on the basis, a series of convolution, batch normalization and activation functions are used for extracting features of the image and downsampling the image to 8 times, 16 times and 32 times, and due to different downsampling multiplying power, large objects in a training set are usually predicted in a small-size prediction feature map, and small objects are predicted in a large-size prediction feature map.
Working principle: in the invention, the detection performance of the Anchor-Based algorithm is very sensitive to the size, the aspect ratio and the number of Anchor frames, in order to improve the detection rate of the ship water gauge reading, the K-Means algorithm is adopted to cluster the Anchor frames, and in the YOLOv5 target detection algorithm, the target boundary frame is usually represented by the coordinates of an upper left corner point and a lower right corner point and is marked as (x) 1 ,y 1 ,x 2 ,y 2 ) When clustering the target bounding boxes, only the width and the height of the bounding boxes are required to be used as features, however, the sizes of actually sampled pictures are different, so that the width and the height of the target bounding boxes are required to be normalized by using the width and the height of the pictures, and the calculation formula is as follows:
it is found in practice that if the euclidean distance in the conventional K-Means clustering algorithm is used as a measure, clusters of large target bounding boxes will have larger errors than clusters of small target bounding boxes in the clustering iteration, so the distance from each target bounding box to each cluster center is calculated by adopting (Intersection over Union, overlap ratio) in the invention, and if the calculation formula of the overlap ratio IoU and the calculation formula of the distance d are as follows:
d(box,anchor)=1-IoU(box,anchor)
in the above formula, the intersection (box, anchor) represents the intersection of the target bounding box and the anchor frame, and the unit (box, anchor) represents the union of the target bounding box and the anchor frame, and it is known from the above formula that if the target bounding box and the anchor frame completely overlap, i.e., ioU =1, the distance between the anchor frame and the target bounding box is 0. In the iterative process, repeatedly calculating the distance d, distributing each target boundary frame to the anchor frame closest to the target boundary frame to form clusters, calculating the average value of the width and the height of all the target boundary frames in each cluster, updating the width and the height of the anchor frame until no change occurs, enabling the anchor frame after K-Means clustering to be more in line with the boundary frame aspect ratio of a ship water gauge reading data set, and enabling integer areas and decimal areas to be predicted better in a YOLOv5 training stage;
in practical application, errors often exist in an effective reading area obtained only through a YOLOv5 target detection model under the influence of water color, and particularly when water scale reflection occurs on the water surface, the target detection model is prone to false detection, a Mask RCNN segmentation network is introduced to conduct example segmentation on a ship area and a water area, ship water boundary lines can be obtained through classifying each pixel point of an image, if false detection occurs at the boundary lines through a YOLOv5 target detection algorithm, y-axis coordinates of right lower corner points of a target boundary frame are located below the ship water boundary lines, and at the moment, right lower corner coordinates of the target boundary frame exceeding the ship water boundary lines are restrained on the boundary lines through post-processing logic. Likewise, if the upper left corner of the bounding box of the target deduced by YOLOv5 is below the ship water boundary, then the bounding box is considered as an invalid target, and such targets are filtered directly in post-processing logic, ensuring that the target fed into the subsequent OCR character recognition algorithm is just a water gauge reading observable at the ship water boundary. Through the combined action of multiple models, the false detection rate of the water gauge readings of the ship is greatly reduced, and the method is remarkable in that different models can be easily applied to other similar water gauge reading tasks by cascading, and a good detection effect can be obtained.
It will be apparent to those skilled in the art that various modifications and variations can be made to the present invention without departing from the spirit or scope of the invention. Thus, it is intended that the present invention also include such modifications and alterations insofar as they come within the scope of the appended claims or the equivalents thereof.
Claims (3)
1. The ship water gauge reading detection and identification method based on machine vision is characterized by comprising the following steps of:
s1, acquiring a picture of a water gauge area of a past ship through an unmanned plane or a wharf fixed monitoring camera, wherein the image acquisition is required to ensure that light is sufficient, and the water gauge graduation and the water content boundary line of the ship are clearly visible;
s2, performing example segmentation on a ship area and a water area by using a Mask-RCNN model to obtain a ship water boundary;
s3, taking M as a unit, using Arabic numerals to represent integer and decimal parts, respectively predicting the integer part and the decimal part of the scale of the ship water gauge by using a YOLOv5 target detection algorithm to obtain the coordinate position of a boundary frame of the water gauge reading and the category of each boundary frame, wherein the integer part is marked as "int" category, and the decimal part is marked as "float" category;
s4, respectively finding out a target boundary box with the largest y-axis coordinate of a center point in an 'int' type and a 'float' type target by taking the upper left corner of an original image as a coordinate origin, namely an effective reading area, wherein the area further obtains a more accurate area to be predicted by combining the ship water boundary line obtained in the S2;
s5, character prediction is carried out on the area to be predicted through an OCR algorithm, accuracy of water gauge reading prediction is guaranteed, interference of water surface fluctuation on prediction is avoided, all water gauge readings detected within 10 seconds are taken in actual pushing, and an average value is taken as a final prediction result.
2. The method for detecting and identifying the ship water gauge reading based on the machine vision according to claim 1, wherein in the step S3, under the marking strategy, different types of target bounding boxes have different sizes and aspect ratios, so that the target can be predicted in different prediction feature diagrams by a subsequent YOLOv5 target detection algorithm.
3. The ship water gauge reading detection and identification method Based on machine vision as claimed in claim 1, wherein in S3, YOLOv5 belongs to an object detection algorithm of Anchor-Based, namely, regression prediction is needed by using an priori Anchor Box (Anchor Box), the basic principle is that an input image is scaled to a size required by a model through preprocessing operation, and on the basis, feature extraction is performed on the image by using a series of convolution, batch normalization and activation functions and downsampling is performed to 8 times, 16 times and 32 times, and large objects in a training set are usually predicted in a small-size prediction feature map and small objects are predicted in a large-size prediction feature map due to different downsampling multiplying power.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311003627.8A CN117037132A (en) | 2023-08-10 | 2023-08-10 | Ship water gauge reading detection and identification method based on machine vision |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311003627.8A CN117037132A (en) | 2023-08-10 | 2023-08-10 | Ship water gauge reading detection and identification method based on machine vision |
Publications (1)
Publication Number | Publication Date |
---|---|
CN117037132A true CN117037132A (en) | 2023-11-10 |
Family
ID=88631228
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202311003627.8A Withdrawn CN117037132A (en) | 2023-08-10 | 2023-08-10 | Ship water gauge reading detection and identification method based on machine vision |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117037132A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117437495A (en) * | 2023-12-20 | 2024-01-23 | 武汉理工大学 | Water gauge line identification method and device for ship, electronic equipment and storage device |
CN117854045A (en) * | 2024-03-04 | 2024-04-09 | 东北大学 | Automatic driving-oriented vehicle target detection method |
-
2023
- 2023-08-10 CN CN202311003627.8A patent/CN117037132A/en not_active Withdrawn
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117437495A (en) * | 2023-12-20 | 2024-01-23 | 武汉理工大学 | Water gauge line identification method and device for ship, electronic equipment and storage device |
CN117854045A (en) * | 2024-03-04 | 2024-04-09 | 东北大学 | Automatic driving-oriented vehicle target detection method |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021238030A1 (en) | Water level monitoring method for performing scale recognition on the basis of partitioning by clustering | |
CN109993166B (en) | Automatic reading identification method for pointer instrument based on scale searching | |
US20210374466A1 (en) | Water level monitoring method based on cluster partition and scale recognition | |
CN117037132A (en) | Ship water gauge reading detection and identification method based on machine vision | |
CN111160120A (en) | Fast R-CNN article detection method based on transfer learning | |
CN111598098B (en) | Water gauge water line detection and effectiveness identification method based on full convolution neural network | |
CN110781885A (en) | Text detection method, device, medium and electronic equipment based on image processing | |
CN102975826A (en) | Portable ship water gauge automatic detection and identification method based on machine vision | |
CN111652213A (en) | Ship water gauge reading identification method based on deep learning | |
CN107220962B (en) | Image detection method and device for tunnel cracks | |
CN110619328A (en) | Intelligent ship water gauge reading identification method based on image processing and deep learning | |
CN113077392B (en) | High-accuracy automatic reading method for fuzzy photo of pointer table | |
CN112734729B (en) | Water gauge water level line image detection method and device suitable for night light supplement condition and storage medium | |
CN115147418B (en) | Compression training method and device for defect detection model | |
CN112132131A (en) | Measuring cylinder liquid level identification method and device | |
CN114241469A (en) | Information identification method and device for electricity meter rotation process | |
CN115775236A (en) | Surface tiny defect visual detection method and system based on multi-scale feature fusion | |
CN116228780A (en) | Silicon wafer defect detection method and system based on computer vision | |
CN116188756A (en) | Instrument angle correction and indication recognition method based on deep learning | |
CN117557565B (en) | Detection method and device for lithium battery pole piece | |
CN113255555A (en) | Method, system, processing equipment and storage medium for identifying Chinese traffic sign board | |
CN117315670A (en) | Water meter reading area detection method based on computer vision | |
CN116935369A (en) | Ship water gauge reading method and system based on computer vision | |
CN114742849B (en) | Leveling instrument distance measuring method based on image enhancement | |
CN116486212A (en) | Water gauge identification method, system and storage medium based on computer vision |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
WW01 | Invention patent application withdrawn after publication | ||
WW01 | Invention patent application withdrawn after publication |
Application publication date: 20231110 |