CN112949389A - Haze image target detection method based on improved target detection network - Google Patents

Haze image target detection method based on improved target detection network Download PDF

Info

Publication number
CN112949389A
CN112949389A CN202110115251.4A CN202110115251A CN112949389A CN 112949389 A CN112949389 A CN 112949389A CN 202110115251 A CN202110115251 A CN 202110115251A CN 112949389 A CN112949389 A CN 112949389A
Authority
CN
China
Prior art keywords
image
haze
network
target
target detection
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110115251.4A
Other languages
Chinese (zh)
Inventor
许悦雷
张凡
加尔肯别克
胡璐娟
崔祺
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Northwestern Polytechnical University
Original Assignee
Northwestern Polytechnical University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Northwestern Polytechnical University filed Critical Northwestern Polytechnical University
Priority to CN202110115251.4A priority Critical patent/CN112949389A/en
Publication of CN112949389A publication Critical patent/CN112949389A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • G06V20/13Satellite images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/30Noise filtering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Remote Sensing (AREA)
  • Astronomy & Astrophysics (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a haze image target detection method based on an improved target detection network, which comprises the steps of firstly adopting a rapid defogging algorithm based on optimized contrast enhancement, effectively removing haze in real time and enhancing the contrast information of an image; then sending the defogged image into an improved Yolov4 target detection network for detection, and outputting target information under the haze weather condition; meanwhile, in order to improve the feature extraction capability of the YOLOV4 network, the invention replaces the residual blocks in the YOLOV4 main network with the dense link blocks, and compared with the residual blocks, the dense link blocks enable the network to improve the feature expression capability and improve the feature extraction capability of the network. Finally, experiments prove that the target detection capability under the haze weather condition can be improved by using the method.

Description

Haze image target detection method based on improved target detection network
Technical Field
The invention belongs to the technical field of image processing, and particularly relates to a haze image target detection method.
Background
The target detection and identification refers to positioning and identifying an interested target in an image by using a certain technical means. Currently, there are two main types of target detection methods, sliding window based methods and regional targeting based methods. The former uses a window to traverse the whole image, extracts the image characteristics in the window at each position, and finally uses a classifier to classify the characteristics; the latter adopts a saliency detection or target method to extract a candidate region of a target, so that a large amount of calculation of a sliding window is avoided, but as the situation that a part of the target is framed or the adjacent target is distinguished is likely to occur, a certain influence is exerted on a detection result, and a plurality of methods are proposed to improve the problem, for example, a document proposes a region extraction method with stronger target, so that the number of candidate windows is reduced and the quality of the candidate windows is improved.
Although the current deep learning network is greatly developed in the field of target detection, target identification is basically performed on the basis of clear images, and identification accuracy is greatly reduced when the target is under complex conditions such as haze and the like.
In reality, rain, snow, fog, dust and the like are often encountered, and in severe weather, the performance of sensors for city security, high-speed monitoring and reconnaissance monitoring is rapidly reduced or even fails, so that the degradation problems of loss of details, increase of noise, reduction of contrast and the like of images acquired by a computer system are caused, and great influence is caused on post-processing such as image feature extraction, detection segmentation, target identification and the like. Fog, for example, is the most common extreme weather that affects visual performance, and image visual distortion from such severe weather conditions can affect the performance of many outdoor vision systems, negatively impacting target recognition accuracy.
The performance of a monitoring and reconnaissance sensor can be greatly influenced by complicated and changeable weather conditions such as haze and the like, the acquired image noise is increased, details are lost, the contrast is reduced and the like, the target identification and detection are used as a preprocessing part of most computer vision applications, and the identification effect of the target identification and detection inevitably can be interfered by haze conditions of different degrees. Therefore, the method can quickly and accurately remove the haze interference in the input image, and can stably and normally work in various complex states, thereby having extremely high value for improving the accuracy of target identification and detection. Some traditional defogging algorithms have the problems that a defogged picture can generate a halo artifact, the color distortion is serious, haze can be locally eliminated, but thick fog in some places cannot be eliminated, the instantaneity is difficult to guarantee, and the like.
Disclosure of Invention
In order to overcome the defects of the prior art, the invention provides a haze image target detection method based on an improved target detection network, which comprises the steps of firstly adopting a rapid defogging algorithm based on optimized contrast enhancement, effectively removing haze in real time and enhancing the contrast information of an image; then sending the defogged image into an improved Yolov4 target detection network for detection, and outputting target information under the haze weather condition; meanwhile, in order to improve the feature extraction capability of the YOLOV4 network, the invention replaces the residual blocks in the YOLOV4 main network with the dense link blocks, and compared with the residual blocks, the dense link blocks enable the network to improve the feature expression capability and improve the feature extraction capability of the network. Finally, experiments prove that the target detection capability under the haze weather condition can be improved by using the method.
The technical scheme adopted by the invention for solving the technical problem comprises the following steps:
step 1: acquiring a haze weather image containing a target;
step 2: defining a haze degradation model as shown in formula (1):
I(p)=t(p)J(p)+(1-t(p))A (1)
wherein J (p) ═ J (J)r(p),Jg(p),Jb(p))TRepresenting the image after haze removal, Jr(p),Jg(p),Jb(p) pixel values of the image after haze removal in three channels of r, g, and b, respectively, and I (p) ═ Ir(p),Ig(p),Ib(p))TRepresenting an image containing haze Ir(p),Ig(p),Ib(p) each isRepresenting pixel values of an image containing haze in three channels of r, g and b, wherein p represents the image; a represents the atmospheric light value of the image, t (p) epsilon [0, 1]Represents a transmittance;
and step 3: determining an atmospheric light value of the haze weather image by using a hierarchical search method based on quadtree subdivision;
dividing a haze weather image into four rectangular areas, subtracting the standard deviation of pixel values in each rectangular area from the average pixel value of each rectangular area, then selecting the rectangular area with the largest calculation result, further dividing the rectangular area into four smaller rectangular areas, and repeating the process until the size of the selected rectangular area is smaller than a pre-specified threshold;
in the finally selected rectangular area, calculating the distance between each pixel point and the white pixel, namely calculating I (I)r(p),Ig(p),Ib(p)) - (255, 255, 255) l, finding a pixel point with the minimum distance to the white pixel, and taking the pixel value of the pixel point as the atmospheric light value of the haze weather image;
and 4, step 4: calculating an optimal transmittance based on the image block;
step 4-1: dividing the haze weather image into a plurality of image blocks b × b, and assuming that the scene depth of each image block is consistent; setting a transmittance for each image block, the haze degradation model for each image block is written as:
Figure BDA0002920375610000021
step 4-2: the mean square error contrast is adopted to represent the contrast of the haze-removed image block of the haze weather image block, and the mean square error contrast is represented by the formula (3):
Figure BDA0002920375610000031
where c e (r, g, b) represents the color channel index, Jc(p) represents the image block pixel values after haze removal,
Figure BDA0002920375610000032
is Jc(p), N is the number of pixels of the image block; according to formula (2), CMSEThe rewrite is:
Figure BDA0002920375610000033
wherein, Ic(p) represents a haze weather image block pixel value,
Figure BDA0002920375610000034
is Ic(p) average value;
step 4-3: obtaining the optimum transmittance t*
Defining a contrast cost Econtrast
Figure BDA0002920375610000035
Defining information loss cost Eloss
Eloss=∑c∈(r,g,b)p∈B{(min{0,Jc(p)})2+(max{0,Jc(p)-255})2} (6)
Where min {0, J }c(p) } and max {0, Jc(p) -255} represent pixel values in the output defogged image that are truncated due to pixel underflow and overflow, respectively;
minimizing the overall cost function E to obtain the optimum transmission t*The overall cost function E is expressed as:
E=EcontrastLEloss (7)
where λ L is a weighting parameter that controls the relative importance of contrast cost and information loss cost by controlling λ LLA balance is struck between contrast enhancement and information loss;
and 5: recovering the haze image according to the formula (2) after obtaining the atmospheric light value and the optimal transmittance of the haze weather image;
step 6: labeling the target position and the type of the haze-removing image, and making a target data set;
and 7: performing data enhancement work on a target data set, wherein the enhancement method comprises but is not limited to rotation, translation, gamma transformation, MixUp, Mosaic and fuzzification to obtain a new target data set;
and 8: replacing the residual blocks in the YOLOV4 network with dense link blocks to obtain an improved YOLOV4 network;
a dense link block is made up of a number of 3 x 3 convolutional layers followed by a 1 x 1 convolutional layer, each 3 x 3 convolutional layer is directly connected to all the following 1 x 1 convolutional layers, and the output of the dense link block is represented as:
Figure BDA0002920375610000036
wherein XkRepresents a characteristic diagram, W, of the 3X 3 convolutional layers preceding the 1X 1 convolutional layerskRepresenting the weight, Y, assigned to each signature associated with a 1 x 1 convolutional layeriRepresenting the output of the 1 × 1 convolutional layer in the dense linked block;
and step 9: training the improved YOLOV4 network by taking the image in the new target data set as a sample to obtain a target detection model of the image after haze removal;
step 10: and 2, the detected target is the haze image target detection model from the step 2 to the step 9, the haze image to be detected is input into the haze image target detection model, and the detected target is output.
Preferably, b is 32.
The invention has the following beneficial effects:
1. compared with the traditional atmospheric light calculation method, the method can effectively avoid the problem that the selection of the atmospheric light is inaccurate due to the fact that the object brighter than the atmospheric light can cause.
2. The method establishes a cost function consisting of a contrast item and an information loss item, and through minimizing the cost function, the defogging algorithm furthest reduces the information loss caused by the truncation of the pixel value while enhancing the contrast, stores the information in an optimal mode, and comprehensively considers the problem that the information loss is increased when the contrast is enhanced compared with the traditional method.
3. The method uses the dense link mode to replace a residual link mode in the YOLOV4 network, effectively improves the feature expression capability of the target detection network, and improves the detection capability of the network on complex small targets.
Drawings
FIG. 1 is a flow chart of a haze removal method according to the present invention.
FIG. 2 is a schematic diagram of hierarchical search based on quadtree subdivision according to the method of the present invention.
Fig. 3 is a diagram of a mapping between input (fogged) image pixel values to output (dehazed) image pixel values.
Fig. 4 shows a set of image defogging results under different transmittances according to the embodiment of the invention, in which the original images are sequentially displayed from left to right, and the transmission values are t equal to 0.1, t equal to 0.3, t equal to 0.5, and t equal to 0.7, respectively.
Fig. 5 is a YOLOV4 network residual block diagram.
FIG. 6 is a dense link block diagram.
Detailed Description
The invention is further illustrated with reference to the following figures and examples.
In order to improve the accuracy of target detection and identification under the haze weather condition, the main problems to be solved include the following two points: (1) how to improve the image contrast on the basis of minimum image information loss and effectively remove haze; (2) how to effectively improve the accuracy of target detection and identification. The measures adopted by the invention are as follows: (1) establishing a cost function consisting of a contrast item and an information loss item, and reducing the information loss caused by pixel value truncation to the maximum extent while enhancing the contrast by minimizing the cost function through a defogging algorithm, so that the information is stored in an optimal mode; (2) the dense link mode is used for replacing a residual link mode in the YOLOV4 network, the extraction capability of the network on the detail features of the target is effectively improved, the detection capability of the network on the complex small target is effectively improved, and therefore the detection precision of the overall target is effectively improved.
According to the method, the haze image is subjected to defogging treatment, the YOLOV4 network residual block is replaced by an intensive link mode, fine granularity features of the target can be better extracted, and finally the defogged image is subjected to target identification, so that a target identification task under the haze condition is realized.
A haze image target detection method based on an improved target detection network comprises the following steps:
step 1: acquiring a haze weather image containing a target;
step 2: defining a haze degradation model as shown in formula (1):
i (p) ═ t (p) J (p) +(1-t (p)) a (1) wherein J (p) ═ J (J)r(p),Jg(p),Jb(p))TRepresenting the image after haze removal, Jr(p),Jg(p),Jb(p) pixel values of the image after haze removal in three channels of r, g, and b, respectively, and I (p) ═ Ir(p),Ig(p),Ib(p))TRepresenting an image containing haze Ir(p),Ig(p),Ib(p) respectively representing pixel values of the image containing haze in three channels of r, g and b, wherein p represents the image; a represents the atmospheric light value of the image, t (p) epsilon [0, 1]Represents a transmittance;
and step 3: as shown in fig. 2, determining the atmospheric light value of the haze weather image using a hierarchical search method based on quadtree subdivision;
atmospheric light in a haze degradation model is generally considered to be the brightest color in an image because a large amount of haze produces bright colors. However, in this approach, objects brighter than the atmospheric light may cause inaccuracies in the atmospheric light selection. In order to estimate the atmospheric light more reliably, the fact that: in foggy areas, such as the sky, the variance of the pixel values is typically low.
Dividing a haze weather image into four rectangular areas, subtracting the standard deviation of pixel values in each rectangular area from the average pixel value of each rectangular area, then selecting the rectangular area with the largest calculation result, further dividing the rectangular area into four smaller rectangular areas, and repeating the process until the size of the selected rectangular area is smaller than a pre-specified threshold;
in the finally selected rectangular area, calculating the distance between each pixel point and the white pixel, namely calculating I (I)r(p),Ig(p),Ib(p)) - (255, 255, 255) l, finding a pixel point with the minimum distance to a white pixel, taking the pixel value of the pixel point as an atmospheric light value of a haze weather image, and selecting the atmospheric light which is as bright as possible; in the traditional method, atmospheric light is generally considered as the brightest color in an image, however, an object brighter than the atmospheric light may cause inaccurate selection of the atmospheric light, and in contrast, the atmospheric light calculation method provided by the invention can effectively avoid the problem;
and 4, step 4: calculating an optimal transmittance based on the image block;
step 4-1: dividing the haze weather image into a plurality of image blocks b × b, and assuming that the scene depth of each image block is consistent; setting a transmittance for each image block, the haze degradation model for each image block is written as:
Figure BDA0002920375610000061
after estimation of the atmospheric light a, j (p) depends on the choice of the transmission t (p). Generally, the contrast generated by haze is low, and the contrast of the haze-removed block is increased along with the decrease of the estimated t (p), and it is desirable to estimate the optimal t (p) so that the haze-removed block has the maximum contrast;
step 4-2: the mean square error contrast is adopted to represent the contrast of the haze-removed image block of the haze weather image block, and the mean square error contrast is represented by the formula (3):
Figure BDA0002920375610000062
where c e (r, g, b) represents the color channel index, Jc(p) represents the image block pixel values after haze removal,
Figure BDA0002920375610000063
is Jc(p), N is the number of pixels of the image block; according to formula (2), CMSEThe rewrite is:
Figure BDA0002920375610000064
wherein, Ic(p) represents a haze weather image block pixel value,
Figure BDA0002920375610000065
is Ic(p) average value; the MSE contrast is a decreasing function with respect to t (p). Since contrast is inversely related to the transmission t (p), a smaller value of t (p) may be selected to improve the contrast of the recovered block. However, as t (p) becomes smaller, more pixel values are truncated in the restored image, and as shown in fig. 4, the original image and the defogged image with transmission values t 0.1, t 0.3, t 0.5, and t 0.7 are sequentially displayed from left to right, and the smaller t, the more the pixel level truncation is serious; therefore, enhancing contrast and reducing information loss are contradictory processes;
step 4-3: obtaining the optimum transmittance t*
Defining a contrast cost E by taking the negative sum of the MSE contrasts of the three color channels of each blockcontrastBy minimizing EcontrastMSE contrast can be maximized;
Figure BDA0002920375610000066
defining information loss cost Eloss
Eloss=∑c∈(r,g,b)p∈B{(min{0,Jc(p)})2+(max{0,Jc(p)-255})2} (6)
Where min {0, J }c(p) } and max {0, Jc(p) -255} represent pixel values in the output defogged image that are truncated due to pixel underflow and overflow, respectively, as shown in fig. 3;
minimizing the overall cost function E to obtain the optimum transmission t*The overall cost function E is expressed as:
E=EcontrastLEloss (7)
wherein λ L is a weighting parameter that controls the relative importance of contrast cost and information loss cost, by controlling λ L, a balance is struck between contrast enhancement and information loss;
and 5: as shown in fig. 1, after obtaining the atmospheric light value and the optimal transmittance of the haze weather image, the haze image is restored according to the formula (2);
step 6: labeling the target position and the type of the haze-removing image, and making a target data set;
and 7: performing data enhancement work on a target data set, wherein the enhancement method comprises but is not limited to rotation, translation, gamma transformation, MixUp, Mosaic and fuzzification to obtain a new target data set;
and 8: the YOLOV4 target detection network is optimized and improved on the basis of the YOLOV3 network. The main work comprises the following steps: 1) the data are enhanced by using a Mosaic data enhancement technology, so that the background of a detected object is effectively enriched; 2) using a self-confrontation training data enhancement method that improves the robustness of the model by adding noise to the data; 3) modifying a spatial-wise attention mechanism (spatial-wise attention) of a spatial attention module into a point-wise attention mechanism (point-wise attention), and replacing shortcut connections in a path aggregation network with cascade connections; 4) the Batch Normalization (BN) is replaced by cross-batch normalization (CmBN), 4 mini lots inside the batch are isolated to the outside as a whole, and then statistics of the first three moments are summarized in the mini lots.
In order to improve the network feature extraction capability, the invention replaces the residual block in the Yolov4 network with the dense chainAnd (7) connecting blocks. YOLOV4 uses CSPDarknet53 as a backbone network extraction feature, and CSPDarknet53 is composed of adding CSP blocks on the basis of YOLOV3Darknet53 network. Each CSP block consists of a number of residual blocks, as shown in fig. 5. The residual block can improve the propagation efficiency of the gradient during network training, and the network training speed and the network characteristic propagation efficiency are improved at the same time. Suppose XiRepresenting the output of the i-th layer of the network, Xi+1Is the output of the (i +1) th layer. The output of the residual layer in the network is:
Yi=(Xi+Xi-1)W
wherein W represents the added feature weight; it can be seen that the features output by the ith layer and the (i +1) th layer share the same weight; although the efficiency of network training can be improved, the output of the two layers share one weight, so that the diversity of the network on feature expression is limited, and the capability of the network on extracting tiny target features in a complex environment is limited.
To solve this problem, the residual blocks in the YOLOV4 network are replaced with dense link blocks, resulting in an improved YOLOV4 network;
as shown in fig. 6, a dense link block is composed of a plurality of 3 × 3 convolutional layers followed by 1 × 1 convolutional layers, each 3 × 3 convolutional layer is directly connected to all the following 1 × 1 convolutional layers, and the output of the dense link block is represented as:
Figure BDA0002920375610000081
wherein XkRepresents a characteristic diagram, W, of the 3X 3 convolutional layers preceding the 1X 1 convolutional layerskRepresenting the weight, Y, assigned to each signature associated with a 1 x 1 convolutional layeriRepresenting the output of the 1 × 1 convolutional layer in the dense linked block; it can be seen that the dense linked blocks have a stronger feature expression capability than the residual linked blocks because the dense linked blocks assign different weights to each feature map.
And step 9: training the improved YOLOV4 network by taking the image in the new target data set as a sample to obtain a target detection model of the image after haze removal;
step 10: and 2, the detected target is the haze image target detection model from the step 2 to the step 9, the haze image to be detected is input into the haze image target detection model, and the detected target is output.
The specific embodiment is as follows:
in order to verify the effectiveness of the algorithm, 11243 pictures of different scenes are collected under the haze condition, and the algorithm is verified.
To verify the effectiveness of the defogging algorithm and the modified YOLOV4 network, verification was performed on the haze image dataset and the defogged image dataset using original YOLOV4 and modified YOLOV4, respectively, and the experimental results are shown in table 1. As can be seen from the table, under the same data set, the precision of the improved YOLOV4 is improved to a certain extent compared with that of the original YOLOV4, because the improved YOLOV4 has better network feature extraction capability; under the condition of adopting the same model, the target recognition precision in the defogged image is improved to a certain extent compared with that in the haze image, and the detection accuracy of the deep learning detection network on the target under the haze condition can be effectively improved through the defogging treatment. Therefore, the defogging algorithm based on optimized contrast enhancement is combined with the improved YOLOV4 network, the detection and identification rate of the target under the haze condition is improved to a great extent in numerical value, and the method is extremely effective.
Table 1 comparative experiments on haze and defogging datasets for Yolov4 and modified Yolov4
Figure BDA0002920375610000082
Figure BDA0002920375610000091
In the table, AP represents average accuracy, AP50Denotes an average accuracy when IoU is 0.5 or more, AP75The average accuracy at IoU of 0.75 or more was represented as a strict mode measurement.

Claims (2)

1. A haze image target detection method based on an improved target detection network is characterized by comprising the following steps:
step 1: acquiring a haze weather image containing a target;
step 2: defining a haze degradation model as shown in formula (1):
i (p) ═ t (p) J (p) +(1-t (p)) a (1) wherein J (p) ═ J (J)r(p),Jg(p),Jb(p))TRepresenting the image after haze removal, Jr(p),Jg(p),Jb(p) pixel values of the image after haze removal in three channels of r, g, and b, respectively, and I (p) ═ Ir(p),Ig(p),Ib(p))TRepresenting an image containing haze Ir(p),Ig(p),Ib(p) respectively representing pixel values of the image containing haze in three channels of r, g and b, wherein p represents the image; a represents the atmospheric light value of the image, t (p) epsilon [0, 1]Represents a transmittance;
and step 3: determining an atmospheric light value of the haze weather image by using a hierarchical search method based on quadtree subdivision;
dividing a haze weather image into four rectangular areas, subtracting the standard deviation of pixel values in each rectangular area from the average pixel value of each rectangular area, then selecting the rectangular area with the largest calculation result, further dividing the rectangular area into four smaller rectangular areas, and repeating the process until the size of the selected rectangular area is smaller than a pre-specified threshold;
in the finally selected rectangular area, calculating the distance between each pixel point and the white pixel, namely calculating I (I)r(p),Ig(p),Ib(p)) - (255, 255, 255) l, finding a pixel point with the minimum distance to the white pixel, and taking the pixel value of the pixel point as the atmospheric light value of the haze weather image;
and 4, step 4: calculating an optimal transmittance based on the image block;
step 4-1: dividing the haze weather image into a plurality of image blocks b × b, and assuming that the scene depth of each image block is consistent; setting a transmittance for each image block, the haze degradation model for each image block is written as:
Figure FDA0002920375600000011
step 4-2: the mean square error contrast is adopted to represent the contrast of the haze-removed image block of the haze weather image block, and the mean square error contrast is represented by the formula (3):
Figure FDA0002920375600000015
where c e (r, g, b) represents the color channel index, Jc(p) represents the image block pixel values after haze removal,
Figure FDA0002920375600000012
is Jc(p), N is the number of pixels of the image block; according to formula (2), CMSEThe rewrite is:
Figure FDA0002920375600000013
wherein, Ic(p) represents a haze weather image block pixel value,
Figure FDA0002920375600000014
is Ic(p) average value;
step 4-3: obtaining the optimum transmittance t*
Defining a contrast cost Econtrast
Figure FDA0002920375600000021
Defining information loss cost Eloss
Eloss=∑c∈(r,g,b)p∈B{(min{0,Jc(p)})2+(max{0,Jc(p)-255})2} (6)
Where min {0, J }c(p) } and max {0, Jc(p) -255} represent pixel values in the output defogged image that are truncated due to pixel underflow and overflow, respectively;
minimizing the overall cost function E to obtain the optimum transmission t*The overall cost function E is expressed as:
E=EcontrastLEloss (7)
wherein λLIs a weighting parameter that controls the relative importance of contrast cost and information loss cost by controlling lambdaLA balance is struck between contrast enhancement and information loss;
and 5: recovering the haze image according to the formula (2) after obtaining the atmospheric light value and the optimal transmittance of the haze weather image;
step 6: labeling the target position and the type of the haze-removing image, and making a target data set;
and 7: performing data enhancement work on a target data set, wherein the enhancement method comprises but is not limited to rotation, translation, gamma transformation, MixUp, Mosaic and fuzzification to obtain a new target data set;
and 8: replacing the residual blocks in the YOLOV4 network with dense link blocks to obtain an improved YOLOV4 network;
a dense link block is made up of a number of 3 x 3 convolutional layers followed by a 1 x 1 convolutional layer, each 3 x 3 convolutional layer is directly connected to all the following 1 x 1 convolutional layers, and the output of the dense link block is represented as:
Figure FDA0002920375600000022
wherein XkRepresents a characteristic diagram, W, of the 3X 3 convolutional layers preceding the 1X 1 convolutional layerskRepresenting the weight, Y, assigned to each signature associated with a 1 x 1 convolutional layeriPresentation secretCollecting the output of 1 × 1 convolutional layer in the link block;
and step 9: training the improved YOLOV4 network by taking the image in the new target data set as a sample to obtain a target detection model of the image after haze removal;
step 10: and 2, the detected target is the haze image target detection model from the step 2 to the step 9, the haze image to be detected is input into the haze image target detection model, and the detected target is output.
2. The method for detecting the haze image target based on the improved target detection network as claimed in claim 1, wherein b is 32.
CN202110115251.4A 2021-01-28 2021-01-28 Haze image target detection method based on improved target detection network Pending CN112949389A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110115251.4A CN112949389A (en) 2021-01-28 2021-01-28 Haze image target detection method based on improved target detection network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110115251.4A CN112949389A (en) 2021-01-28 2021-01-28 Haze image target detection method based on improved target detection network

Publications (1)

Publication Number Publication Date
CN112949389A true CN112949389A (en) 2021-06-11

Family

ID=76238342

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110115251.4A Pending CN112949389A (en) 2021-01-28 2021-01-28 Haze image target detection method based on improved target detection network

Country Status (1)

Country Link
CN (1) CN112949389A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113763261A (en) * 2021-06-29 2021-12-07 中国科学院沈阳自动化研究所 Real-time detection method for far and small targets under sea fog meteorological condition

Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104978719A (en) * 2015-06-16 2015-10-14 浙江工业大学 Self-adaptive traffic video real-time defogging method based on temporal-spatial coherence
CN105898111A (en) * 2016-05-06 2016-08-24 西安理工大学 Video defogging method based on spectral clustering
CN105989583A (en) * 2016-07-19 2016-10-05 河海大学 Image defogging method
CN107248146A (en) * 2017-05-22 2017-10-13 哈尔滨工程大学 A kind of UUV Layer Near The Sea Surfaces visible images defogging method
CN108647648A (en) * 2018-05-14 2018-10-12 电子科技大学 A kind of Ship Recognition system and method under visible light conditions based on convolutional neural networks
CN108765302A (en) * 2018-03-29 2018-11-06 西安电子科技大学 The real-time defogging method of image based on GPU
CN108876855A (en) * 2018-05-28 2018-11-23 哈尔滨工程大学 A kind of sea cucumber detection and binocular visual positioning method based on deep learning
CN109919879A (en) * 2019-03-13 2019-06-21 重庆邮电大学 A kind of image defogging method based on dark channel prior Yu bright channel prior
CN110210354A (en) * 2019-05-23 2019-09-06 南京邮电大学 A kind of detection of haze weather traffic mark with know method for distinguishing
CN110428371A (en) * 2019-07-03 2019-11-08 深圳大学 Image defogging method, system, storage medium and electronic equipment based on super-pixel segmentation
CN110807496A (en) * 2019-11-12 2020-02-18 智慧视通(杭州)科技发展有限公司 Dense target detection method
CN110826379A (en) * 2018-08-13 2020-02-21 中国科学院长春光学精密机械与物理研究所 Target detection method based on feature multiplexing and YOLOv3
CN110991311A (en) * 2019-11-28 2020-04-10 江南大学 Target detection method based on dense connection deep network
CN111027511A (en) * 2019-12-23 2020-04-17 西安电子科技大学 Remote sensing image ship detection method based on region of interest block extraction
CN111161170A (en) * 2019-12-18 2020-05-15 江苏科技大学 Underwater image comprehensive enhancement method for target recognition
CN111414807A (en) * 2020-02-28 2020-07-14 浙江树人学院(浙江树人大学) Tidal water identification and crisis early warning method based on YO L O technology
CN111695514A (en) * 2020-06-12 2020-09-22 长安大学 Vehicle detection method in foggy days based on deep learning
CN112084866A (en) * 2020-08-07 2020-12-15 浙江工业大学 Target detection method based on improved YOLO v4 algorithm
CN112232351A (en) * 2020-11-09 2021-01-15 浙江工业职业技术学院 License plate recognition system based on deep neural network

Patent Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104978719A (en) * 2015-06-16 2015-10-14 浙江工业大学 Self-adaptive traffic video real-time defogging method based on temporal-spatial coherence
CN105898111A (en) * 2016-05-06 2016-08-24 西安理工大学 Video defogging method based on spectral clustering
CN105989583A (en) * 2016-07-19 2016-10-05 河海大学 Image defogging method
CN107248146A (en) * 2017-05-22 2017-10-13 哈尔滨工程大学 A kind of UUV Layer Near The Sea Surfaces visible images defogging method
CN108765302A (en) * 2018-03-29 2018-11-06 西安电子科技大学 The real-time defogging method of image based on GPU
CN108647648A (en) * 2018-05-14 2018-10-12 电子科技大学 A kind of Ship Recognition system and method under visible light conditions based on convolutional neural networks
CN108876855A (en) * 2018-05-28 2018-11-23 哈尔滨工程大学 A kind of sea cucumber detection and binocular visual positioning method based on deep learning
CN110826379A (en) * 2018-08-13 2020-02-21 中国科学院长春光学精密机械与物理研究所 Target detection method based on feature multiplexing and YOLOv3
CN109919879A (en) * 2019-03-13 2019-06-21 重庆邮电大学 A kind of image defogging method based on dark channel prior Yu bright channel prior
CN110210354A (en) * 2019-05-23 2019-09-06 南京邮电大学 A kind of detection of haze weather traffic mark with know method for distinguishing
CN110428371A (en) * 2019-07-03 2019-11-08 深圳大学 Image defogging method, system, storage medium and electronic equipment based on super-pixel segmentation
CN110807496A (en) * 2019-11-12 2020-02-18 智慧视通(杭州)科技发展有限公司 Dense target detection method
CN110991311A (en) * 2019-11-28 2020-04-10 江南大学 Target detection method based on dense connection deep network
CN111161170A (en) * 2019-12-18 2020-05-15 江苏科技大学 Underwater image comprehensive enhancement method for target recognition
CN111027511A (en) * 2019-12-23 2020-04-17 西安电子科技大学 Remote sensing image ship detection method based on region of interest block extraction
CN111414807A (en) * 2020-02-28 2020-07-14 浙江树人学院(浙江树人大学) Tidal water identification and crisis early warning method based on YO L O technology
CN111695514A (en) * 2020-06-12 2020-09-22 长安大学 Vehicle detection method in foggy days based on deep learning
CN112084866A (en) * 2020-08-07 2020-12-15 浙江工业大学 Target detection method based on improved YOLO v4 algorithm
CN112232351A (en) * 2020-11-09 2021-01-15 浙江工业职业技术学院 License plate recognition system based on deep neural network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
吴仁彪 等: "雾霾环境下基于PLATE- YOLO的车牌检测方法", 《信号处理》 *
张玥 等: "雾霾图像增强算法对目标检测性能影响研究", 《计算机测量与控制》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113763261A (en) * 2021-06-29 2021-12-07 中国科学院沈阳自动化研究所 Real-time detection method for far and small targets under sea fog meteorological condition
CN113763261B (en) * 2021-06-29 2023-12-26 中国科学院沈阳自动化研究所 Real-time detection method for far small target under sea fog weather condition

Similar Documents

Publication Publication Date Title
CN107301623B (en) Traffic image defogging method and system based on dark channel and image segmentation
CN110490914B (en) Image fusion method based on brightness self-adaption and significance detection
CN107680054B (en) Multi-source image fusion method in haze environment
CN111415363B (en) Image edge identification method
CN112819772B (en) High-precision rapid pattern detection and recognition method
WO2022027931A1 (en) Video image-based foreground detection method for vehicle in motion
CN111062293B (en) Unmanned aerial vehicle forest flame identification method based on deep learning
CN111539980B (en) Multi-target tracking method based on visible light
CN113592911A (en) Apparent enhanced depth target tracking method
CN110782407A (en) Single image defogging method based on sky region probability segmentation
CN111242026A (en) Remote sensing image target detection method based on spatial hierarchy perception module and metric learning
CN112200746A (en) Defogging method and device for traffic scene image in foggy day
CN116883868A (en) Unmanned aerial vehicle intelligent cruising detection method based on adaptive image defogging
CN110349113B (en) Adaptive image defogging method based on dark primary color priori improvement
CN107085830B (en) Single image defogging method based on propagation filtering
CN111274964A (en) Detection method for analyzing water surface pollutants based on visual saliency of unmanned aerial vehicle
CN112949389A (en) Haze image target detection method based on improved target detection network
CN109949239B (en) Self-adaptive sharpening method suitable for multi-concentration multi-scene haze image
CN112288726A (en) Method for detecting foreign matters on belt surface of underground belt conveyor
Shi et al. Image haze removal using dark channel prior and minimizing energy function
CN110633705A (en) Low-illumination imaging license plate recognition method and device
CN109801238B (en) Weak illumination image enhancement method based on simplified atmospheric scattering model
CN109961413B (en) Image defogging iterative algorithm for optimized estimation of atmospheric light direction
Chen et al. Research on safe distance measuring method of front vehicle in foggy environment
CN110599517A (en) Target feature description method based on local feature and global HSV feature combination

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20210611