CN115170523B - Low-complexity infrared dim target detection method based on local contrast - Google Patents

Low-complexity infrared dim target detection method based on local contrast Download PDF

Info

Publication number
CN115170523B
CN115170523B CN202210833296.XA CN202210833296A CN115170523B CN 115170523 B CN115170523 B CN 115170523B CN 202210833296 A CN202210833296 A CN 202210833296A CN 115170523 B CN115170523 B CN 115170523B
Authority
CN
China
Prior art keywords
area
sub
value
contrast
target
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210833296.XA
Other languages
Chinese (zh)
Other versions
CN115170523A (en
Inventor
马永奎
王俊杰
张佳岩
赵洪林
单成兆
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Harbin Institute of Technology
Original Assignee
Harbin Institute of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Harbin Institute of Technology filed Critical Harbin Institute of Technology
Priority to CN202210833296.XA priority Critical patent/CN115170523B/en
Publication of CN115170523A publication Critical patent/CN115170523A/en
Application granted granted Critical
Publication of CN115170523B publication Critical patent/CN115170523B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/90Dynamic range modification of images or parts thereof
    • G06T5/94Dynamic range modification of images or parts thereof based on local image properties, e.g. for local contrast enhancement
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/136Segmentation; Edge detection involving thresholding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10048Infrared image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20021Dividing image into blocks, subimages or windows
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20024Filtering details
    • G06T2207/20032Median filtering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20036Morphological image processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Quality & Reliability (AREA)
  • Image Analysis (AREA)

Abstract

A low-complexity infrared weak and small target detection method based on local contrast relates to the technical field of target detection, and aims at solving the problem of low target detection speed under a complex background in the prior art, the method utilizes the characteristics that the infrared weak and small target has larger difference from the background and has higher contrast locally to detect the infrared weak and small target, and the transmission bandwidth of a digital system is fully utilized through an efficient filtering template in the implementation process, so that the image traversing time is reduced, and the system processing speed is greatly improved; meanwhile, the algorithm complexity is reduced, floating point operation and nonlinear operation are reduced, the processing time delay of the system is reduced, the design of a large-scale assembly line is facilitated, and the hardware implementation difficulty is reduced. Finally, the method can realize reliable high-speed infrared weak and small target detection, the throughput of the actually measured system reaches 20Gbps, the detection rate is higher than 90%, the false alarm rate is lower than 10%, and the high-speed detection can be realized in engineering application.

Description

Low-complexity infrared dim target detection method based on local contrast
Technical Field
The invention relates to the technical field of target detection, in particular to a low-complexity infrared dim target detection method based on local contrast.
Background
The infrared imaging target detection system has the advantages of capability of working day and night, no interference from natural environments such as cloud and fog, and the like, and strong anti-interference capability. Therefore, the infrared imaging system is widely applied to the military, industrial and civil fields of all-weather security monitoring, unmanned aerial vehicle disaster relief detection and weapon guidance.
Low signal-to-noise ratio small object detection in complex contexts is a very challenging task because objects in an image are typically very small and lack specific shape, texture and structure information, which results in small objects often occupying very small pixels in an image and being strongly disturbed by background clutter or random noise. In order to solve the above problems, the mechanism of the human visual system has been applied in the field of infrared dim target detection in recent years, which mainly uses the visual attention mechanism, the brightness and contrast sensitivity mechanism of human eyes, and the like, and adopts saliency maps to realize feature extraction. The detection algorithm based on human eye vision is mainly divided into 2 types, the first type is an algorithm based on spectrum residual error, the main principle is that the spectrum of an infrared image is subjected to a low-pass filter to obtain a smoothed spectrum, then the smoothed spectrum is subjected to difference with the original spectrum to obtain spectrum residual error, and the small target position is detected based on high-frequency components in the residual error. The other type is a detection algorithm based on local contrast, which uses the big difference between the brightness of a small target and the neighborhood of the small target, and adopts the idea of a sliding window to detect the contrast in the sliding window and search the position of the small target.
Because the local contrast measurement algorithm has small calculation amount and good detection effect compared with a frequency spectrum residual error method, the method is used and improved by a plurality of scholars, the improvement thinking is to improve the mode of calculating the contrast and to improve the window, but the algorithm is designed aiming at software implementation, and a plurality of floating points and nonlinear operations are used, so that the method is not beneficial to parallel implementation of a high-speed digital image processing system based on an FPGA; due to the serial processing capability of the CPU or the DSP, the algorithm has high processing time delay and poor instantaneity, and meanwhile, the filtering window is not selected too large, so that the time for traversing a pair of images is too long, and the limited bandwidth resource in the transmission process of the digital signal is not fully utilized.
Disclosure of Invention
The purpose of the invention is that: aiming at the problem of low target detection speed under a complex background in the prior art, the low-complexity infrared weak and small target detection method based on local contrast is provided.
The technical scheme adopted by the invention for solving the technical problems is as follows:
a low-complexity infrared dim target detection method based on local contrast comprises the following steps:
step one: acquiring an infrared image of an object to be identified, and preprocessing the infrared image of the object to be identified, wherein the preprocessing comprises morphological open operation filtering processing and maximum median filtering processing;
step two: the method comprises the steps of carrying out contrast detection on a preprocessed infrared image by using a high-efficiency filtering template, wherein the high-efficiency filtering template is a square template, and the high-efficiency filtering template is 2 n ×2 n The subblocks are formed, n is more than or equal to 2 and less than or equal to 5, and the size of each subblock is 8 multiplied by 8 pixels;
the contrast detection comprises the following steps:
step two,: utilizing 3 multiplied by 3 sub-blocks in the efficient filtering template as a processing unit, wherein the central sub-block of the processing unit is a target area to be processed, and the rest sub-blocks are background areas;
step two: the processing unit is used as a measuring window to process the preprocessed infrared image, and the specific steps are as follows:
step two, one by one: sequencing all pixels in a target area to be processed by gray values, taking the maximum value and the next maximum value of the gray values, and calculating the average value of the maximum value and the next maximum value, wherein the average value is used as an A value;
step two, two: calculating the average value of gray values of pixels in the sub-blocks corresponding to the background area, wherein the average value is taken as a B value;
step two, one and three: acquiring three sub-blocks corresponding to one direction of 8 directions of a target area to be processed, and then acquiring the average value of B values of the three sub-blocks, wherein the value is taken as a C value;
step two, one and four: the value A and the value C are subjected to difference, and the result is the contrast measurement value of the corresponding direction in 8 directions of the target area to be processed;
step two and step five: repeating the second step, the third step and the fourth step to obtain contrast measurement values in 8 directions;
the target area in the processing unit is area 0 The background area is area from left to right and from top to bottom 1 、area 2 、area 3 、area 4 、area 5 、area 6 、area 7 、area 8
The 8 directions are specifically as follows: up direction neighborhood (area) 1 ,area 2 ,area 3 ) Down direction neighborhood (area) 6 ,area 7 ,area 8 ) Left direction neighborhood (area) 1 ,area 4 ,area 6 ) Right direction neighborhood (area) 3 ,area 5 ,area 8 ) A first quadrant neighborhood (area) 2 ,area 3 ,area 5 ) A second quadrant neighborhood (area) 2 ,area 1 ,area 4 ) A third quadrant neighborhood (area) 4 ,area 6 ,area 7 ) A fourth quadrant neighborhood (area) 5 ,area 8 ,area 7 );
Step three: taking the target area to be processed as a pixel of the saliency map, taking absolute values of contrast measurement values in 8 directions, and summing the absolute values to obtain a gray value of the pixel;
step four: repeating the second step and the third step until the infrared image is processed, so as to obtain a saliency map;
step five: and (3) carrying out threshold segmentation on the saliency map by using the constant false alarm rate, extracting the region with the contrast larger than the threshold value from the segmentation result, and carrying out position output.
Further, the efficient filtering template is a double-layer template, the efficient filtering template comprises a first-layer template and a second-layer template, and the first-layer template is 2 n ×2 n Square templates of sub-blocks, the second layer template is 2 n -1×2 n -a square template of 1 sub-block, every four adjacent sub-blocks in the first layer template corresponding to one sub-block in the second layer template, the intersection points of the four adjacent sub-blocks in the first layer template coinciding with the center points of the sub-blocks in the second layer template.
Further, the specific steps of contrast detection are as follows:
step 1: area of the target area 0 All pixels in the target region are ordered according to gray values, and the average value of the maximum pixel and the next maximum pixel of the gray values is used as the saliency brightness Lmax of the target region;
step 2: carrying out average pooling operation on each background area to obtain the average value of the gray level of each background area, and using the average value of the gray level of each background area to represent the saliency brightness Lmean of the background area i ,i=1,2,…,8;
Step 3: saliency Lmean according to background region i Obtaining the saliency Lneighbor of each direction m
Step 4: and obtaining a contrast measurement value by utilizing the difference between the saliency brightness of the target area and the saliency brightness of the target area in 8 directions.
Further, the saliency luminance Lmax of the target area is expressed as:
Figure BDA0003746408690000031
/>
wherein G is 1 、G 2 Representing the maximum value pixel and the next-maximum value pixel of the target area.
Further, the salient luminance Lmean of the background area i Expressed as:
Figure BDA0003746408690000032
wherein G is i (s, t) represents the gray value at the position (s, t) of the i-th block region, i being the number of the 3 sub-blocks required for the contrast in the current direction.
Further, the significance luminance Lneighbor of each direction m Expressed as:
Figure BDA0003746408690000033
further, the contrast measurement is expressed as:
contrast m =Lmax-Lneighbor m
where m=1, 2, …,8.
Further, the threshold is expressed as:
T h =m c +k·σ
wherein m is c The background clutter mean sigma is the root mean square of noise.
Further, the infrared image is acquired by an infrared camera.
The beneficial effects of the invention are as follows:
the infrared weak and small target detection method utilizes the characteristic that the infrared weak and small target has larger difference with the background and has higher contrast in local to carry out infrared weak and small target detection, and the transmission bandwidth of the digital system is fully utilized through the efficient filtering template in the implementation process, so that the image traversing time is reduced, and the system processing speed is greatly improved; meanwhile, the algorithm complexity is reduced, floating point operation and nonlinear operation are reduced, the processing time delay of the system is reduced, the design of a large-scale assembly line is facilitated, and the hardware implementation difficulty is reduced. Finally, the method can realize reliable high-speed infrared weak and small target detection, the throughput of the actually measured system reaches 20Gbps, the detection rate is higher than 90%, the false alarm rate is lower than 10%, and the high-speed detection can be realized in engineering application.
The method utilizes the characteristic that the infrared dim target has larger difference with the background and has higher contrast locally to detect the infrared dim target, a novel efficient filtering template is also used in detection, the efficient filtering template is a square template with the size of 2 and the even power number of pixels, and the divided sub-blocks in the template are squares with the size of 2 and the power number of pixels, and as the digital system generally takes the power of 2 as the transmission bit width, the method can be matched with the transmission bandwidth of various digital systems. The transmission bandwidth of the digital system is fully utilized through the efficient filtering template, so that the image traversing time can be reduced, and the system processing speed is greatly improved; meanwhile, the algorithm complexity is reduced, floating point operation and nonlinear operation are reduced, the processing time delay of the system is reduced, the design of a large-scale assembly line is facilitated, and the hardware implementation difficulty is reduced. Finally, reliable and rapid detection of the infrared weak target is realized.
Drawings
FIG. 1 is a flow chart of the present application;
FIG. 2 is a first layer schematic of a high efficiency filtering template;
FIG. 3 is a second layer schematic of the high efficiency filter template;
FIG. 4 is a schematic diagram of a measurement cell;
FIG. 5 is a schematic view of 8 measurement contrast directions;
FIG. 6 is a block diagram of a system hardware platform;
FIG. 7 is a schematic diagram of an embodiment of the present application;
FIG. 8 is a schematic diagram of an original infrared image in infrared small target detection;
FIG. 9 is a schematic view of a salient image obtained after processing in infrared dim target detection;
FIG. 10 is a schematic diagram of the detection result in infrared dim target detection;
FIG. 11 is a schematic diagram of a generic filtering template 1;
FIG. 12 is a schematic diagram of a generic filtering template 2;
FIG. 13 is a schematic diagram of a generic filtering template 3;
FIG. 14 is a schematic diagram of a high efficiency filtering template usage process;
FIG. 15 is a diagram of the actual correspondence of the high-efficiency filtering templates;
FIG. 16 is a schematic diagram of a relationship between a transmission unit and a processing unit;
FIG. 17 is a schematic diagram of a saliency map generation process 1;
FIG. 18 is a schematic diagram of a saliency map generation process 2;
FIG. 19 is a schematic diagram of a saliency map generation process 3;
FIG. 20 is an original view;
FIG. 21 is a saliency map;
fig. 22 is a schematic diagram of correspondence of the saliency map.
Detailed Description
It should be noted in particular that, without conflict, the various embodiments disclosed herein may be combined with each other.
The first embodiment is as follows: referring to fig. 1, the method for detecting a low-complexity infrared weak and small target based on local contrast according to the present embodiment specifically includes the following steps:
step 1: preprocessing an infrared image;
step 1.1: morphological open operation filtering: morphological opening operations are typically used to smooth the contours of objects, eliminating elongated protrusions. Highlight noise in the preprocessed image will be filtered out. The morphological open operation process is as follows:
Figure BDA0003746408690000051
the formula represents that the structural element B performs the following operation on the image A: first B corrodes a, then B swells a.
Step 1.2: maximum median filtering: since small objects are easily submerged by background clutter of an infrared image, the infrared background can be predicted by utilizing the regional correlation of the background clutter, and the object or noise and the background have no correlation and cannot be predicted, so that after the predicted background is subtracted from the original image, only a residual image containing the object and the noise is left. The method is realized by adopting a maximum median filtering method, the gray value of each pixel point is set to be the maximum value of the median values along different directions in a window of a certain field of the point, and the detail characteristics of a small target can be well protected. The algorithm process is to take a sliding window on the image, calculate the median value on each line along the horizontal, vertical and two diagonal directions respectively, compare the four median values, and then take the original value at the position of the largest substituted center pixel. The maximum median filter formula is:
y(x,y) Max-median =max(z 1 ,z 2 ,z 3 ,z 4 ) (2)
z in formula (VI) i Representing the condition of a certain dotted lineThe median of the dry pixels corresponds to the mathematical expression:
z 1 =median{x(m,n-N),...,x(m,n),...,x(m,n+N)} (3)
z 2 =median{x(m-N,n),...,x(m,n),...,x(m-N,n)} (4)
z 3 =median{x(m+N,n-N),...,x(m,n),...,x(m-N,n+N)} (5)
z 4 =median{x(m-N,n-N),...,x(m,n),...,x(m+N,n+N)} (6)
the method can effectively inhibit the background, enhance the contrast between the target and the surrounding background, and facilitate the subsequent processing. The residual image which only contains the target and the noise after the background inhibition can be obtained by making the difference between the background predicted by the method and the original image.
Step 2: multidirectional contrast detection:
the application utilizes the high-efficiency filtering template to carry out contrast detection on the preprocessed infrared image, the high-efficiency filtering template is a square template, and the high-efficiency filtering template comprises 2 n ×2 n Sub-blocks, n is more than or equal to 2 and less than or equal to 5, and the size of each sub-block is 8 multiplied by 8 pixels; processing the infrared image by taking 3 multiplied by 3 subblocks in the efficient filtering template as a processing unit, wherein the central subblock of the processing unit is a target area to be processed, and the rest subblocks are background areas;
for example: the efficient filtering template comprises 4×4 sub-blocks, and the processing unit is 3×3 sub-blocks, as shown in fig. 2, where the target area to be processed is four sub-blocks of "5, 6, 9, 10";
taking a sub-block '10' as an example, firstly sequencing all pixels in the sub-block '10' by gray values, then taking the maximum value and the next maximum value of the gray values, and then calculating the average value of the maximum value and the next maximum value;
then calculating the mean value of the pixel values of three sub-blocks in each of the 8 directions of the sub-block '10', namely the mean value of '13, 14 and 15', the mean value of '5, 6 and 7', the mean value of '7, 11 and 15', the mean value of '5, 9 and 13', the mean value of '9, 13 and 14', the mean value of '14, 15 and 11', the mean value of '11, 7 and 6', and the mean value of '9, 5 and 6';
respectively differencing the average value of the maximum value and the next maximum value with the average value of the pixel values of the three sub-blocks in each of the 8 directions to obtain 8 contrast measurement values;
the saliency map is a compressed sub-block, the sub-block corresponding to the target area to be processed is compressed to be used as a pixel of the saliency map, and the pixel value of the pixel is 8 contrast measurement values, takes absolute values and sums the absolute values.
In addition, the high efficiency filtering template may be 2 layers, the first layer comprising 2 n ×2 n Sub-blocks, the second layer comprising 2 n -1×2 n -1 sub-block, the first layer and the second layer being square templates, each four adjacent sub-blocks in the first layer template corresponding to one sub-block in the second layer template, the intersection points of the four adjacent sub-blocks in the first layer template coinciding with the center points of the sub-blocks in the second layer template;
in application, for example, the first layer comprises 4×4 sub-blocks, the second layer comprises 3×3 sub-blocks, the first layer template and the second layer template are parallelized, the result of the first layer template is used as the first layer of the saliency map, the result of the second layer template is used as the second layer of the saliency map, the two layers of efficient filtering templates can obtain 5 pixel points of the saliency map at a time, and the one layer of efficient filtering templates can only obtain 4 pixel points at a time. The order of the sliding of the sub-blocks is from top to bottom and from left to right.
area 0 In order to enhance the contrast of the target, the present application sorts all pixels of the target region, and takes out the average value of the maximum value and the next-largest value to represent the significant brightness of the region, where the definition of the brightness as Lmax has a calculation formula:
Figure BDA0003746408690000071
g in 1 、G 2 Representing the maximum value pixel and the next-maximum value pixel of the target area.
area 1 ~area 8 The area is a reference background area, the inventionThe mean-pooling operation is performed on each region, and the mean value is used to represent the salient brightness of the region, and the brightness is recorded as Lmean i Then there is a calculation formula:
Figure BDA0003746408690000072
g in i (s, t) represents the gradation value at the position (s, t) of the i-th block region.
After having the salient luminance of each region, the present application continues to define the neighborhood luminance in eight directions. First define a neighborhood (area 1 ,area 2 ,area 3 ) For the upward neighborhood, a similar definition (area 6 ,area 7 ,area 8 ) Is a down direction neighborhood (area) 1 ,area 4 ,area 6 ) Is left direction neighborhood (area) 3 ,area 5 ,area 8 ) Is a right direction neighborhood (area) 2 ,area 3 ,area 5 ) Is a first quadrant neighborhood (area) 2 ,area 1 ,area 4 ) Is a second quadrant neighborhood (area) 4 ,area 6 ,area 7 ) Is a third quadrant neighborhood (area) 5 ,area 8 ,area 7 ) Is a fourth quadrant neighborhood. Defining each neighborhood brightness as Lneighbor i Then there is a calculation formula:
Figure BDA0003746408690000073
lmean in k Representing the saliency brightness of the three neighborhoods in the corresponding direction, m=1, 2, …, representing 8 directions, i representing the number of 3 sub-blocks needed to calculate the contrast in the current direction.
After defining the saliency luminance in the eight-direction neighborhood, the present application represents the contrast in a certain direction by using the difference between the saliency luminance of the target area and the saliency luminance in that direction i Namely the formula:
contrast m =Lmax-Lneighbor m (10)
when the target sliding window traverses the whole picture, the contrast of the infrared image in 8 directions in a certain area can be obtained.
Step 3: constant false alarm rate threshold segmentation: the constant false alarm rate characteristic can keep the false alarm rate of the infrared detection system unchanged. Infrared image noise can be generally described by a gaussian probability density function with zero mean:
Figure BDA0003746408690000074
m is in c Is the mean value of background clutter; m is m t The real strength of the target point; sigma is the root mean square of noise; x is x i Is the gray value somewhere in the image.
Further signal-to-noise ratio and threshold signal-to-noise ratio of the signal may be defined:
Figure BDA0003746408690000081
Figure BDA0003746408690000082
t in h Representing a threshold segmentation threshold, the system considers that an object is detected when the gray value at a certain place in the image is higher than the threshold, and then the detection probability of the system can be calculated as follows:
Figure BDA0003746408690000083
order the
Figure BDA0003746408690000084
Then there are:
Figure BDA0003746408690000085
false alarm probability P of system fa The threshold signal-to-noise ratio can be used to represent:
Figure BDA0003746408690000086
should keep T h With the mean value m of background clutter c And the noise root mean square sigma dynamic change can keep the constant false alarm rate of the system, so that a threshold value calculation formula can be obtained:
T h =m c +TNR·σ (17)
in order to facilitate design and adjustment, TNR is replaced by an inputtable parameter k in the subsequent design, different false alarm probabilities are set according to specific scenes, so that different k values are calculated for input, and a final threshold calculation formula is as follows:
T h =m c +k·σ (18)
and (3) carrying out threshold segmentation on the saliency map by using the constant false alarm rate, if the contrast ratio in the segmentation result is larger than the threshold value, proving that the region has high local contrast ratio, and outputting the region coordinates to finish infrared weak and small target detection.
Aiming at the two-layer efficient filtering templates, the parameters of the 2 layers of the saliency map are processed, and the pixels of the saliency map obtained by the two-layer efficient filtering templates are more and denser, so that the result is more accurate.
The infrared dim target detection method utilizes the characteristic that the infrared dim target has larger difference from the background and has higher contrast locally to detect the infrared dim target, and in the implementation process, the transmission bandwidth of a digital system is fully utilized through the high-efficiency filtering template, wherein the high-efficiency filtering template is a square template with the size of 2 and the number of pixels of even power, and the divided sub-blocks in the template are squares with the size of 2 and the number of pixels of power, so that the image traversing time can be reduced by using the template, and the system processing speed is greatly improved; meanwhile, the algorithm complexity is reduced, floating point operation and nonlinear operation are reduced, the processing time delay of the system is reduced, the design of a large-scale assembly line is facilitated, and the hardware implementation difficulty is reduced. Finally, the method can realize reliable high-speed infrared weak and small target detection, the throughput of the actually measured system reaches 20Gbps, the detection rate is higher than 90%, the false alarm rate is lower than 10%, the high-speed detection can be realized in engineering application, and the processing speed index is far higher than that of the similar algorithm.
Examples:
the invention is realized on a high-speed digital image processing board designed by an FPGA and DSP dual-processor architecture, the FPGA is a main processor model Kintex7, and is responsible for realizing most algorithm functions and peripheral chip control, the DSP chip is a coprocessor, and is responsible for managing processed data and communicating with an upper computer. The actual hardware platform and chip model are shown in fig. 5: the total processing rate is 20Gbps, the input of a 8X SRIO port of the system is equivalent to 8 paths of parallel processing, and each path requires the minimum rate of 2.5Gbps; the single-path SRIO adopts a working mode of 3.125Gbps, and the working main frequency is 125MHz; the caching chip adopts a DDR3 chip, the size is 2GB, the working speed is 1333MT/s, 20 pictures can be cached, the reading and writing operation clock frequency is 166MHz, and the reading and writing bit width is 512 bits.
In fig. 7, the specific implementation steps are as follows:
step 1: high-speed transmission of digital images;
image data acquired by the infrared camera is input into the system through a high-speed interface. The invention adopts 2 SRIO interfaces with 4X working modes, the single-port rate is set to be 3.125Gbps, and the highest throughput rate which can be achieved is 25Gbps. In order to meet the stability of high-speed data transmission and reduce transmission failure and rate loss caused by data retransmission, an SRIO interface generally needs a low-jitter high-precision clock, and the Si5368 clock chip is adopted to meet the requirement in the invention, and the Si5368 clock chip stably outputs a 125MHz high-precision clock.
Step 2: a cache of digital images;
the original infrared image is cached in a cache chip to wait for processing. The invention uses DDR3 chip to buffer data, 8 pieces DDR3 are altogether, the chip model is MT41J256M8HX_15E, and the control module is designed based on MIG IP core in VIVADO 2017.4.
Step 3: and (5) rapidly detecting the infrared weak and small target.
The infrared weak target step is as described in the detailed description.
Fig. 8, fig. 9 and fig. 10 show the processing results of the method in the practical application, fig. 8 represents the input original image, wherein 6 infrared dim targets in different backgrounds exist, fig. 9 is a significant graph output after contrast detection, and fig. 10 is a detection result output after threshold segmentation.
Further elaborating on efficient filtering templates
Distinction of efficient filter templates from other filter templates: the high-efficiency filtering templates are all 2-power-of-power pixels in length and width and are divided into 2-power sub-blocks. For example, the template in this application is 32×32 pixels in size, and then one sub-block is formed every 8×8 pixels. The common filtering templates generally do not satisfy the power of 2 pixels in length and width. A common filtering template is typically centrosymmetric and is typically divided into odd sub-blocks, e.g. 5 x 5, 7 x7, which results in that the length and width of the template must not be a power of 2 pixels. The first layer of the efficient filtering template is shown in fig. 2, and the second layer of the efficient filtering template is shown in fig. 3. In practical application, the efficient filtering template is 2 n ×2 n The square template of each sub-block is more than or equal to 2 and less than or equal to 5, the size of each sub-block is 8 multiplied by 8 pixels, the effective filtering template is necessarily square, the lower limit of 4 multiplied by 4 sub-blocks is at least capable of containing a processing unit for the efficient filtering template, and the upper limit of 32 multiplied by 32 sub-blocks is larger in resource consumption after the size of the template is increased, so that the effective filtering template is not practical any more. The specific size is configured with the actual situation of the digital system.
Fig. 14 shows a filtering template actually used in the present application, where a black square represents a pixel, a painted portion represents a sub-block, and three sub-blocks are divided, and a matrix of 32×32 pixels is transmitted in a digital system, which is a basic "transmission unit", and 3×3 sub-blocks in the template are taken out to form a "processing unit", where the processing unit is used to detect contrast, and a "transmission unit" includes 5 "processing units", where the actual correspondence is as shown in fig. 15:
the processing unit refers to a square area of 3×3 sub-block size (fixed value), as shown in fig. 16:
wherein area is 0 ~area 8 Representing 8 sub-blocks, the weak and small target occupying less than 9×9 pixels in the infrared image according to the international optical engineering council definition, so each sub-block is defined as 8×8 pixels in size, e.g. area 0 The actual size of (a) is 8 x 8 pixels (fixed value). From this, it can be deduced that a processing unit must be 24 x 24 pixels in size.
The left side of fig. 15 represents transmission of a high-efficiency filtering template matrix in the system, the middle of the dotted line of fig. 15 represents a contrast measurement process, 3×3 sub-blocks are taken out from the high-efficiency filtering template to form a "processing unit", the contrast of the central gray area of each processing unit is calculated according to a contrast measurement algorithm, and the contrast of 5 small sub-blocks is obtained after calculation, such as the gray sub-blocks on the right side of fig. 15. The center of the processing unit is defined as area 0 The background is defined as area 1 ~area 8 . The white background of fig. 15 represents the efficient filtering template, i.e. the transmission unit, and the color part represents one processing unit.
Saliency map generation process
Fig. 17 is a view showing the removal of one processing unit from the "transmission unit" to calculate the surrounding sub-block mean and the maximum/minimum value of the sub-block at the center. The resulting data may continue to calculate contrast for a certain direction and generate a saliency map, e.g. calculate contrast for sub-block 10:
upper direction contrast= (1/2) × (maximum value+next largest value) - (1/3) × (upper direction field pixel mean sum)
=(1/2)×(200+195)-(1/3)×(36+96+66)
=131.5
Lower contrast= (1/2) × (200+195) - (1/3) × (55+38+78)
=140.5
First quadrant direction contrast= (1/2) × (200+195) - (1/3) × (96+66+48)
=127.5
Similarly, the contrast in 8 directions is calculated:
left direction contrast=158.8, right direction contrast=133.5, second quadrant direction contrast=145.2, third quadrant contrast=158.2, fourth quadrant contrast=142.8.
Such a calculation process is present 5 times in total in one transmission unit.
The "transmission unit" is a sliding window in the original view from the scale of the whole figure, as shown in fig. 18:
the sliding window slides on the original image to obtain transmission units, and the 1 transmission unit divides 5 processing units for contrast detection and can obtain the contrast of 5 central areas in all directions.
A saliency map is then calculated. The saliency map is also a "map" consisting of pixels, introducing the calculation rules for each pixel:
after calculating the contrast in each direction of the sub-block 10 in fig. 5, the contrast in each direction may be taken as an absolute value and summed to obtain 1 pixel of the saliency map.
For example, the 10 th sub-block of the current sliding window corresponds to the pixel value of the saliency map, the upward direction contrast +|the downward direction contrast |+ …
=|131.5|+|140.5|+|158.8|+|133.5|+|127.5|+|145.2|+|158.2|+|142.8|
=1138
The absolute value must be taken because the calculated contrast in a certain direction is not necessarily positive.
Similarly, sub-blocks 5, 6, 9 and 20 of the sliding window can calculate a pixel of the saliency map, and the saliency map is double-layered because the sliding window is double-layered. The 1 transmission unit can calculate 5 pixels of the saliency map. As shown in fig. 7.
The sliding window slides the distance of two sub-blocks (2×8=16 pixels) on the original image each time, so that 4 blue sub-blocks of the first layer can be seamlessly spliced into the first layer of the saliency map, and the second layer of the saliency map is gapped. As shown in fig. 20 and 21.
The white part in fig. 20 and 21 is the original image, the gray part is the calculated saliency map, and 1 pixel in the saliency map represents 8×8 pixels (1 sub-block) in the original image; FIG. 20 is a first layer of saliency maps, with the left dashed box representing 4 saliency map pixels calculated at position 1 of the slide window (gray patches representing 1 pixel of the saliency map), and the right dashed box representing 4 saliency map pixels calculated at position 2 after sliding the slide window by a distance of 2 sub-blocks; fig. 21 is the second layer of saliency maps, the 5 th saliency map pixel calculated for each sliding window. The correspondence is as shown in fig. 22:
after the saliency map is calculated, each pixel of the saliency map is compared with a threshold value.
It should be noted that the detailed description is merely for explaining and describing the technical solution of the present invention, and the scope of protection of the claims should not be limited thereto. All changes which come within the meaning and range of equivalency of the claims and the specification are to be embraced within their scope.

Claims (9)

1. A low-complexity infrared dim target detection method based on local contrast is characterized by comprising the following steps of: the method comprises the following steps:
step one: acquiring an infrared image of an object to be identified, and preprocessing the infrared image of the object to be identified, wherein the preprocessing comprises morphological open operation filtering processing and maximum median filtering processing;
step two: the method comprises the steps of carrying out contrast detection on a preprocessed infrared image by using a high-efficiency filtering template, wherein the high-efficiency filtering template is a square template, and the high-efficiency filtering template is 2 n ×2 n The subblocks are formed, n is more than or equal to 2 and less than or equal to 5, and the size of each subblock is 8 multiplied by 8 pixels;
the contrast detection comprises the following steps:
step two,: utilizing 3 multiplied by 3 sub-blocks in the efficient filtering template as a processing unit, wherein the central sub-block of the processing unit is a target area to be processed, and the rest sub-blocks are background areas;
step two: the processing unit is used as a measuring window to process the preprocessed infrared image, and the specific steps are as follows:
step two, one by one: sequencing all pixels in a target area to be processed by gray values, taking the maximum value and the next maximum value of the gray values, and calculating the average value of the maximum value and the next maximum value, wherein the average value is used as an A value;
step two, two: calculating the average value of gray values of pixels in the sub-blocks corresponding to the background area, wherein the average value is taken as a B value;
step two, one and three: acquiring three sub-blocks corresponding to one direction of 8 directions of a target area to be processed, and then acquiring the average value of B values of the three sub-blocks, wherein the value is taken as a C value;
step two, one and four: the value A and the value C are subjected to difference, and the result is the contrast measurement value of the corresponding direction in 8 directions of the target area to be processed;
step two and step five: repeating the second step, the third step and the fourth step to obtain contrast measurement values in 8 directions;
the target area in the processing unit is area 0 The background area is area from left to right and from top to bottom 1 、area 2 、area 3 、area 4 、area 5 、area 6 、area 7 、area 8
The 8 directions are specifically as follows: up direction neighborhood (area) 1 ,area 2 ,area 3 ) Down direction neighborhood (area) 6 ,area 7 ,area 8 ) Left direction neighborhood (area) 1 ,area 4 ,area 6 ) Right direction neighborhood (area) 3 ,area 5 ,area 8 ) A first quadrant neighborhood (area) 2 ,area 3 ,area 5 ) A second quadrant neighborhood (area) 2 ,area 1 ,area 4 ) A third quadrant neighborhood (area) 4 ,area 6 ,area 7 ) A fourth quadrant neighborhood (area) 5 ,area 8 ,area 7 );
Step three: taking the target area to be processed as a pixel of the saliency map, taking absolute values of contrast measurement values in 8 directions, and summing the absolute values to obtain a gray value of the pixel;
step four: repeating the second step and the third step until the infrared image is processed, so as to obtain a saliency map;
step five: and (3) carrying out threshold segmentation on the saliency map by using the constant false alarm rate, extracting the region with the contrast larger than the threshold value from the segmentation result, and carrying out position output.
2. The method for detecting the low-complexity infrared dim target based on the local contrast according to claim 1, wherein the efficient filtering template is a double-layer template, the efficient filtering template comprises a first-layer template and a second-layer template, and the first-layer template is 2 n ×2 n Square templates of sub-blocks, the second layer template is 2 n -1×2 n -a square template of 1 sub-block, every four adjacent sub-blocks in the first layer template corresponding to one sub-block in the second layer template, the intersection points of the four adjacent sub-blocks in the first layer template coinciding with the center points of the sub-blocks in the second layer template.
3. The method for detecting the low-complexity infrared dim target based on the local contrast according to claim 2, wherein the specific steps of the contrast detection are as follows:
step 1: area of the target area 0 All pixels in the target region are ordered according to gray values, and the average value of the maximum pixel and the next maximum pixel of the gray values is used as the saliency brightness Lmax of the target region;
step 2: carrying out average pooling operation on each background area to obtain the average value of the gray level of each background area, and using the average value of the gray level of each background area to represent the saliency brightness Lmean of the background area i ,i=1,2,…,8;
Step 3: saliency Lmean according to background region i Obtaining the saliency Lneighbor of each direction m
Step 4: and obtaining a contrast measurement value by utilizing the difference between the saliency brightness of the target area and the saliency brightness of the target area in 8 directions.
4. A low-complexity infrared small target detection method based on local contrast according to claim 3, characterized in that the saliency luminance Lmax of the target area is expressed as:
Figure FDA0003746408680000021
wherein G is 1 、G 2 Representing the maximum value pixel and the next-maximum value pixel of the target area.
5. The method for detecting a low-complexity infrared dim target based on local contrast according to claim 4, wherein said background region has a significant luminance Lmean i Expressed as:
Figure FDA0003746408680000022
wherein G is i (s, t) represents the gray value at the position (s, t) of the i-th block region, i being the number of the 3 sub-blocks required for the contrast in the current direction.
6. The method for detecting low-complexity infrared dim target based on local contrast according to claim 5, wherein said salient luminance Lneighbor in each direction is characterized by m Expressed as:
Figure FDA0003746408680000023
7. a method of detecting a low-complexity infrared small target based on local contrast according to claim 6, wherein the contrast measure is expressed as:
contrast m =Lmax-Lneighbor m
where m=1, 2, …,8.
8. A low-complexity infrared small target detection method based on local contrast according to claim 1, characterized in that the threshold value is expressed as:
T h =m c +k·σ
wherein m is c The background clutter mean sigma is the root mean square of noise.
9. The method for detecting the low-complexity infrared dim target based on the local contrast according to claim 1, wherein the infrared image is acquired by an infrared camera.
CN202210833296.XA 2022-07-14 2022-07-14 Low-complexity infrared dim target detection method based on local contrast Active CN115170523B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210833296.XA CN115170523B (en) 2022-07-14 2022-07-14 Low-complexity infrared dim target detection method based on local contrast

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210833296.XA CN115170523B (en) 2022-07-14 2022-07-14 Low-complexity infrared dim target detection method based on local contrast

Publications (2)

Publication Number Publication Date
CN115170523A CN115170523A (en) 2022-10-11
CN115170523B true CN115170523B (en) 2023-05-05

Family

ID=83495410

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210833296.XA Active CN115170523B (en) 2022-07-14 2022-07-14 Low-complexity infrared dim target detection method based on local contrast

Country Status (1)

Country Link
CN (1) CN115170523B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117671801B (en) * 2024-02-02 2024-04-23 中科方寸知微(南京)科技有限公司 Real-time target detection method and system based on binary reduction

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107403433A (en) * 2016-05-20 2017-11-28 南京理工大学 A kind of complicated cloud infrared small target in background detection method
CN110135312A (en) * 2019-05-06 2019-08-16 电子科技大学 A kind of quick small target detecting method based on classification LCM
CN110751667A (en) * 2019-09-17 2020-02-04 中国科学院上海技术物理研究所 Method for detecting infrared dim small target under complex background based on human visual system
CN114549642A (en) * 2022-02-10 2022-05-27 中国科学院上海技术物理研究所 Low-contrast infrared weak and small target detection method

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103646232B (en) * 2013-09-30 2016-08-17 华中科技大学 Aircraft ground moving target infrared image identification device

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107403433A (en) * 2016-05-20 2017-11-28 南京理工大学 A kind of complicated cloud infrared small target in background detection method
CN110135312A (en) * 2019-05-06 2019-08-16 电子科技大学 A kind of quick small target detecting method based on classification LCM
CN110751667A (en) * 2019-09-17 2020-02-04 中国科学院上海技术物理研究所 Method for detecting infrared dim small target under complex background based on human visual system
CN114549642A (en) * 2022-02-10 2022-05-27 中国科学院上海技术物理研究所 Low-contrast infrared weak and small target detection method

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
刘旭 ; 崔文楠 ; .采用人类视觉对比机制的红外弱小目标检测.红外技术.2020,(第06期),全文. *
王晓阳 ; 彭真明 ; 张萍 ; 孟晔铭 ; .局部对比度结合区域显著性红外弱小目标检测.强激光与粒子束.2015,(第09期),全文. *
王铎 ; .基于形态学和邻域差值的红外小目标检测算法.光电技术应用.2016,(第02期),全文. *

Also Published As

Publication number Publication date
CN115170523A (en) 2022-10-11

Similar Documents

Publication Publication Date Title
Han et al. Underwater image processing and object detection based on deep CNN method
CN110472627B (en) End-to-end SAR image recognition method, device and storage medium
CN109685732B (en) High-precision depth image restoration method based on boundary capture
CN110782477A (en) Moving target rapid detection method based on sequence image and computer vision system
CN111476159A (en) Method and device for training and detecting detection model based on double-angle regression
CN107403134B (en) Local gradient trilateral-based image domain multi-scale infrared dim target detection method
CN104834915B (en) A kind of small infrared target detection method under complicated skies background
CN107248174A (en) A kind of method for tracking target based on TLD algorithms
CN112364865B (en) Method for detecting small moving target in complex scene
CN115170523B (en) Low-complexity infrared dim target detection method based on local contrast
Chen et al. A color-guided, region-adaptive and depth-selective unified framework for Kinect depth recovery
Liu et al. Image edge recognition of virtual reality scene based on multi-operator dynamic weight detection
CN114972423A (en) Aerial video moving target detection method and system
CN116612468A (en) Three-dimensional target detection method based on multi-mode fusion and depth attention mechanism
CN115272876A (en) Remote sensing image ship target detection method based on deep learning
CN112329764A (en) Infrared dim target detection method based on TV-L1 model
Sun et al. IRDCLNet: Instance segmentation of ship images based on interference reduction and dynamic contour learning in foggy scenes
CN111709964A (en) PCBA target edge detection method
Liu et al. [Retracted] Self‐Correction Ship Tracking and Counting with Variable Time Window Based on YOLOv3
CN113205494A (en) Infrared small target detection method and system based on adaptive scale image block weighting difference measurement
CN110910497A (en) Method and system for realizing augmented reality map
CN115908202A (en) ToF depth image denoising method based on expansion modeling and multi-mode fusion
CN113834447B (en) High-dynamic laser light bar self-adaptive imaging processing method in outdoor complex environment
CN111008555B (en) Unmanned aerial vehicle image small and weak target enhancement extraction method
CN115953589B (en) Engine cylinder block aperture size measurement method based on depth camera

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant