CN102496021A - Wavelet transform-based thresholding method of image - Google Patents

Wavelet transform-based thresholding method of image Download PDF

Info

Publication number
CN102496021A
CN102496021A CN2011103763913A CN201110376391A CN102496021A CN 102496021 A CN102496021 A CN 102496021A CN 2011103763913 A CN2011103763913 A CN 2011103763913A CN 201110376391 A CN201110376391 A CN 201110376391A CN 102496021 A CN102496021 A CN 102496021A
Authority
CN
China
Prior art keywords
image
gray
threshold
sigma
wavelet
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN2011103763913A
Other languages
Chinese (zh)
Inventor
王恺
杨巨峰
李娇凤
焦姣
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nankai University
Original Assignee
Nankai University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nankai University filed Critical Nankai University
Priority to CN2011103763913A priority Critical patent/CN102496021A/en
Publication of CN102496021A publication Critical patent/CN102496021A/en
Pending legal-status Critical Current

Links

Images

Landscapes

  • Image Processing (AREA)

Abstract

The invention provides a wavelet transform-based thresholding method of an image, and belongs to the field of image processing. In the method, a gray scale of a complete natural scenery image is subjected to wavelet decomposition by utilizing the excellent denoising characteristic of wavelet; foreground characters in the image are removed as noise by virtue of low-pass filter so as to acquire approximate background distribution and foreground distribution; a global threshold is calculated according to the foreground distribution, the global threshold and the background distribution are superposed to form a local threshold which is finally used for image thresholding. The thresholding method provided by the invention can be used for quickly and effectively separating the character part as the foreground to eliminate interference of a complex background, and provides advantages to subsequent character cutting and identifying work. According to the thresholding method provided by the invention, the problem of OCR (optical character recognition) of the natural scenery image can be effectively solved.

Description

Image binaryzation method based on wavelet transformation
Technical field
The invention belongs to technical field of image processing, be specifically related to a kind of image binaryzation method based on wavelet transformation.
Background technology
For the image that contains literal, the purpose of binaryzation is normally separated word segment as prospect.The effect of binaryzation directly has influence on follow-up literal and cuts apart and identification.Than file and picture binary coding method, natural scene image is carried out binaryzation require adaptation of methods property stronger, and can handle multiple complex situations simultaneously.
Research to image binaryzation in recent years deepens continuously, but also considerably less to the research of natural scene image binaryzation specially, only has the people that the natural scene image of monocase has been carried out binaryzation.But for a complete natural scene image, how navigating to character area itself is exactly a difficult problem.Therefore how research is carried out binaryzation to complete image and is had more universal significance.
Wavelet theory (Wavelet theory) is considered to the important breakthrough on mathematical analysis and method in recent years, under the scientist of different ambits makes joint efforts, nowadays solid mathematical theory basis and broad application background has been arranged.At mathematics circle, wavelet analysis is counted as the milestone on the Fourier analysis development history, and it is the perfect adaptation of functional analysis, Fourier analysis, batten analysis, harmonic analysis, numerical analysis.The place that wavelet analysis is superior to the Fourier conversion is; It all has good localization property in time domain and frequency domain; Through changing sampling step length, can focus on any details of object, make people both can see " forest "; Can see " trees " again, so be called as " school microscop ".Wavelet transformation use sharp-pointed gradually temporal resolution for the radio-frequency component of signal in case move nearly observation signal become branch soon, for low-frequency component use sharp-pointed gradually frequency resolution in case move the remote viewing signal become branch (overall variation trend) slowly.The signal analysis representation feature of small echo this " not only seeing trees but also see forest " is very effective to analyzing non-stationary signal.At present, wavelet transformation is widely used in the every field of signal Processing: handle like voice signal, and Digital Image Processing, Digital Video Processing, nonlinear properties processing etc. have become strong tool in scientific research and the practical application at present.
Wavelet filtering is the important application of wavelet transformation in image processing field.The basic thought of wavelet filteration method is: through the multilayer wavelet transformation, represent the absolute value of wavelet coefficient of original image information bigger to original image, and represent the absolute value of wavelet coefficient of noise signal less relatively.Through threshold value is set, with absolute value less than or filter greater than the wavelet coefficient of threshold value, thereby reach the effect of filtering.This research utilizes the good denoising characteristic of small echo just, with the prospect literal in the image as noise removal.Literal identification in the natural scene image can be by means of existing OCR technology; But different with document is; This literal is embedded in the middle of the complex background, and how eliminating the complex background interference better is the key issue that the natural scene image binaryzation will solve.
Summary of the invention
The objective of the invention is to be directed against specially the research of natural scene image binaryzation, further explore how complete image is carried out binaryzation, propose a kind of image binaryzation method based on wavelet transformation.
The present invention utilize wavelet transformation with the prospect literal in the image as noise removal; Thereby background distributions that obtains being similar to and prospect distribute; Form local threshold according to prospect Distribution calculation overall situation binary-state threshold and with global threshold and background distributions stack back again, finally be used for image binaryzation.
The concrete generative process of the inventive method comprises the steps:
Step 1, read in a width of cloth natural scene coloured image, be converted into gray-scale map.
Step 2, background distributions are similar to, and earlier gray-scale map are made L layer wavelet decomposition, obtain the detail coefficients of L layer approximation coefficient LL and three directions, are respectively level detail coefficient HL, vertical detail coefficient LH and diagonal detail coefficient HH.According to a large amount of experiments, it is best that the decomposition number of plies L of small echo gets 6 layers of effect; Pass through the image of the level and smooth word segment of LPF again, and do 1 layer of wavelet reconstruction, obtain the thumbnail of background distributions; Utilize image interpolation that the background distributions thumbnail is amplified to original image size at last, promptly obtain the background distributions figure that is similar to.
Step 3, prospect distribute approximate, and the difference image that the former gray-scale map in background distributions figure and the step 1 is obtained as difference operation is the prospect distribution plan.
Step 4, the threshold value of choosing on a kind of overall binarization method calculating difference image are global threshold.
Step 5, with the background distributions figure stack that global threshold and step 2 obtain, can obtain the binaryzation local threshold of each pixel in the former gray-scale map.
Step 6, the binary-state threshold that obtains according to step 5 are bianry image with the former greyscale image transitions in the step 1.
Wherein, step 1 is described transfers coloured image to gray-scale map GRAY and adopts following formula:
GRAY(x,y)=0.2989*R(I(x,y))+0.5870*G(I(x,y))+0.1140*B(I(x,y))
Wherein, R (.), G (.) and B (.) represent to get redness, green and blue component respectively.
The approximate employing of the described background distributions of step 2 anti-symmetrical bi-orthogonal wavelet is done after the multilayer wavelet decomposition to gray level image; According to the visu thrink method setting threshold that proposes by Dohono; Only the Word message in the top decomposition result is handled, then top decomposition result is carried out reconstruct, get rid of the interference of noise and contextual factor; Adopt bicubic interpolation that the background thumbnail that reconstruct obtains is amplified to original image size at last, promptly obtain the background distributions figure BG that is similar to.
The described prospect of step 3 distributes approximate owing to consider that it is that image is done smoothing processing in essence that the small echo filter is made an uproar; Some pixel (the especially background around the literal) can deepen (it is big that the value of pixel becomes) when filter is made an uproar; When making difference operation, negative value might occur, all directly be changed to 0 with this situation that negative value occurs this moment; Therefore, adopt following formula calculating prospect distribution FG:
FG ( x , y ) = BG ( x , y ) - GRAY ( x , y ) , ifGRAY ( x , y ) < BG ( x , y ) 0 , otherwise .
The described global threshold of step 4 calculates, and adopts fairly simple global threshold computing method: the average μ and the standard deviation sigma of all pixels during at first calculating prospect distributes; Again [| μ-σ |, | μ+σ |] the global threshold GT that looks on the interval gray-scale value to distribute as prospect with minimum projection's number, its computing method shown in following formula,
&mu; = &Sigma; i = 0 255 i * H ( i ) / ( M * N )
&sigma; = 1 M * N - 1 &Sigma; i = 0 255 i * ( H ( i ) - &mu; ) 2
GT = arg min i = &mu; - &sigma; &mu; + &sigma; ( H ( i ) )
Wherein, M and N are respectively the line number and the columns of prospect distribution plan, and H (i) expression prospect distribution grey level histogram is calculated as follows:
H ( i ) = &Sigma; y = 1 N &Sigma; x = 1 M A ( i , FG ( x , y ) ) , i &Element; [ 0,255 ] , A ( m , n ) = 1 , ifm = n 0 , otherwise .
The described local threshold of step 5 calculates according to following formula global threshold and background distributions is superposeed, obtain each pixel in the former gray-scale map (x, binaryzation local threshold LT y) (x, y),
LT(x,y)=GT+BG(x,y)。
The local binary-state threshold that the binaryzation of the described gray level image of step 6 calculates according to step 5; To each the pixel (x in the former gray-scale map; Y), if gray-scale value less than the local binary-state threshold LT of this point (x, y); Then this point is turned to the foreground point by two-value, otherwise this point is turned to background dot by two-value.
Advantage of the present invention and good effect:
The method that the present invention proposes; Gray-scale map to complete natural scene image carries out wavelet decomposition; The prospect that the filtering literal obtains being similar to distributes, and forms the local threshold that finally is used for image binaryzation according to prospect Distribution calculation overall situation binary-state threshold and with global threshold and background distributions stack back again.In order to verify the validity of the inventive method, chosen Otsu, Pavlidis, Liu, Bernsen, Niblack, Sauvola, Gatos, Multi-Scale, 9 kinds of important binarization methods such as block compare test.These binarization methods had both comprised global approach, and partial approach is also arranged; The existing generally acknowledged method of similar research that obtains early also has the up-to-date proposition binarization method of researcher.When selecting, method covers comprehensively.Test is carried out on the ICDAR2003 sample set, and the result shows that the inventive method all is superior to above-mentioned 9 kinds of methods on recall rate, accuracy rate, three indexs of F-measure, and the processing time of single image about 1 second, satisfy application request.It is thus clear that the binarization method based on wavelet transformation that the present invention proposes is adapted to the complicated scene image of background more.
Description of drawings
Fig. 1 is the binaryzation process synoptic diagram based on wavelet transformation;
Fig. 2 is the row extended matrix;
Fig. 3 is the partiting row sampling matrix.
Fig. 4 is a binaryzation process synoptic diagram, wherein, (a) is the gray-scale map after the conversion, (b) individual layer wavelet decomposition result, (c) individual layer wavelet decomposition figure, (d) background distributions figure, (e) prospect distribution plan, (f) prospect distribution grey level histogram, (g) binary map.
Embodiment
Fig. 1 has provided idiographic flow of the present invention, combines embodiment of the invention further explain at present:
1. coloured image transfers gray level image to
Get a fabric width and the high colored natural scene image I that is respectively W=388 and H=543, transfer it to gray level image GRAY according to following formula earlier.Have for
Figure BDA0000111442260000041
y ∈ [1, H]:
GRAY(x,y)=0.2989*R(I(x,y))+0.5870*G(I(x,y))+0.1140*B(I(x,y))
Wherein, R (.), G (.) and B (.) represent to get redness, green and blue component respectively.
Gray-scale map after the conversion is shown in (a) among Fig. 4.
2. background distributions is approximate
Earlier gray level image GRAY is made L layer wavelet decomposition, obtain the detail coefficients of L layer approximation coefficient LL and three directions, be respectively level detail coefficient HL, vertical detail coefficient LH and diagonal detail coefficient HH, shown in (b) among Fig. 4.According to a large amount of experiments, it is best that the decomposition number of plies L of small echo gets 6 layers of effect, adopts anti-symmetrical bi-orthogonal wavelet in the present embodiment, and decomposing number of plies L is 6; Pass through the image of the level and smooth word segment of LPF again, and do 1 layer of wavelet reconstruction, obtain the thumbnail of background distributions; Utilize image interpolation that the background distributions thumbnail is amplified to original image size at last, promptly obtain the background distributions figure BG that is similar to, shown in (d) among Fig. 4.The concrete disposal route in each step is following:
1) wavelet decomposition
Gray level image is carried out 1 layer of wavelet decomposition, be about to the detail coefficients that gray level image is decomposed into approximation coefficient and three directions, shown in (b) among Fig. 4; It is exactly that the approximation coefficient LL among the L-1 layer result is decomposed that gray level image is made L layer wavelet decomposition, obtains new approximation coefficient and detail coefficients.The present invention utilizes anti-symmetrical bi-orthogonal wavelet to make wavelet decomposition; Low pass that it is corresponding and Hi-pass filter matrix of coefficients are respectively
Figure BDA0000111442260000042
and
Figure BDA0000111442260000043
and establish that matrix f to be decomposed is of a size of M*N in the present embodiment; M and N be the height and width of correspondence image respectively; M=H=543, N=W=388.The decomposition step of wavelet transformation is following:
Fig. 2 is the row extended matrix
Figure BDA0000111442260000044
Fig. 3 is the partiting row sampling matrix
Figure BDA0000111442260000045
Respectively expand the matrix f that row obtain M capable (N+2) row by following formula about with matrix f when decomposing (1 layer f=GRAY) Ar:
f ar = f * ( E r N ) T
We get the partial pixel dot matrix f among the matrix f to be decomposed *Operation is done and being specified, the process for the row expansion as follows, below the partial pixel dot matrix f that all takes out with the symbolic representation of upper right corner band " * " *Processing procedure.
f * = 97 96 95 94 97 96 95 94 96 96 95 94
f ar * = 97 97 96 95 94 94 97 97 96 95 94 94 96 96 96 95 94 94
Press following formula with f ArRespectively with
Figure BDA0000111442260000049
With
Figure BDA00001114422600000410
Make convolution, obtain the matrix f of M capable (N+1) row HAnd f G:
f H = f Ar &CircleTimes; H ~ T , Promptly f H ( x - 2 , y - 1 ) = &Sigma; k = x - 2 x - 1 f Ar ( k , y - 1 ) H ~ ( x - k , 1 )
f G = f Ar &CircleTimes; G ~ T , Promptly f G ( x - 2 , y - 1 ) = &Sigma; k = x - 2 x - 1 f Ar ( k , y - 1 ) G ~ ( x - k , 1 )
f H * = 137.2 136.5 135.1 133.6 132.9 137.2 136.5 135.1 133.6 132.9 135.8 135.8 135.1 133.6 132.9
f G * = 0 0.7 0.7 0.7 0 0 0.7 0.7 0.7 0 0 0 0.7 0.7 0
Get f respectively by following formula HAnd f GIn even column and do the row expansion, obtain the matrix f of (M+2) row (N+1)/2 row EHAnd f EG:
f EH = E r M * f H * ( S r N + 1 ) T
f EH * = 136.5 133.6 136.5 133.6 136.5 133.6 135.8 133.6 135.8 133.6
f EG = E r M * f G * ( S r N + 1 ) T
f EG * = 0.7 0.7 0.7 0.7 0.7 0.7 0 0.7 0 0.7
Press following formula with f EHAnd f EGRespectively with
Figure BDA0000111442260000057
With
Figure BDA0000111442260000058
Make convolution, can obtain the matrix of four (M+1) row ((N+1)/2) row:
MA = f EH &CircleTimes; H ~ , Promptly MA ( x , y ) = &Sigma; k = x - 2 x - 1 f EH ( k , y - 1 ) H ~ ( x - k , 1 )
MA * = 193.0 189 193.0 189 192.5 189 192.0 189
MD H = f EH &CircleTimes; G ~ , Promptly M D H ( x , y ) = &Sigma; k = x - 2 x - 1 f EH ( k , y - 1 ) G ~ ( x - k , 1 )
MD H * = 0 0 0 0 0.5 0 0 0
MD V = f EG &CircleTimes; H ~ , Promptly M D V ( x , y ) = &Sigma; k = x - 2 x - 1 f EG ( k , y - 1 ) H ~ ( x - k , 1 )
MD V * = 1.0 1.0 1.0 1.0 0.5 1.0 0 1.0
MD D = f EG &CircleTimes; G ~ , Promptly M D D ( x , y ) = &Sigma; k = x - 2 x - 1 f EG ( k , y - 1 ) G ~ ( x - k , 1 )
MD D * = 0 0 0 0 0.5 0 0 0
Wherein, k is the filtering window span.
Press following formula again to matrix M A, MD H, MD V, MD DAt a distance from the row sampling, can obtain 1 layer of wavelet decomposition result, shown in (b) among Fig. 4, LL is the approximation coefficient of the 1st layer of wavelet decomposition, and HL is that level detail coefficient, LH are the vertical detail coefficient, and HH is the diagonal detail coefficient.Each matrix of consequence is ((M+1)/2) row ((N+1)/2) row, is followed successively by A, the D of matrix of the partial pixel point composition of taking-up from left to right H, D VAnd D D:
A = MA * ( S r N + 1 ) T , D H = MD H * ( S r N + 1 ) T , D V = MD V * ( S r N + 1 ) T , D D = MD D * ( S r N + 1 ) T , A , D H , D V, D DUpper left, upper right, down left and lower right-most portion, i.e. LL, HL, LH, HH shown in (b) 1 layer of wavelet decomposition result among Fig. 4 in the difference corresponding diagram 4 in (c).
A * = 193.0 189 192.0 189 D H * = 0 0 0 0 D V * = 1.0 1.0 0 1.0 D D * = 0 0 0 0
By above-mentioned steps A is carried out wavelet decomposition, promptly obtain 1 layer of wavelet decomposition result, the rest may be inferred, can calculate 6 layers of wavelet decomposition result.Because the gray-scale map text color is shallow than background color, need in the present embodiment doing wavelet decomposition after its inverse again.
2) wavelet filtering
The wavelet filtering process be exactly with among the wavelet filtering result greater than threshold value T nDetail coefficients be changed to 0, keep less than T nValue, that is:
D H ( x , y ) = D H ( x , y ) , if D H ( x , y ) < T n 0 , otherwise
D V ( x , y ) = D V ( x , y ) , if D V ( x , y ) < T n 0 , otherwise
D D ( x , y ) = D D ( x , y ) , if D D ( x , y ) < T n 0 , otherwise
The threshold value the most frequently used according to wavelet filtering, the visu thrink method preset threshold that is promptly proposed by Dohono is handled the Word message in the top decomposition result, word segment.Threshold value calculation method is formula as follows:
T n=σ n*sqrt(2lnN)
Wherein, σ n=c/0.6745, c are the intermediate value of the absolute value of small echo detail coefficients, N=3*H L* W L, H LAnd W LBe the row and column that obtains matrix of coefficients after the L layer wavelet decomposition, the threshold value that calculates in the present embodiment is 204.6.Detail coefficients after 6 layers of wavelet decomposition is as follows, and we find after doing filtering according to threshold value that because the partial pixel point that we take out belongs to background pixel, the value of detail coefficients remains less than threshold value the most at last, and word segment will be filtered.
D H * = [ 2.1 ] D V * = [ 7.9 ] D D * = [ - 1.2 ]
3) wavelet reconstruction
Wavelet reconstruction is the inverse operation of wavelet decomposition, to top coefficient of wavelet decomposition A L,
Figure BDA0000111442260000072
Utilize coefficient of wavelet decomposition with
Figure BDA0000111442260000073
With
Figure BDA0000111442260000074
Associate matrix
Figure BDA0000111442260000075
With
Figure BDA0000111442260000076
The convolution sum decomposition result is carried out reconstruct, shown in following formula:
f r = F H L * ( ( E H L * A &CircleTimes; H ) * ( E W L ) T &CircleTimes; H T ) * ( F W L ) T
+ F H L * ( ( E H L * D H &CircleTimes; G ) * ( E W L ) T &CircleTimes; H T ) * ( F W L ) T
+ F H L * ( ( E H L * D V &CircleTimes; H ) * ( E W L ) T &CircleTimes; G T ) * ( F W L ) T
+ F H L * ( ( E H L * D D &CircleTimes; G ) * ( E W L ) T &CircleTimes; G T ) * ( F W L ) T
f r * = 95 95 95 95
Wherein,
4) image interpolation
The present invention adopts bicubic interpolation, utilizes and to treat that the gray scale of 16 points is done cubic interpolation around the sampled point, considers that not only the gray scale of 4 direct neighbor points influences, and considers the influence of gray-value variation rate between each adjoint point.The bicubic interpolation algorithm need be chosen the interpolation basis function and come fitting data, and its form is shown below:
S ( w ) = 1 - 2 | w | 2 + + | w | 3 | w | < 1 4 - 8 | w | + 5 | w | 2 - | w | 3 1 &le; | w | < 2 0 | w | &GreaterEqual; 2
Carry out bicubic interpolation according to following formula, can obtain the image array f after the interpolation R, promptly approximate background distributions figure BG is shown in (d) among Fig. 4.f RIdentical with the original image matrix size:
f R(i+u,j+v)=A*B*C
f R * = 95 95 95 95 95 95 95 95 95 95 95 95
Wherein, A, B, C are matrix, and its form is following:
A=[S(1+u)?S(u)?S(1-u)?S(2-u)]
B = f ( i - 1 , j - 2 ) f ( i , j - 2 ) f ( i + 1 , j - 2 ) f ( i + 2 , j - 2 ) f ( i - 1 , j - 1 ) f ( i , j - 1 ) f ( i + 1 , j - 1 ) f ( i + 2 , j - 1 ) f ( i - 1 , j ) f ( i , j ) f ( i + 1 , j ) f ( i + 2 , j ) f ( i - 1 , j + 1 ) f ( i , j + 1 ) f ( i + 1 , j + 1 ) f ( i + 1 , j + 1 )
C=[S(1+v)?S(v)?S(1-v)?S(2-v)] T
3. prospect distributes approximate
Background distributions figure BG and former gray-scale map GRAY are made difference operation can obtain prospect distribution plan FG.Obviously, in prospect distributed, the value of former gray-scale map background pixel point should level off to 0, the value of foreground pixel point should be away from 0.Because it is that image is done smoothing processing in essence that the small echo filter is made an uproar; Some pixel (the especially background around the literal) can deepen (it is big that the value of pixel becomes) when filter is made an uproar; When making difference operation, negative value might occur, all directly be changed to 0 with this situation that negative value occurs this moment.Therefore, the prospect that finally is calculated as follows distributes:
FG ( x , y ) = BG ( x , y ) - G ( x , y ) , ifG ( x , y ) < BG ( x , y ) 0 , otherwise
The prospect distribution plan that calculates in the present embodiment is obtained as difference operation according to following formula with (d) background distributions figure by (a) former gray-scale map among Fig. 4 shown in (e) among Fig. 4.
4. global threshold calculates
In prospect distributed, a large amount of background pixel point values were 0 or approach 0 that the value of foreground pixel point is then away from 0.Because in the prospect distribution plan, preceding background difference is bigger, therefore adopt fairly simple global threshold computing method here: the average μ and the standard deviation sigma of all pixels during at first calculating prospect distributes; Again [| μ-σ |, | μ+σ |] the global threshold GT that looks on the interval gray-scale value to distribute as prospect with minimum projection's number, its computing method are shown in following formula.
&mu; = &Sigma; i = 0 255 i * H ( i ) / ( M * N )
&sigma; = 1 M * N - 1 &Sigma; i = 0 255 i * ( H ( i ) - &mu; ) 2
GT = arg min i = &mu; - &sigma; &mu; + &sigma; ( H ( i ) )
H (i) expression prospect distribution grey level histogram, its computing method are shown below:
H ( i ) = &Sigma; y = 1 N &Sigma; x = 1 M A ( i , FG ( x , y ) ) , i &Element; [ 0,255 ]
Wherein, A ( m , n ) = 1 , Ifm = n 0 , Otherwise .
Present embodiment calculates prospect distribution grey level histogram according to above-mentioned formula, shown in (f) among Fig. 4.The average μ of all pixels=15 in the prospect distribution plan, standard deviation sigma=30, [| μ-σ |, | μ+σ |] the global threshold GT=45 that finds on the interval, such as among the figure (f) mark.
5. local threshold calculates
By following formula global threshold GT and background distributions BG are superposeed, can obtain each pixel among the former gray-scale map GRAY (x, binaryzation local threshold LT y) (x, y).The place that literal is arranged in former gray-scale map, its local threshold is bigger, makes can well be kept in the binary image Chinese words; And be the place of background in former gray-scale map, its local threshold is less, makes most of background in binary image, remain background.
LT(x,y)=GT+BG(x,y)
6. binaryzation
Convert gray level image GRAY into bianry image BW by following formula.(x, y), (x, y), then this point is turned to the foreground point by two-value, otherwise this point is turned to background dot by two-value if gray-scale value is less than the local binary-state threshold LT of this point for each pixel among the former gray-scale map GRAY.The binary map that (g) is converted to for (a) among Fig. 4.
BW ( x , y ) = 1 , ifGRAY ( x , y ) < LT ( x , y ) 0 , otherwise

Claims (7)

1. the image binaryzation method based on wavelet transformation is characterized in that this method comprises the steps:
Step 1, read in a width of cloth natural scene coloured image, be converted into gray-scale map;
Step 2, background distributions are similar to, and earlier gray-scale map are made L layer wavelet decomposition, obtain the detail coefficients of L layer approximation coefficient LL and three directions, are respectively level detail coefficient HL, vertical detail coefficient LH and diagonal detail coefficient HH; Pass through the image of the level and smooth word segment of LPF again, and do 1 layer of wavelet reconstruction, obtain the thumbnail of background distributions; Utilize image interpolation that the background distributions thumbnail is amplified to original image size at last, promptly obtain the background distributions figure that is similar to;
Step 3, prospect distribute approximate, and the difference image that the former gray-scale map in background distributions figure and the step 1 is obtained as difference operation is the prospect distribution plan;
Step 4, the threshold value of choosing on a kind of overall binarization method calculating difference image are global threshold;
Step 5, with the background distributions figure stack that global threshold and step 2 obtain, can obtain the binaryzation local threshold of each pixel in the former gray-scale map;
Step 6, the binary-state threshold that obtains according to step 5 are bianry image with the former greyscale image transitions in the step 1.
2. method according to claim 1 is characterized in that step 1 is described and transfers coloured image to gray-scale map GRAY and adopt following formula:
GRAY(x,y)=0.2989*R(I(x,y))+0.5870*G(I(x,y))+0.1140*B(I(x,y))
Wherein, R (.), G (.) and B (.) represent to get redness, green and blue component respectively.
3. method according to claim 1; It is characterized in that the approximate anti-symmetrical bi-orthogonal wavelet that adopts of the described background distributions of step 2 does after the multilayer wavelet decomposition gray level image; According to the visu thrink method setting threshold that proposes by Dohono; Only the Word message in the top decomposition result is handled, then top decomposition result is carried out reconstruct, get rid of the interference of noise and contextual factor; Adopt bicubic interpolation that the background thumbnail that reconstruct obtains is amplified to original image size at last, promptly obtain the background distributions figure BG that is similar to.
4. method according to claim 1; It is characterized in that the described prospect of step 3 distributes approximate owing to consider that it is that image is done smoothing processing in essence that the small echo filter is made an uproar; Some pixel can deepen pixel when filter is made an uproar value becomes big; Especially the background pixel point around the literal negative value might occur when making difference operation, and all directly be changed to 0 with this situation that negative value occurs this moment; Therefore, adopt following formula to calculate prospect distribution plan FG:
FG ( x , y ) = BG ( x , y ) - GRAY ( x , y ) , ifGRAY ( x , y ) < BG ( x , y ) 0 , otherwise .
5. method according to claim 1 is characterized in that the described global threshold of step 4 calculates, and adopts fairly simple global threshold computing method: the average μ and the standard deviation sigma of all pixels during at first calculating prospect distributes; Again [| μ-σ |, | μ+σ |] the global threshold GT that looks on the interval gray-scale value to distribute as prospect with minimum projection's number, its computing method shown in following formula,
&mu; = &Sigma; i = 0 255 i * H ( i ) / ( M * N )
&sigma; = 1 M * N - 1 &Sigma; i = 0 255 i * ( H ( i ) - &mu; ) 2
GT = arg min i = &mu; - &sigma; &mu; + &sigma; ( H ( i ) )
Wherein, M and N are respectively the line number and the columns of prospect distribution plan, and H (i) expression prospect distribution grey level histogram is calculated as follows:
H ( i ) = &Sigma; y = 1 N &Sigma; x = 1 M A ( i , FG ( x , y ) ) , i &Element; [ 0,255 ] , A ( m , n ) = 1 , ifm = n 0 , otherwise .
6. method according to claim 1 is characterized in that the described local threshold of step 5 calculates according to following formula global threshold and background distributions are superposeed, obtain each pixel in the former gray-scale map (x, binaryzation local threshold LT y) (x, y),
LT(x,y)=GT+BG(x,y)。
7. method according to claim 1; It is characterized in that the local binary-state threshold that the binaryzation of the described gray level image of step 6 calculates according to step 5, to each pixel in the former gray-scale map (x, y); If gray-scale value is less than the local binary-state threshold LT (x of this point; Y), then this point is turned to the foreground point by two-value, otherwise this point is turned to background dot by two-value.
CN2011103763913A 2011-11-23 2011-11-23 Wavelet transform-based thresholding method of image Pending CN102496021A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN2011103763913A CN102496021A (en) 2011-11-23 2011-11-23 Wavelet transform-based thresholding method of image

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN2011103763913A CN102496021A (en) 2011-11-23 2011-11-23 Wavelet transform-based thresholding method of image

Publications (1)

Publication Number Publication Date
CN102496021A true CN102496021A (en) 2012-06-13

Family

ID=46187846

Family Applications (1)

Application Number Title Priority Date Filing Date
CN2011103763913A Pending CN102496021A (en) 2011-11-23 2011-11-23 Wavelet transform-based thresholding method of image

Country Status (1)

Country Link
CN (1) CN102496021A (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105335972A (en) * 2015-10-20 2016-02-17 江南大学 Warp knitting fabric defect detection method based on wavelet contourlet transformation and visual saliency
CN105447491A (en) * 2014-07-25 2016-03-30 北京大学深圳研究生院 Signboard image binaryzation method and device
WO2017020723A1 (en) * 2015-08-04 2017-02-09 阿里巴巴集团控股有限公司 Character segmentation method and device and electronic device
CN106447904A (en) * 2016-09-09 2017-02-22 深圳怡化电脑股份有限公司 Paper money identifying method and apparatus
CN106778761A (en) * 2016-12-23 2017-05-31 潘敏 A kind of processing method of vehicle transaction invoice
WO2017128603A1 (en) * 2016-01-26 2017-08-03 上海葡萄纬度科技有限公司 Educational toy kit and mirror position detection method thereof
CN108734669A (en) * 2017-04-24 2018-11-02 南京理工大学 Image denoising method based on wavelet transformation Wiener filtering and edge detection
RU2743224C1 (en) * 2020-04-07 2021-02-16 Федеральное государственное казенное военное образовательное учреждение высшего образования "Военный учебно-научный центр Военно-воздушных сил "Военно-воздушная академия имени профессора Н.Е. Жуковского и Ю.А. Гагарина" (г. Воронеж) Министерства обороны Российской Федерации Method of detecting small objects on images
CN113365082A (en) * 2021-07-12 2021-09-07 北京爱芯科技有限公司 Haar wavelet transform hardware device
US11302286B2 (en) 2018-07-20 2022-04-12 Huawei Technologies Co., Ltd. Picture obtaining method and apparatus and picture processing method and apparatus

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6292583B1 (en) * 1997-09-30 2001-09-18 Advantest Corporation Image information processing apparatus
JP3520166B2 (en) * 1996-11-19 2004-04-19 株式会社リコー Image processing device
CN1941838A (en) * 2005-09-29 2007-04-04 株式会社理光 File and picture binary coding method
CN101093538A (en) * 2006-06-19 2007-12-26 电子科技大学 Method for identifying iris based on zero crossing indication of wavelet transforms

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3520166B2 (en) * 1996-11-19 2004-04-19 株式会社リコー Image processing device
US6292583B1 (en) * 1997-09-30 2001-09-18 Advantest Corporation Image information processing apparatus
CN1941838A (en) * 2005-09-29 2007-04-04 株式会社理光 File and picture binary coding method
CN101093538A (en) * 2006-06-19 2007-12-26 电子科技大学 Method for identifying iris based on zero crossing indication of wavelet transforms

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
王蓉晖 等: "基于小波能量的自然纹理背景中多目标检测方法", 《长春理工大学学报》, vol. 28, no. 3, 30 September 2005 (2005-09-30) *
许礼武 等: "基于小波分解的车牌定位算法", 《计算机工程》, vol. 32, no. 21, 30 November 2006 (2006-11-30) *

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105447491A (en) * 2014-07-25 2016-03-30 北京大学深圳研究生院 Signboard image binaryzation method and device
WO2017020723A1 (en) * 2015-08-04 2017-02-09 阿里巴巴集团控股有限公司 Character segmentation method and device and electronic device
US10552705B2 (en) 2015-08-04 2020-02-04 Alibaba Group Holding Limited Character segmentation method, apparatus and electronic device
CN105335972A (en) * 2015-10-20 2016-02-17 江南大学 Warp knitting fabric defect detection method based on wavelet contourlet transformation and visual saliency
CN105335972B (en) * 2015-10-20 2018-11-30 江南大学 Knitted fabric defect detection method based on small echo contourlet transform and vision significance
WO2017128603A1 (en) * 2016-01-26 2017-08-03 上海葡萄纬度科技有限公司 Educational toy kit and mirror position detection method thereof
CN106447904A (en) * 2016-09-09 2017-02-22 深圳怡化电脑股份有限公司 Paper money identifying method and apparatus
CN106778761A (en) * 2016-12-23 2017-05-31 潘敏 A kind of processing method of vehicle transaction invoice
CN108734669A (en) * 2017-04-24 2018-11-02 南京理工大学 Image denoising method based on wavelet transformation Wiener filtering and edge detection
US11302286B2 (en) 2018-07-20 2022-04-12 Huawei Technologies Co., Ltd. Picture obtaining method and apparatus and picture processing method and apparatus
RU2743224C1 (en) * 2020-04-07 2021-02-16 Федеральное государственное казенное военное образовательное учреждение высшего образования "Военный учебно-научный центр Военно-воздушных сил "Военно-воздушная академия имени профессора Н.Е. Жуковского и Ю.А. Гагарина" (г. Воронеж) Министерства обороны Российской Федерации Method of detecting small objects on images
CN113365082A (en) * 2021-07-12 2021-09-07 北京爱芯科技有限公司 Haar wavelet transform hardware device

Similar Documents

Publication Publication Date Title
CN102496021A (en) Wavelet transform-based thresholding method of image
CN111062892B (en) Single image rain removing method based on composite residual error network and deep supervision
CN101872472B (en) Method for super-resolution reconstruction of facial image on basis of sample learning
CN109102469B (en) Remote sensing image panchromatic sharpening method based on convolutional neural network
US20180232857A1 (en) Method and device for super-resolution image reconstruction based on dictionary matching
CN102005037B (en) Multimodality image fusion method combining multi-scale bilateral filtering and direction filtering
Zhang et al. Region of interest extraction in remote sensing images by saliency analysis with the normal directional lifting wavelet transform
CN1822046A (en) Infrared and visible light image fusion method based on regional property fuzzy
CN101430759A (en) Optimized recognition pretreatment method for human face
CN101404084A (en) Infrared image background suppression method based on Wavelet and Curvelet conversion
CN107909560A (en) A kind of multi-focus image fusing method and system based on SiR
CN112365414A (en) Image defogging method based on double-path residual convolution neural network
CN104537678A (en) Method for removing cloud and mist from single remote sensing image
CN103218811A (en) Statistical distribution-based satellite multi-spectral image waveband registration method
CN112712500A (en) Remote sensing image target extraction method based on deep neural network
CN111652033A (en) Lane line detection method based on OpenCV
CN103324958A (en) License plate locating method based on projection method and SVM under complex background
CN102496144A (en) NSCT (nonsubsampled contourlet transform) sonar image enhancement method based on HSV (hue, saturation and value) color space
CN106971377A (en) A kind of removing rain based on single image method decomposed based on sparse and low-rank matrix
CN104463822A (en) Multi-focus image fusing method and device based on multi-scale overall filtering
CN105139362A (en) Image enhancing method based on frequency domain
CN105069755A (en) Multi-channel haze image enhancement processing device and method
CN108764287B (en) Target detection method and system based on deep learning and packet convolution
CN105654070A (en) Low-resolution face recognition method
CN112734675A (en) Image rain removing method based on pyramid model and non-local enhanced dense block

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C02 Deemed withdrawal of patent application after publication (patent law 2001)
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20120613