CN102867195A - Method for detecting and identifying a plurality of types of objects in remote sensing image - Google Patents

Method for detecting and identifying a plurality of types of objects in remote sensing image Download PDF

Info

Publication number
CN102867195A
CN102867195A CN2012103006458A CN201210300645A CN102867195A CN 102867195 A CN102867195 A CN 102867195A CN 2012103006458 A CN2012103006458 A CN 2012103006458A CN 201210300645 A CN201210300645 A CN 201210300645A CN 102867195 A CN102867195 A CN 102867195A
Authority
CN
China
Prior art keywords
image
matrix
class
target
remote sensing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN2012103006458A
Other languages
Chinese (zh)
Other versions
CN102867195B (en
Inventor
韩军伟
周培诚
王东阳
郭雷
程塨
李晖晖
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Northwestern Polytechnical University
Original Assignee
Northwestern Polytechnical University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Northwestern Polytechnical University filed Critical Northwestern Polytechnical University
Priority to CN201210300645.8A priority Critical patent/CN102867195B/en
Publication of CN102867195A publication Critical patent/CN102867195A/en
Application granted granted Critical
Publication of CN102867195B publication Critical patent/CN102867195B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Image Analysis (AREA)

Abstract

The invention relates to a method for detecting and identifying a plurality of types of objects in a remote sensing image based on sparse representation dictionary learning. The method is technically characterized by comprising the following steps of: training preprocessed training data to form a dictionary by using a sparse representation dictionary training method; performing sparse coding on the dictionary obtained by training by using an image sub-block in a test image, calculating a sparse representation coefficient to further obtain a reconstruction error of the image sub-block, and determining a candidate target area by thresholding the reconstruction error; and accurately detecting and identifying a plurality of types of objects in the remote sensing image by using post processing. By the method, a plurality of types of objects in the remote sensing image under a complex background can be detected and identified. The method is high in detection and identification accuracy and low in false alarm rate.

Description

A kind of remote sensing images multi-class targets detection and Identification method
Technical field
The present invention relates to a kind of remote sensing images multi-class targets detection and Identification method, can be applied to the polytype Target detection and identification under the complex background remote sensing images.
Background technology
An application as Remote Sensing Image Processing Technology, object detection and recognition under the complex background remote sensing images is a gordian technique in the fields such as military surveillance and precision strike, also be study hotspot and the difficult point in this field always, there is important military and civilian to be worth, is subject to people and more and more pays close attention to.
Remote Sensing Target detects the two kinds of methods that mainly contain at present.A kind of is that some shape, the geometric properties that has by the detection target in remote sensing images solves target detection problems, but because the remote sensing images background is complicated, exist a large amount of shape, geometric properties similar with target, only rely on these features to detect target and a large amount of undetected, flase drop can occur.Another kind is based on the thought of classification, wherein modal is Bag-of-Words(BoW) sorting technique, the method at first is that image is extracted SIFT feature and cluster, with cluster centre as the one group of standard base (image-region of standard) in the image space, then can organize the standard base with this image is carried out vector representation, at last resulting vector is classified and thresholding by use svm classifier device, obtain testing result; But the BoW method although the SIFT feature of extracting has yardstick and rotational invariance, has only been utilized the statistical nature of characteristic area, and has been ignored the spatial information of characteristic area, therefore uses the method verification and measurement ratio of BoW low, and false alarm rate is high; And another sorting technique Linear Spatial Pyramid Matching Using Sparse Coding(ScSPM) although considered the spatial information of characteristic area, resulting vectorial dimension for classification is too high, operand is excessive.In addition, most also only limits to simple target is detected based on the object detection method of classification, can not detect and identify a plurality of targets simultaneously.
Summary of the invention
The technical matters that solves
For fear of the deficiencies in the prior art part, the present invention proposes a kind of method of the remote sensing images multi-class targets detection and Identification based on the rarefaction representation dictionary learning.This method can automatically detect from the remote sensing images of complex background and identify dissimilar target, has higher accuracy of detection and lower false alarm rate.
Technical scheme
A kind of remote sensing images multi-class targets detection and Identification method is characterized in that step is as follows:
Step 1: use the method training dictionary based on the rarefaction representation dictionary learning, concrete steps are as follows:
Step a1 training image is processed in earlier stage: at first with generic target unification to a principal direction in the original image, then will unify after the direction image along 0 ° to 360 °, according to step-length
Figure BDA00002042658500021
Rotate to be
Figure BDA00002042658500022
The image of individual different directions; The original image of different classes of target is all processed according to the method described above, obtained
Figure BDA00002042658500023
The class training image, wherein p is the different classes of number of targets that will detect,
Figure BDA00002042658500024
Be the anglec of rotation, c is the total number of classification of different target different directions image in the resulting training image; Wherein:
Figure BDA00002042658500025
For rounding downwards;
Step b1 data pre-service: adopt method of weighted mean pair Three components of the RGB of class training image are weighted and on average obtain gray level image, then gray level image are carried out down-sampling and process, and obtain the image of n * n size; The image of n * n size is carried out energy normalized process and obtain normalized image, again normalized image is converted to n 2The column vector of * 1 dimension as the row in the training data, obtains pretreated training dataset U=[U with column vector 1, U 2..., U c], U wherein iThe subdata collection of corresponding i class among the training dataset U, i=1,2 ..., c;
Step c1 trains dictionary: the FDDL software package by Fisher Discrimination Dictionary Learning for Sparse Representation issue is trained known training dataset U=[U 1, U 2..., U c], obtain dictionary D=[D 1, D 2..., D c], wherein, D iThe sub-dictionary corresponding with the i class;
Step 2 sparse coding: according to the resulting dictionary D=[D of training 1, D 2..., D c], each subimage block in the test pattern is carried out sparse coding, obtain sparse coefficient corresponding to each subimage block, concrete treatment step is as follows:
Step a2 test pattern pre-service: at first use the method for weighted mean described in the step b1 that test pattern is converted into the test gray level image, then use size to obtain subimage block along the test gray level image with interval steps b slip as the moving window of S * S; The subimage block down-sampling is processed the image that size is n * n, then carry out energy normalized and process, the image transitions after again energy normalized being processed is a n 2The column vector β of * 1 dimension represents pixel grey scale value information by the resulting subimage block of moving window with column vector β;
Step b2 sparse coding: each subimage block is passed through Optimized model
α ^ = arg min | | α | | 1 s . t . | | β - Dα | | 2 2 ≤ ϵ
Obtain the sparse coding coefficient of corresponding each subimage block
Figure BDA00002042658500031
Wherein
Figure BDA00002042658500032
Be and sub-dictionary D iCorresponding coefficient vector, ε>0th, allowable error, || || 1Be l 1Norm, || || 2Be l 2Norm;
Step c2 asks for reconstructed error: according to the sparse coding coefficient
Figure BDA00002042658500033
Calculate the reconstructed error e of each subimage block and each class i, get e=min{e iAs the reconstructed error of this subimage block, and record its corresponding classification
Figure BDA00002042658500034
Then judge whether comprise target in this subimage block according to the magnitude relationship between reconstructed error e and the predefined threshold tau: if e<τ illustrate to comprise target, otherwise, illustrate that this subimage block is background;
Step 3 object detection and recognition:
Step a3: will judge the corresponding reconstructed error e of each subimage block that comprises target among the step c2, and form a reconstructed error matrix E=(e of the same size with the test gray level image, as to represent candidate target region St) P * QWherein, e StThe value of locating in coordinate points (s, t) for the reconstruct error matrix, e st = 0 e &GreaterEqual; &tau; e e < &tau; , P * Q is the size of test pattern, s=1, and 2 ... P, t=1,2 ... Q;
Will be among the step c2 judge the corresponding classification C of each subimage block that comprises target, forms one with test that gray level image is of the same size, the classification matrix L that represents the candidate target classification=(C St) P * QC wherein StThe value of locating in coordinate points (s, t) for the classification matrix, C st = 0 e &GreaterEqual; &tau; C e < &tau; ;
Step b3: change big or small G time that slides window S * S, repeating step 2 ~ step a3G time, the G that obtains a reconstructed error matrix and G classification matrix, the span of G is 5 ~ 10; G the reconstructed error matrix that obtains formed a multiple dimensioned reconstructed error matrix M AP=(e Stg) P * Q * GWherein, e StgBe the element among the matrix M AP, its value is the corresponding e of reconstructed error matrix that the sliding window size of the g time change obtains St, P * Q * G is the size of multiple dimensioned reconstructed error matrix, g=1, and 2 ... G;
G the classification matrix that obtains consisted of a multiple dimensioned classification Matrix C LASS=(C Stg) P * Q * GWherein, C StgBe the element among the Matrix C LASS, its value is the corresponding C of classification matrix that the sliding window size of the g time change obtains StObtain a minimal reconstruction error matrix (map (s, t)) according to multiple dimensioned reconstructed error matrix M AP P * Q, wherein map (s, t) is the value that corresponding minimal reconstruction error matrix is located in coordinate points (s, t),
Figure BDA00002042658500037
Then obtain the minimum classification matrix (class (s, t)) of corresponding minimal reconstruction error matrix P * Q, wherein class (s, t) is the value that minimum classification matrix is located in coordinate points (s, t),
Figure BDA00002042658500041
Obtain Scale Matrixes scale=(scale (s, t)) according to multiple dimensioned reconstructed error matrix M AP P * Q, scale (s, t) is the value that corresponding Scale Matrixes is located in coordinate points (r, t), scale ( s , t ) = 0 e st = 0 arg min g { e stg } e st &NotEqual; 0 ;
Step c3: ask for minimal reconstruction error matrix (map (s, t)) P * QThe local neighborhood minimal value as the target response value that detects, the local neighborhood minimal value is at minimal reconstruction error matrix (map (s, t)) P * QIn corresponding coordinate be the center of target, according to the center at (class (s, t)) P * Q(scale (s, t)) P * QThe corresponding classification of target and scale size are found in the position of middle correspondence.
Described calculated with weighted average method formula is f (x, y)=0.3R (x, y)+0.59G (x, y)+0.11B (x, y), in the formula, f (x, y) gray level image that obtains for method of weighted mean is at the gray-scale value of pixel (x, y), R (x, y), G (x, y) and B (x, y) be respectively the training image of input at three component values of the RGB of pixel (x, y).
Described energy normalized computing formula is In the formula, f Norm(x, y) is the gray-scale value after f (x, y) the process energy normalized, and u and v are respectively the row and column size of gray level image.
Described l 1The computing formula of norm is
| | z | | 1 = &Sigma; k = 1 M | &xi; k |
In the formula, z is that size is the vector of M * 1, ξ kBe the element of vectorial z, k=1,2 ..., M.
Described l 2The computing formula of norm is
| | z | | 1 = &Sigma; k = 1 M | &xi; k | 2
In the formula, z is that size is the vector of M * 1, ξ kBe the element of vectorial z, k=1,2 ..., M.
Described reconstructed error e iComputing formula be
e i = | | &beta; - D i &alpha; ^ i | | 2 2 + &gamma; | | &alpha; ^ - m i | | 2 2
In the formula, γ is predefined weights, and the span of γ is 0 ~ 1, m iTo Y iIn the element of the every delegation mean vector of averaging and obtaining; Y iBe U iThe optimum code coefficient that process dictionary D sparse coding obtains.
The described anglec of rotation
Figure BDA00002042658500047
Span is 0 ° to 90 °.
Described FDDL software package parameter lambda 1Scope be 0.001 ~ 0.01, λ 2Scope be 0.01 ~ 0.1.
The span of described S is the integer between 40 ~ 90, and the span of b is the integer between 1 ~ 15.
The span of described threshold tau is 0 ~ 1.
Beneficial effect
The method of a kind of remote sensing images multi-class targets detection and Identification based on the rarefaction representation dictionary learning that the present invention proposes, at first use pretreated training data to train redundant dictionary, then use the resulting dictionary of training to carry out sparse coding to the subimage block in the test pattern, obtain its rarefaction representation coefficient, and then obtain the reconstructed error of subimage block by the rarefaction representation coefficient, and it is carried out thresholding process, determine candidate target region; Finally by crossing the accurate detection and Identification of some post-processed realizations to the remote sensing images multi-class targets.
The present invention automatically detects and identifies the target of plurality of classes the remote sensing images under complex background.Facts have proved, the method has higher detection and accuracy of identification and lower false alarm rate.
Description of drawings
Fig. 1: the basic flow sheet of the inventive method
Fig. 2: the training data in the inventive method
Fig. 3: the partial detection of the inventive method
(a) Aircraft Targets testing result (red square frame represents Aircraft Targets, and yellow square frame is false-alarm)
(b) Ship Target Detection result (white box representative Ship Target)
(c) oil depot target detection result (blue square frame represents the oil depot target)
(d) aircraft, Ship Target Detection result
(e) aircraft, oil depot target detection result
(f) naval vessel, oil depot target detection result
Embodiment
Now the invention will be further described in conjunction with embodiment, accompanying drawing:
The hardware environment that is used for implementing is: Intel Pentium 2.93GHz CPU computing machine, 2.0GB internal memory, the software environment of operation is: Matlab R2011a and Windows XP.Choose 100 width of cloth and carried out the multi-class targets test experience from the remote sensing images that Google Earth obtains, mainly comprised tertiary target: aircraft, naval vessel, oil depot, wherein, totally 200 of Aircraft Targets, totally 120 of Ship Targets, totally 420 of oil depot targets.
Implementation of the present invention is as follows:
1, training redundant dictionary: use the method training dictionary based on the rarefaction representation dictionary learning, detailed process is as follows:
(1.1) training image is processed in earlier stage: concrete processing procedure is: at first with generic target unification to a principal direction in the original image, then will unify after the direction image along 0 ° to 360 °, every 10 ° of rotations once, obtain 36 class training datas, the original image of different classes of target is all processed according to above method, finally obtained 55 class training image, i.e. c=55, wherein, aircraft is totally 36 classes, naval vessel 18 classes, oil depot 1 class;
(1.2) data pre-service: adopt method of weighted mean that three components of RGB of 55 class training images are weighted and on average obtain gray level image, then gray level image being carried out down-sampling processes, obtain the image of 15 * 15 sizes, the image of 15 * 15 sizes is carried out energy normalized to be processed and obtains normalized image, again normalized image is converted to the column vector of 255 * 1 dimensions, this column vector as the row in the training data, is obtained pretreated training dataset U=[U 1, U 2..., U c], U wherein iThe subdata collection of corresponding i class among the training dataset U, i=1,2 ..., c;
(1.3) train known training dataset U=[U by the FDDL software package of Lei Zhang issue 1, U 2..., U c], obtain dictionary D=[D 1, D 2..., D c], wherein, D iThe sub-dictionary corresponding with the i class; The software package parameter lambda 1=0.005, λ 2=0.05;
The FDDL software package of described Lei Zhang is seen paper: Meng Yang, Lei Zhang, Xiangchu Feng, David Zhang.Fisher Discrimination Dictionary Learning for Sparse Representation[C] .ICCV, 2011
2, sparse coding: the dictionary D=[D that obtains according to training 1, D 2..., D c], each subimage block in the test pattern is carried out sparse coding, obtain sparse coefficient corresponding to each subimage block, concrete treatment step is as follows:
(2.1) test pattern pre-service: at first use the method for weighted mean described in (1.1) that test pattern is converted into the test gray level image, then use size to obtain subimage block along testing gray level image take interval steps as 5 pixels slips as the moving window of S * S, the S initial value gets 90; To using resulting each subimage block of moving window, it is down sampled to size is 15 * 15 image, then carrying out energy normalized processes, image transitions after again energy normalized being processed is the column vector β of one 225 * 1 dimension, represents pixel grey scale value information by the resulting subimage block of moving window with column vector β;
(2.2) sparse coding: to each subimage block by passing through Optimized model:
&alpha; ^ = arg min | | &alpha; | | 1 s . t . | | &beta; - D&alpha; | | 2 2 &le; &epsiv;
Obtain the sparse coding coefficient vector of corresponding each subimage block
Figure BDA00002042658500062
Wherein
Figure BDA00002042658500063
Be and sub-dictionary D i
Corresponding coefficient vector, allowable error ε=0.15, || || 1Be l 1Norm, || || 2Be l 2Norm;
(2.3) ask for reconstructed error: according to the sparse coding coefficient
Figure BDA00002042658500071
Calculate the reconstructed error of subimage block picture piece and each class
Figure BDA00002042658500072
E=min{e is got in weights γ=0.5 iAs the reconstructed error of this subimage block, and record its corresponding classification Then judge whether comprise target in this subimage block according to the magnitude relationship between reconstructed error e and predefined threshold tau=0.3: if e<τ illustrate to comprise target, otherwise, illustrate that this subimage block is background;
3, object detection and recognition:
(3.1) will judge the corresponding reconstructed error e of each subimage block that comprises target in (2.3), form a reconstructed error matrix E=(e of the same size with the test gray level image, as to represent candidate target region St) P * QE wherein StThe value of locating in coordinate points (s, t) for the reconstruct error matrix, e st = 0 e &GreaterEqual; &tau; e e < &tau; , P * Q is the size of test pattern, s=1, and 2 ... P, t=1,2 ... Q; Judge the corresponding classification of each subimage block that comprises target in will (2.3), forms one with test that gray level image is of the same size, the classification matrix L that represents the candidate target classification=(C St) P * QC wherein StThe value of locating in coordinate points (s, t) for the classification matrix, C st = 0 e &GreaterEqual; &tau; C e < &tau; ;
(3.2) change the size of sliding window S * S, become S=90-10 * j, j=1,2 ... G is number of times for a change, repetition 2, step (3.1) G time altogether, the G that obtains a reconstructed error matrix and G classification matrix; G the reconstructed error matrix that obtains formed a multiple dimensioned reconstructed error matrix M AP=(e Stg) P * Q * GWherein, e StgBe the element among the matrix M AP, its value is the corresponding e of reconstructed error matrix that the sliding window size of the g time change obtains St, P * Q * G is the size of multiple dimensioned reconstructed error matrix, g=1, and 2 ... G; G the classification matrix that obtains consisted of a multiple dimensioned classification Matrix C LASS=(C Stg) P * Q * GWherein, C StgBe the element among the Matrix C LASS, its value is the corresponding C of classification matrix that the sliding window size of the g time change obtains StObtain a minimal reconstruction error matrix (map (s, t)) according to multiple dimensioned reconstructed error matrix M AP P * Q, wherein map (s, t) is the value that corresponding minimal reconstruction error matrix is located in coordinate points (s, t),
Figure BDA00002042658500076
Then obtain the minimum classification matrix (class (s, t)) of corresponding minimal reconstruction error matrix P * Q, wherein class (s, t) is the value that minimum classification matrix is located in coordinate points (s, t),
Figure BDA00002042658500077
Obtain Scale Matrixes according to multiple dimensioned reconstructed error matrix M AP
Figure BDA00002042658500078
Scale (s, t) is the value that corresponding Scale Matrixes is located in coordinate points (r, t),
scale ( s , t ) = 0 e st = 0 arg min g { e stg } e st &NotEqual; 0 ;
(3.3): ask for minimal reconstruction error matrix (map (s, t)) P * QThe local neighborhood minimal value as the target response value that detects, the local neighborhood minimal value is at minimal reconstruction error matrix (map (s, t)) P * QIn corresponding coordinate be the center of target, just can be at (class (s, t)) according to the center P * Q(scale (s, t)) P * QThe corresponding classification of target and scale size are found in the position of middle correspondence.
Described calculated with weighted average method formula is
f(x,y)=0.3R(x,y)+0.59G(x,y)+0.11B(x,y)
In the formula, the gray level image that f (x, y) obtains for method of weighted mean is at pixel (x, y) gray-scale value, R (x, y), G (x, y) and B (x, y) be respectively the training image of input at three component values of the RGB of pixel (x, y).
Described energy normalized computing formula is
f norm ( x , y ) = f ( x , y ) &Sigma; x = 1 u &Sigma; y = 1 v [ f ( x , y ) ] 2
In the formula, f Norm(x, y) is the gray-scale value after f (x, y) the process energy normalized, and u and v are respectively the row and column size of gray level image, u=15, v=15.
Described l 1The computing formula of norm is
| | z | | 1 = &Sigma; k = 1 M | &xi; k |
In the formula, z is that size is the vector of M * 1, ξ kBe the element of vectorial z, k=1,2 ..., M.
Described l 2The computing formula of norm is
| | z | | 1 = &Sigma; k = 1 M | &xi; k | 2
Described reconstructed error e iComputing formula be
e i = | | &beta; - D i &alpha; ^ i | | 2 2 + &gamma; | | &alpha; ^ - m i | | 2 2
In the formula, γ is predefined weights, γ=0.5, m iTo Y iIn the element of the every delegation mean vector of averaging and obtaining; Y iBe U iThe optimum code coefficient that process dictionary D sparse coding obtains;
Select correct verification and measurement ratio and false alarm rate that validity of the present invention is assessed.Wherein, the ratio of the target number that correct verification and measurement ratio is defined as correct detection and total target number, false alarm rate is defined as false-alarm number and the correct target number that detects and the ratio of false-alarm number sum.Simultaneously, the testing result of gained of the present invention and multi-class targets detection algorithm based on BoW are contrasted, comparing result is as shown in table 1.Correct verification and measurement ratio and false alarm rate have all shown the validity of the inventive method.
Table 1 evaluation
Figure BDA00002042658500092

Claims (10)

1. remote sensing images multi-class targets detection and Identification method, its design procedure is as follows:
Step 1: use the method training dictionary based on the rarefaction representation dictionary learning, concrete steps are as follows:
Step a1 training image is processed in earlier stage: at first with generic target unification to a principal direction in the original image, then will unify after the direction image along 0 ° to 360 °, according to step-length Rotate to be
Figure FDA00002042658400012
The image of individual different directions; The original image of different classes of target is all processed according to the method described above, obtained
Figure FDA00002042658400013
The class training image, wherein p is the different classes of number of targets that will detect,
Figure FDA00002042658400014
Be the anglec of rotation, c is the total number of classification of different target different directions image in the resulting training image; Wherein: For rounding downwards;
Step b1 data pre-service: adopt method of weighted mean pair
Figure FDA00002042658400016
Three components of the RGB of class training image are weighted and on average obtain gray level image, then gray level image are carried out down-sampling and process, and obtain the image of n * n size; The image of n * n size is carried out energy normalized process and obtain normalized image, again normalized image is converted to n 2The column vector of * 1 dimension as the row in the training data, obtains pretreated training dataset U=[U with column vector 1, U 2..., U c], U wherein 1The subdata collection of corresponding i class among the training dataset U, i=1,2 ..., c;
Step c1 trains dictionary: the FDDL software package by Fisher Discrimination Dictionary Learning for Sparse Representation issue is trained known training dataset U=[U 1, U 2.., U c], obtain dictionary D=[D 1, D 2..., D c], wherein, D iThe sub-dictionary corresponding with the i class;
Step 2 sparse coding: according to the resulting dictionary D=[D of training 1, D 2..., D c], each subimage block in the test pattern is carried out sparse coding, obtain sparse coefficient corresponding to each subimage block, concrete treatment step is as follows:
Step a2 test pattern pre-service: at first use the method for weighted mean described in the step b1 that test pattern is converted into the test gray level image, then use size to obtain subimage block along the test gray level image with interval steps b slip as the moving window of S * S; The subimage block down-sampling is processed the image that size is n * n, then carry out energy normalized and process, the image transitions after again energy normalized being processed is a n 2The column vector β of * 1 dimension represents pixel grey scale value information by the resulting subimage block of moving window with column vector β;
Step b2 sparse coding: each subimage block is passed through Optimized model
&alpha; ^ = arg min | | &alpha; | | 1 s . t . | | &beta; - D&alpha; | | 2 2 &le; &epsiv;
Obtain the sparse coding coefficient of corresponding each subimage block Wherein
Figure FDA00002042658400023
Be and sub-dictionary D iCorresponding coefficient vector, ε>0th, allowable error, || || 1Be l 1Norm, || || 2Be l 2Norm;
Step c2 asks for reconstructed error: according to the sparse coding coefficient
Figure FDA00002042658400024
Calculate the reconstructed error e of each subimage block and each class i, get e=min{e iAs the reconstructed error of this subimage block, and record its corresponding classification
Figure FDA00002042658400025
Then judge whether comprise target in this subimage block according to the magnitude relationship between reconstructed error e and the predefined threshold tau: if e<τ illustrate to comprise target, otherwise, illustrate that this subimage block is background;
Step 3 object detection and recognition:
Step a3: will judge the corresponding reconstructed error e of each subimage block that comprises target among the step c2, and form a reconstructed error matrix E=(e of the same size with the test gray level image, as to represent candidate target region St) P * QWherein, e StThe value of locating in coordinate points (s, t) for the reconstruct error matrix, e st = 0 e &GreaterEqual; &tau; e e < &tau; , P * Q is the size of test pattern, s=1, and 2 ... P, t=1,2 ... Q;
Will be among the step c2 judge the corresponding classification C of each subimage block that comprises target, forms one with test that gray level image is of the same size, the classification matrix L that represents the candidate target classification=(C St) P * QC wherein StThe value of locating in coordinate points (s, t) for the classification matrix, C st = 0 e &GreaterEqual; &tau; C e < &tau; ;
Step b3: change big or small G time that slides window S * S, repeating step 2 ~ step a3G time, the G that obtains a reconstructed error matrix and G classification matrix, the span of G is 5 ~ 10; G the reconstructed error matrix that obtains formed a multiple dimensioned reconstructed error matrix M AP=(e Stg) P * Q * GWherein, e StgBe the element among the matrix M AP, its value is the corresponding e of reconstructed error matrix that the sliding window size of the g time change obtains St, P * Q * G is the size of multiple dimensioned reconstructed error matrix, g=1, and 2 ... G;
G the classification matrix that obtains consisted of a multiple dimensioned classification Matrix C LASS=(C Stg) P * Q * GWherein, C StgBe the element among the Matrix C LASS, its value is the corresponding C of classification matrix that the sliding window size of the g time change obtains StObtain a minimal reconstruction error matrix (map (s, t)) according to multiple dimensioned reconstructed error matrix M AP P * Q, wherein map (s, t) is the value that corresponding minimal reconstruction error matrix is located in coordinate points (s, t),
Figure FDA00002042658400031
Then obtain the minimum classification matrix (class (s, t)) of corresponding minimal reconstruction error matrix P * Q, wherein class (s, t) is the value that minimum classification matrix is located in coordinate points (s, t),
Figure FDA00002042658400032
Obtain Scale Matrixes scale=(scale (s, t)) according to multiple dimensioned reconstructed error matrix M AP P * Q, scale (s, t) is the value that corresponding Scale Matrixes is located in coordinate points (r, t), scale ( s , t ) = 0 e st = 0 arg min g { e stg } e st &NotEqual; 0 ;
Step c3: ask for minimal reconstruction error matrix (map (s, t)) P * QThe local neighborhood minimal value as the target response value that detects, the local neighborhood minimal value is at minimal reconstruction error matrix (map (s, t)) P * QIn corresponding coordinate be the center of target, according to the center at (class (s, t)) P * Q(scale (s, t)) P * QThe corresponding classification of target and scale size are found in the position of middle correspondence.
2. described remote sensing images multi-class targets detection and Identification method according to claim 1, it is characterized in that: described calculated with weighted average method formula is f (x, y)=0.3R (x, y)+0.59G (x, y)+0.11B (x, y), in the formula, the gray level image that f (x, y) obtains for method of weighted mean is at pixel (x, y) gray-scale value, R (x, y), G (x, y) and B (x, y) be respectively the training image of input at three component values of the RGB of pixel (x, y).
3. described remote sensing images multi-class targets detection and Identification method according to claim 1, it is characterized in that: described energy normalized computing formula is
Figure FDA00002042658400034
In the formula, f Norm(x, y) is the gray-scale value after f (x, y) the process energy normalized, and u and v are respectively the row and column size of gray level image.
4. described remote sensing images multi-class targets detection and Identification method according to claim 1 is characterized in that: described l 1The computing formula of norm is
Figure FDA00002042658400035
In the formula, z is that size is the vector of M * 1, ξ kBe the element of vectorial z,
k=1,2,…,M。
5. described remote sensing images multi-class targets detection and Identification method according to claim 1 is characterized in that: described l 2The computing formula of norm is
Figure FDA00002042658400041
In the formula, z is that size is the vector of M * 1, ξ kBe the element of vectorial z,
k=1,2,…,M。
6. described remote sensing images multi-class targets detection and Identification method according to claim 1 is characterized in that: described reconstructed error e iComputing formula be
Figure FDA00002042658400042
In the formula, γ is predefined weights, the value of γ
Scope is 0 ~ 1, m iTo Y iIn the element of the every delegation mean vector of averaging and obtaining; Y iBe U iThe optimum code coefficient that process dictionary D sparse coding obtains.
7. described remote sensing images multi-class targets detection and Identification method according to claim 1 is characterized in that: the described anglec of rotation
Figure FDA00002042658400043
Span is 0 ° to 90 °.
8. described remote sensing images multi-class targets detection and Identification method according to claim 1 is characterized in that: described FDDL software package parameter lambda 1Scope be 0.001 ~ 0.01, λ 2Scope be 0.01 ~ 0.1.
9. described remote sensing images multi-class targets detection and Identification method according to claim 1, it is characterized in that: the span of described S is the integer between 40 ~ 90, the span of b is the integer between 1 ~ 15.
10. described remote sensing images multi-class targets detection and Identification method according to claim 1, it is characterized in that: the span of described threshold tau is 0 ~ 1.
CN201210300645.8A 2012-08-22 2012-08-22 Method for detecting and identifying a plurality of types of objects in remote sensing image Active CN102867195B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201210300645.8A CN102867195B (en) 2012-08-22 2012-08-22 Method for detecting and identifying a plurality of types of objects in remote sensing image

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201210300645.8A CN102867195B (en) 2012-08-22 2012-08-22 Method for detecting and identifying a plurality of types of objects in remote sensing image

Publications (2)

Publication Number Publication Date
CN102867195A true CN102867195A (en) 2013-01-09
CN102867195B CN102867195B (en) 2014-11-26

Family

ID=47446059

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201210300645.8A Active CN102867195B (en) 2012-08-22 2012-08-22 Method for detecting and identifying a plurality of types of objects in remote sensing image

Country Status (1)

Country Link
CN (1) CN102867195B (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103258210A (en) * 2013-05-27 2013-08-21 中山大学 High-definition image classification method based on dictionary learning
CN103632164A (en) * 2013-11-25 2014-03-12 西北工业大学 Reeling steel state identifying method based on KNN reeling steel picture data of optimized KAP sample
CN104517121A (en) * 2014-12-10 2015-04-15 中国科学院遥感与数字地球研究所 Spatial big data dictionary learning method based on particle swarm optimization
CN105740422A (en) * 2016-01-29 2016-07-06 北京大学 Pedestrian retrieval method and apparatus
CN106067041A (en) * 2016-06-03 2016-11-02 河海大学 A kind of multi-target detection method of based on rarefaction representation of improvement
CN107451595A (en) * 2017-08-04 2017-12-08 河海大学 Infrared image salient region detection method based on hybrid algorithm
CN109190457A (en) * 2018-07-19 2019-01-11 北京市遥感信息研究所 A kind of oil depot complex target rapid detection method based on large format remote sensing images
CN109946076A (en) * 2019-01-25 2019-06-28 西安交通大学 A kind of planet wheel bearing fault identification method of weighted multiscale dictionary learning frame
CN110189328A (en) * 2019-06-11 2019-08-30 北华航天工业学院 A kind of Remote sensing image processing system and its processing method

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100124377A1 (en) * 2008-11-19 2010-05-20 Nec Laboratories America, Inc. Linear spatial pyramid matching using sparse coding
CN102129573A (en) * 2011-03-10 2011-07-20 西安电子科技大学 SAR (Synthetic Aperture Radar) image segmentation method based on dictionary learning and sparse representation
CN102324047A (en) * 2011-09-05 2012-01-18 西安电子科技大学 High spectrum image atural object recognition methods based on sparse nuclear coding SKR

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100124377A1 (en) * 2008-11-19 2010-05-20 Nec Laboratories America, Inc. Linear spatial pyramid matching using sparse coding
CN102129573A (en) * 2011-03-10 2011-07-20 西安电子科技大学 SAR (Synthetic Aperture Radar) image segmentation method based on dictionary learning and sparse representation
CN102324047A (en) * 2011-09-05 2012-01-18 西安电子科技大学 High spectrum image atural object recognition methods based on sparse nuclear coding SKR

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
梁天一 等: "基于稀疏编码的图像语义分类器模型", 《华东理工大学学报(自然科学报)》, vol. 33, no. 6, 31 December 2007 (2007-12-31), pages 827 - 892 *

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103258210A (en) * 2013-05-27 2013-08-21 中山大学 High-definition image classification method based on dictionary learning
CN103258210B (en) * 2013-05-27 2016-09-14 中山大学 A kind of high-definition image classification method based on dictionary learning
CN103632164B (en) * 2013-11-25 2017-03-01 西北工业大学 The volume firm state classification recognition methodss of the KNN coil image data based on KAP sample optimization
CN103632164A (en) * 2013-11-25 2014-03-12 西北工业大学 Reeling steel state identifying method based on KNN reeling steel picture data of optimized KAP sample
CN104517121A (en) * 2014-12-10 2015-04-15 中国科学院遥感与数字地球研究所 Spatial big data dictionary learning method based on particle swarm optimization
CN105740422A (en) * 2016-01-29 2016-07-06 北京大学 Pedestrian retrieval method and apparatus
CN106067041A (en) * 2016-06-03 2016-11-02 河海大学 A kind of multi-target detection method of based on rarefaction representation of improvement
CN106067041B (en) * 2016-06-03 2019-05-31 河海大学 A kind of improved multi-target detection method based on rarefaction representation
CN107451595A (en) * 2017-08-04 2017-12-08 河海大学 Infrared image salient region detection method based on hybrid algorithm
CN109190457A (en) * 2018-07-19 2019-01-11 北京市遥感信息研究所 A kind of oil depot complex target rapid detection method based on large format remote sensing images
CN109190457B (en) * 2018-07-19 2021-12-03 北京市遥感信息研究所 Oil depot cluster target rapid detection method based on large-format remote sensing image
CN109946076A (en) * 2019-01-25 2019-06-28 西安交通大学 A kind of planet wheel bearing fault identification method of weighted multiscale dictionary learning frame
CN109946076B (en) * 2019-01-25 2020-04-28 西安交通大学 Planetary wheel bearing fault identification method of weighted multi-scale dictionary learning framework
CN110189328A (en) * 2019-06-11 2019-08-30 北华航天工业学院 A kind of Remote sensing image processing system and its processing method
CN110189328B (en) * 2019-06-11 2021-02-23 北华航天工业学院 Satellite remote sensing image processing system and processing method thereof

Also Published As

Publication number Publication date
CN102867195B (en) 2014-11-26

Similar Documents

Publication Publication Date Title
CN102867195B (en) Method for detecting and identifying a plurality of types of objects in remote sensing image
CN110348319B (en) Face anti-counterfeiting method based on face depth information and edge image fusion
CN107346436B (en) Visual saliency detection method fusing image classification
CN105608456B (en) A kind of multi-direction Method for text detection based on full convolutional network
Sirmacek et al. Urban area detection using local feature points and spatial voting
Yuan et al. Large-scale solar panel mapping from aerial images using deep convolutional networks
CN103390164B (en) Method for checking object based on depth image and its realize device
CN102542295B (en) Method for detecting landslip from remotely sensed image by adopting image classification technology
CN102324047B (en) Hyper-spectral image ground object recognition method based on sparse kernel representation (SKR)
CN102819740B (en) A kind of Single Infrared Image Frame Dim targets detection and localization method
CN110555475A (en) few-sample target detection method based on semantic information fusion
CN106023220A (en) Vehicle exterior part image segmentation method based on deep learning
CN101930549B (en) Second generation curvelet transform-based static human detection method
CN104008370A (en) Video face identifying method
CN103942540A (en) False fingerprint detection algorithm based on curvelet texture analysis and SVM-KNN classification
CN104182985A (en) Remote sensing image change detection method
CN106845341A (en) A kind of unlicensed vehicle identification method based on virtual number plate
CN102663724B (en) Method for detecting remote sensing image change based on adaptive difference images
CN103745233B (en) The hyperspectral image classification method migrated based on spatial information
CN103646256A (en) Image characteristic sparse reconstruction based image classification method
CN104809471A (en) Hyperspectral image residual error fusion classification method based on space spectrum information
CN111275070B (en) Signature verification method and device based on local feature matching
CN106446854A (en) High-resolution optical remote sensing image target detection method based on rotation invariant HOG feature
CN108073940A (en) A kind of method of 3D object instance object detections in unstructured moving grids
CN103455826A (en) Efficient matching kernel body detection method based on rapid robustness characteristics

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant