CN102867195A - Method for detecting and identifying a plurality of types of objects in remote sensing image - Google Patents
Method for detecting and identifying a plurality of types of objects in remote sensing image Download PDFInfo
- Publication number
- CN102867195A CN102867195A CN2012103006458A CN201210300645A CN102867195A CN 102867195 A CN102867195 A CN 102867195A CN 2012103006458 A CN2012103006458 A CN 2012103006458A CN 201210300645 A CN201210300645 A CN 201210300645A CN 102867195 A CN102867195 A CN 102867195A
- Authority
- CN
- China
- Prior art keywords
- image
- matrix
- class
- target
- remote sensing
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Landscapes
- Image Analysis (AREA)
Abstract
The invention relates to a method for detecting and identifying a plurality of types of objects in a remote sensing image based on sparse representation dictionary learning. The method is technically characterized by comprising the following steps of: training preprocessed training data to form a dictionary by using a sparse representation dictionary training method; performing sparse coding on the dictionary obtained by training by using an image sub-block in a test image, calculating a sparse representation coefficient to further obtain a reconstruction error of the image sub-block, and determining a candidate target area by thresholding the reconstruction error; and accurately detecting and identifying a plurality of types of objects in the remote sensing image by using post processing. By the method, a plurality of types of objects in the remote sensing image under a complex background can be detected and identified. The method is high in detection and identification accuracy and low in false alarm rate.
Description
Technical field
The present invention relates to a kind of remote sensing images multi-class targets detection and Identification method, can be applied to the polytype Target detection and identification under the complex background remote sensing images.
Background technology
An application as Remote Sensing Image Processing Technology, object detection and recognition under the complex background remote sensing images is a gordian technique in the fields such as military surveillance and precision strike, also be study hotspot and the difficult point in this field always, there is important military and civilian to be worth, is subject to people and more and more pays close attention to.
Remote Sensing Target detects the two kinds of methods that mainly contain at present.A kind of is that some shape, the geometric properties that has by the detection target in remote sensing images solves target detection problems, but because the remote sensing images background is complicated, exist a large amount of shape, geometric properties similar with target, only rely on these features to detect target and a large amount of undetected, flase drop can occur.Another kind is based on the thought of classification, wherein modal is Bag-of-Words(BoW) sorting technique, the method at first is that image is extracted SIFT feature and cluster, with cluster centre as the one group of standard base (image-region of standard) in the image space, then can organize the standard base with this image is carried out vector representation, at last resulting vector is classified and thresholding by use svm classifier device, obtain testing result; But the BoW method although the SIFT feature of extracting has yardstick and rotational invariance, has only been utilized the statistical nature of characteristic area, and has been ignored the spatial information of characteristic area, therefore uses the method verification and measurement ratio of BoW low, and false alarm rate is high; And another sorting technique Linear Spatial Pyramid Matching Using Sparse Coding(ScSPM) although considered the spatial information of characteristic area, resulting vectorial dimension for classification is too high, operand is excessive.In addition, most also only limits to simple target is detected based on the object detection method of classification, can not detect and identify a plurality of targets simultaneously.
Summary of the invention
The technical matters that solves
For fear of the deficiencies in the prior art part, the present invention proposes a kind of method of the remote sensing images multi-class targets detection and Identification based on the rarefaction representation dictionary learning.This method can automatically detect from the remote sensing images of complex background and identify dissimilar target, has higher accuracy of detection and lower false alarm rate.
Technical scheme
A kind of remote sensing images multi-class targets detection and Identification method is characterized in that step is as follows:
Step 1: use the method training dictionary based on the rarefaction representation dictionary learning, concrete steps are as follows:
Step a1 training image is processed in earlier stage: at first with generic target unification to a principal direction in the original image, then will unify after the direction image along 0 ° to 360 °, according to step-length
Rotate to be
The image of individual different directions; The original image of different classes of target is all processed according to the method described above, obtained
The class training image, wherein p is the different classes of number of targets that will detect,
Be the anglec of rotation, c is the total number of classification of different target different directions image in the resulting training image; Wherein:
For rounding downwards;
Step b1 data pre-service: adopt method of weighted mean pair
Three components of the RGB of class training image are weighted and on average obtain gray level image, then gray level image are carried out down-sampling and process, and obtain the image of n * n size; The image of n * n size is carried out energy normalized process and obtain normalized image, again normalized image is converted to n
2The column vector of * 1 dimension as the row in the training data, obtains pretreated training dataset U=[U with column vector
1, U
2..., U
c], U wherein
iThe subdata collection of corresponding i class among the training dataset U, i=1,2 ..., c;
Step c1 trains dictionary: the FDDL software package by Fisher Discrimination Dictionary Learning for Sparse Representation issue is trained known training dataset U=[U
1, U
2..., U
c], obtain dictionary D=[D
1, D
2..., D
c], wherein, D
iThe sub-dictionary corresponding with the i class;
Step 2 sparse coding: according to the resulting dictionary D=[D of training
1, D
2..., D
c], each subimage block in the test pattern is carried out sparse coding, obtain sparse coefficient corresponding to each subimage block, concrete treatment step is as follows:
Step a2 test pattern pre-service: at first use the method for weighted mean described in the step b1 that test pattern is converted into the test gray level image, then use size to obtain subimage block along the test gray level image with interval steps b slip as the moving window of S * S; The subimage block down-sampling is processed the image that size is n * n, then carry out energy normalized and process, the image transitions after again energy normalized being processed is a n
2The column vector β of * 1 dimension represents pixel grey scale value information by the resulting subimage block of moving window with column vector β;
Step b2 sparse coding: each subimage block is passed through Optimized model
Obtain the sparse coding coefficient of corresponding each subimage block
Wherein
Be and sub-dictionary D
iCorresponding coefficient vector, ε>0th, allowable error, || ||
1Be l
1Norm, || ||
2Be l
2Norm;
Step c2 asks for reconstructed error: according to the sparse coding coefficient
Calculate the reconstructed error e of each subimage block and each class
i, get e=min{e
iAs the reconstructed error of this subimage block, and record its corresponding classification
Then judge whether comprise target in this subimage block according to the magnitude relationship between reconstructed error e and the predefined threshold tau: if e<τ illustrate to comprise target, otherwise, illustrate that this subimage block is background;
Step 3 object detection and recognition:
Step a3: will judge the corresponding reconstructed error e of each subimage block that comprises target among the step c2, and form a reconstructed error matrix E=(e of the same size with the test gray level image, as to represent candidate target region
St)
P * QWherein, e
StThe value of locating in coordinate points (s, t) for the reconstruct error matrix,
P * Q is the size of test pattern, s=1, and 2 ... P, t=1,2 ... Q;
Will be among the step c2 judge the corresponding classification C of each subimage block that comprises target, forms one with test that gray level image is of the same size, the classification matrix L that represents the candidate target classification=(C
St)
P * QC wherein
StThe value of locating in coordinate points (s, t) for the classification matrix,
Step b3: change big or small G time that slides window S * S, repeating step 2 ~ step a3G time, the G that obtains a reconstructed error matrix and G classification matrix, the span of G is 5 ~ 10; G the reconstructed error matrix that obtains formed a multiple dimensioned reconstructed error matrix M AP=(e
Stg)
P * Q * GWherein, e
StgBe the element among the matrix M AP, its value is the corresponding e of reconstructed error matrix that the sliding window size of the g time change obtains
St, P * Q * G is the size of multiple dimensioned reconstructed error matrix, g=1, and 2 ... G;
G the classification matrix that obtains consisted of a multiple dimensioned classification Matrix C LASS=(C
Stg)
P * Q * GWherein, C
StgBe the element among the Matrix C LASS, its value is the corresponding C of classification matrix that the sliding window size of the g time change obtains
StObtain a minimal reconstruction error matrix (map (s, t)) according to multiple dimensioned reconstructed error matrix M AP
P * Q, wherein map (s, t) is the value that corresponding minimal reconstruction error matrix is located in coordinate points (s, t),
Then obtain the minimum classification matrix (class (s, t)) of corresponding minimal reconstruction error matrix
P * Q, wherein class (s, t) is the value that minimum classification matrix is located in coordinate points (s, t),
Obtain Scale Matrixes scale=(scale (s, t)) according to multiple dimensioned reconstructed error matrix M AP
P * Q, scale (s, t) is the value that corresponding Scale Matrixes is located in coordinate points (r, t),
Step c3: ask for minimal reconstruction error matrix (map (s, t))
P * QThe local neighborhood minimal value as the target response value that detects, the local neighborhood minimal value is at minimal reconstruction error matrix (map (s, t))
P * QIn corresponding coordinate be the center of target, according to the center at (class (s, t))
P * Q(scale (s, t))
P * QThe corresponding classification of target and scale size are found in the position of middle correspondence.
Described calculated with weighted average method formula is f (x, y)=0.3R (x, y)+0.59G (x, y)+0.11B (x, y), in the formula, f (x, y) gray level image that obtains for method of weighted mean is at the gray-scale value of pixel (x, y), R (x, y), G (x, y) and B (x, y) be respectively the training image of input at three component values of the RGB of pixel (x, y).
Described energy normalized computing formula is
In the formula, f
Norm(x, y) is the gray-scale value after f (x, y) the process energy normalized, and u and v are respectively the row and column size of gray level image.
Described l
1The computing formula of norm is
In the formula, z is that size is the vector of M * 1, ξ
kBe the element of vectorial z, k=1,2 ..., M.
Described l
2The computing formula of norm is
In the formula, z is that size is the vector of M * 1, ξ
kBe the element of vectorial z, k=1,2 ..., M.
Described reconstructed error e
iComputing formula be
In the formula, γ is predefined weights, and the span of γ is 0 ~ 1, m
iTo Y
iIn the element of the every delegation mean vector of averaging and obtaining; Y
iBe U
iThe optimum code coefficient that process dictionary D sparse coding obtains.
Described FDDL software package parameter lambda
1Scope be 0.001 ~ 0.01, λ
2Scope be 0.01 ~ 0.1.
The span of described S is the integer between 40 ~ 90, and the span of b is the integer between 1 ~ 15.
The span of described threshold tau is 0 ~ 1.
Beneficial effect
The method of a kind of remote sensing images multi-class targets detection and Identification based on the rarefaction representation dictionary learning that the present invention proposes, at first use pretreated training data to train redundant dictionary, then use the resulting dictionary of training to carry out sparse coding to the subimage block in the test pattern, obtain its rarefaction representation coefficient, and then obtain the reconstructed error of subimage block by the rarefaction representation coefficient, and it is carried out thresholding process, determine candidate target region; Finally by crossing the accurate detection and Identification of some post-processed realizations to the remote sensing images multi-class targets.
The present invention automatically detects and identifies the target of plurality of classes the remote sensing images under complex background.Facts have proved, the method has higher detection and accuracy of identification and lower false alarm rate.
Description of drawings
Fig. 1: the basic flow sheet of the inventive method
Fig. 2: the training data in the inventive method
Fig. 3: the partial detection of the inventive method
(a) Aircraft Targets testing result (red square frame represents Aircraft Targets, and yellow square frame is false-alarm)
(b) Ship Target Detection result (white box representative Ship Target)
(c) oil depot target detection result (blue square frame represents the oil depot target)
(d) aircraft, Ship Target Detection result
(e) aircraft, oil depot target detection result
(f) naval vessel, oil depot target detection result
Embodiment
Now the invention will be further described in conjunction with embodiment, accompanying drawing:
The hardware environment that is used for implementing is: Intel Pentium 2.93GHz CPU computing machine, 2.0GB internal memory, the software environment of operation is: Matlab R2011a and Windows XP.Choose 100 width of cloth and carried out the multi-class targets test experience from the remote sensing images that Google Earth obtains, mainly comprised tertiary target: aircraft, naval vessel, oil depot, wherein, totally 200 of Aircraft Targets, totally 120 of Ship Targets, totally 420 of oil depot targets.
Implementation of the present invention is as follows:
1, training redundant dictionary: use the method training dictionary based on the rarefaction representation dictionary learning, detailed process is as follows:
(1.1) training image is processed in earlier stage: concrete processing procedure is: at first with generic target unification to a principal direction in the original image, then will unify after the direction image along 0 ° to 360 °, every 10 ° of rotations once, obtain 36 class training datas, the original image of different classes of target is all processed according to above method, finally obtained 55 class training image, i.e. c=55, wherein, aircraft is totally 36 classes, naval vessel 18 classes, oil depot 1 class;
(1.2) data pre-service: adopt method of weighted mean that three components of RGB of 55 class training images are weighted and on average obtain gray level image, then gray level image being carried out down-sampling processes, obtain the image of 15 * 15 sizes, the image of 15 * 15 sizes is carried out energy normalized to be processed and obtains normalized image, again normalized image is converted to the column vector of 255 * 1 dimensions, this column vector as the row in the training data, is obtained pretreated training dataset U=[U
1, U
2..., U
c], U wherein
iThe subdata collection of corresponding i class among the training dataset U, i=1,2 ..., c;
(1.3) train known training dataset U=[U by the FDDL software package of Lei Zhang issue
1, U
2..., U
c], obtain dictionary D=[D
1, D
2..., D
c], wherein, D
iThe sub-dictionary corresponding with the i class; The software package parameter lambda
1=0.005, λ
2=0.05;
The FDDL software package of described Lei Zhang is seen paper: Meng Yang, Lei Zhang, Xiangchu Feng, David Zhang.Fisher Discrimination Dictionary Learning for Sparse Representation[C] .ICCV, 2011
2, sparse coding: the dictionary D=[D that obtains according to training
1, D
2..., D
c], each subimage block in the test pattern is carried out sparse coding, obtain sparse coefficient corresponding to each subimage block, concrete treatment step is as follows:
(2.1) test pattern pre-service: at first use the method for weighted mean described in (1.1) that test pattern is converted into the test gray level image, then use size to obtain subimage block along testing gray level image take interval steps as 5 pixels slips as the moving window of S * S, the S initial value gets 90; To using resulting each subimage block of moving window, it is down sampled to size is 15 * 15 image, then carrying out energy normalized processes, image transitions after again energy normalized being processed is the column vector β of one 225 * 1 dimension, represents pixel grey scale value information by the resulting subimage block of moving window with column vector β;
(2.2) sparse coding: to each subimage block by passing through Optimized model:
Obtain the sparse coding coefficient vector of corresponding each subimage block
Wherein
Be and sub-dictionary D
i
Corresponding coefficient vector, allowable error ε=0.15, || ||
1Be l
1Norm, || ||
2Be l
2Norm;
(2.3) ask for reconstructed error: according to the sparse coding coefficient
Calculate the reconstructed error of subimage block picture piece and each class
E=min{e is got in weights γ=0.5
iAs the reconstructed error of this subimage block, and record its corresponding classification
Then judge whether comprise target in this subimage block according to the magnitude relationship between reconstructed error e and predefined threshold tau=0.3: if e<τ illustrate to comprise target, otherwise, illustrate that this subimage block is background;
3, object detection and recognition:
(3.1) will judge the corresponding reconstructed error e of each subimage block that comprises target in (2.3), form a reconstructed error matrix E=(e of the same size with the test gray level image, as to represent candidate target region
St)
P * QE wherein
StThe value of locating in coordinate points (s, t) for the reconstruct error matrix,
P * Q is the size of test pattern, s=1, and 2 ... P, t=1,2 ... Q; Judge the corresponding classification of each subimage block that comprises target in will (2.3), forms one with test that gray level image is of the same size, the classification matrix L that represents the candidate target classification=(C
St)
P * QC wherein
StThe value of locating in coordinate points (s, t) for the classification matrix,
(3.2) change the size of sliding window S * S, become S=90-10 * j, j=1,2 ... G is number of times for a change, repetition 2, step (3.1) G time altogether, the G that obtains a reconstructed error matrix and G classification matrix; G the reconstructed error matrix that obtains formed a multiple dimensioned reconstructed error matrix M AP=(e
Stg)
P * Q * GWherein, e
StgBe the element among the matrix M AP, its value is the corresponding e of reconstructed error matrix that the sliding window size of the g time change obtains
St, P * Q * G is the size of multiple dimensioned reconstructed error matrix, g=1, and 2 ... G; G the classification matrix that obtains consisted of a multiple dimensioned classification Matrix C LASS=(C
Stg)
P * Q * GWherein, C
StgBe the element among the Matrix C LASS, its value is the corresponding C of classification matrix that the sliding window size of the g time change obtains
StObtain a minimal reconstruction error matrix (map (s, t)) according to multiple dimensioned reconstructed error matrix M AP
P * Q, wherein map (s, t) is the value that corresponding minimal reconstruction error matrix is located in coordinate points (s, t),
Then obtain the minimum classification matrix (class (s, t)) of corresponding minimal reconstruction error matrix
P * Q, wherein class (s, t) is the value that minimum classification matrix is located in coordinate points (s, t),
Obtain Scale Matrixes according to multiple dimensioned reconstructed error matrix M AP
Scale (s, t) is the value that corresponding Scale Matrixes is located in coordinate points (r, t),
(3.3): ask for minimal reconstruction error matrix (map (s, t))
P * QThe local neighborhood minimal value as the target response value that detects, the local neighborhood minimal value is at minimal reconstruction error matrix (map (s, t))
P * QIn corresponding coordinate be the center of target, just can be at (class (s, t)) according to the center
P * Q(scale (s, t))
P * QThe corresponding classification of target and scale size are found in the position of middle correspondence.
Described calculated with weighted average method formula is
f(x,y)=0.3R(x,y)+0.59G(x,y)+0.11B(x,y)
In the formula, the gray level image that f (x, y) obtains for method of weighted mean is at pixel (x, y) gray-scale value, R (x, y), G (x, y) and B (x, y) be respectively the training image of input at three component values of the RGB of pixel (x, y).
Described energy normalized computing formula is
In the formula, f
Norm(x, y) is the gray-scale value after f (x, y) the process energy normalized, and u and v are respectively the row and column size of gray level image, u=15, v=15.
Described l
1The computing formula of norm is
In the formula, z is that size is the vector of M * 1, ξ
kBe the element of vectorial z, k=1,2 ..., M.
Described l
2The computing formula of norm is
Described reconstructed error e
iComputing formula be
In the formula, γ is predefined weights, γ=0.5, m
iTo Y
iIn the element of the every delegation mean vector of averaging and obtaining; Y
iBe U
iThe optimum code coefficient that process dictionary D sparse coding obtains;
Select correct verification and measurement ratio and false alarm rate that validity of the present invention is assessed.Wherein, the ratio of the target number that correct verification and measurement ratio is defined as correct detection and total target number, false alarm rate is defined as false-alarm number and the correct target number that detects and the ratio of false-alarm number sum.Simultaneously, the testing result of gained of the present invention and multi-class targets detection algorithm based on BoW are contrasted, comparing result is as shown in table 1.Correct verification and measurement ratio and false alarm rate have all shown the validity of the inventive method.
Table 1 evaluation
Claims (10)
1. remote sensing images multi-class targets detection and Identification method, its design procedure is as follows:
Step 1: use the method training dictionary based on the rarefaction representation dictionary learning, concrete steps are as follows:
Step a1 training image is processed in earlier stage: at first with generic target unification to a principal direction in the original image, then will unify after the direction image along 0 ° to 360 °, according to step-length
Rotate to be
The image of individual different directions; The original image of different classes of target is all processed according to the method described above, obtained
The class training image, wherein p is the different classes of number of targets that will detect,
Be the anglec of rotation, c is the total number of classification of different target different directions image in the resulting training image; Wherein:
For rounding downwards;
Step b1 data pre-service: adopt method of weighted mean pair
Three components of the RGB of class training image are weighted and on average obtain gray level image, then gray level image are carried out down-sampling and process, and obtain the image of n * n size; The image of n * n size is carried out energy normalized process and obtain normalized image, again normalized image is converted to n
2The column vector of * 1 dimension as the row in the training data, obtains pretreated training dataset U=[U with column vector
1, U
2..., U
c], U wherein
1The subdata collection of corresponding i class among the training dataset U, i=1,2 ..., c;
Step c1 trains dictionary: the FDDL software package by Fisher Discrimination Dictionary Learning for Sparse Representation issue is trained known training dataset U=[U
1, U
2.., U
c], obtain dictionary D=[D
1, D
2..., D
c], wherein, D
iThe sub-dictionary corresponding with the i class;
Step 2 sparse coding: according to the resulting dictionary D=[D of training
1, D
2..., D
c], each subimage block in the test pattern is carried out sparse coding, obtain sparse coefficient corresponding to each subimage block, concrete treatment step is as follows:
Step a2 test pattern pre-service: at first use the method for weighted mean described in the step b1 that test pattern is converted into the test gray level image, then use size to obtain subimage block along the test gray level image with interval steps b slip as the moving window of S * S; The subimage block down-sampling is processed the image that size is n * n, then carry out energy normalized and process, the image transitions after again energy normalized being processed is a n
2The column vector β of * 1 dimension represents pixel grey scale value information by the resulting subimage block of moving window with column vector β;
Step b2 sparse coding: each subimage block is passed through Optimized model
Obtain the sparse coding coefficient of corresponding each subimage block
Wherein
Be and sub-dictionary D
iCorresponding coefficient vector, ε>0th, allowable error, || ||
1Be l
1Norm, || ||
2Be l
2Norm;
Step c2 asks for reconstructed error: according to the sparse coding coefficient
Calculate the reconstructed error e of each subimage block and each class
i, get e=min{e
iAs the reconstructed error of this subimage block, and record its corresponding classification
Then judge whether comprise target in this subimage block according to the magnitude relationship between reconstructed error e and the predefined threshold tau: if e<τ illustrate to comprise target, otherwise, illustrate that this subimage block is background;
Step 3 object detection and recognition:
Step a3: will judge the corresponding reconstructed error e of each subimage block that comprises target among the step c2, and form a reconstructed error matrix E=(e of the same size with the test gray level image, as to represent candidate target region
St)
P * QWherein, e
StThe value of locating in coordinate points (s, t) for the reconstruct error matrix,
P * Q is the size of test pattern, s=1, and 2 ... P, t=1,2 ... Q;
Will be among the step c2 judge the corresponding classification C of each subimage block that comprises target, forms one with test that gray level image is of the same size, the classification matrix L that represents the candidate target classification=(C
St)
P * QC wherein
StThe value of locating in coordinate points (s, t) for the classification matrix,
Step b3: change big or small G time that slides window S * S, repeating step 2 ~ step a3G time, the G that obtains a reconstructed error matrix and G classification matrix, the span of G is 5 ~ 10; G the reconstructed error matrix that obtains formed a multiple dimensioned reconstructed error matrix M AP=(e
Stg)
P * Q * GWherein, e
StgBe the element among the matrix M AP, its value is the corresponding e of reconstructed error matrix that the sliding window size of the g time change obtains
St, P * Q * G is the size of multiple dimensioned reconstructed error matrix, g=1, and 2 ... G;
G the classification matrix that obtains consisted of a multiple dimensioned classification Matrix C LASS=(C
Stg)
P * Q * GWherein, C
StgBe the element among the Matrix C LASS, its value is the corresponding C of classification matrix that the sliding window size of the g time change obtains
StObtain a minimal reconstruction error matrix (map (s, t)) according to multiple dimensioned reconstructed error matrix M AP
P * Q, wherein map (s, t) is the value that corresponding minimal reconstruction error matrix is located in coordinate points (s, t),
Then obtain the minimum classification matrix (class (s, t)) of corresponding minimal reconstruction error matrix
P * Q, wherein class (s, t) is the value that minimum classification matrix is located in coordinate points (s, t),
Obtain Scale Matrixes scale=(scale (s, t)) according to multiple dimensioned reconstructed error matrix M AP
P * Q, scale (s, t) is the value that corresponding Scale Matrixes is located in coordinate points (r, t),
Step c3: ask for minimal reconstruction error matrix (map (s, t))
P * QThe local neighborhood minimal value as the target response value that detects, the local neighborhood minimal value is at minimal reconstruction error matrix (map (s, t))
P * QIn corresponding coordinate be the center of target, according to the center at (class (s, t))
P * Q(scale (s, t))
P * QThe corresponding classification of target and scale size are found in the position of middle correspondence.
2. described remote sensing images multi-class targets detection and Identification method according to claim 1, it is characterized in that: described calculated with weighted average method formula is f (x, y)=0.3R (x, y)+0.59G (x, y)+0.11B (x, y), in the formula, the gray level image that f (x, y) obtains for method of weighted mean is at pixel (x, y) gray-scale value, R (x, y), G (x, y) and B (x, y) be respectively the training image of input at three component values of the RGB of pixel (x, y).
3. described remote sensing images multi-class targets detection and Identification method according to claim 1, it is characterized in that: described energy normalized computing formula is
In the formula, f
Norm(x, y) is the gray-scale value after f (x, y) the process energy normalized, and u and v are respectively the row and column size of gray level image.
6. described remote sensing images multi-class targets detection and Identification method according to claim 1 is characterized in that: described reconstructed error e
iComputing formula be
In the formula, γ is predefined weights, the value of γ
Scope is 0 ~ 1, m
iTo Y
iIn the element of the every delegation mean vector of averaging and obtaining; Y
iBe U
iThe optimum code coefficient that process dictionary D sparse coding obtains.
8. described remote sensing images multi-class targets detection and Identification method according to claim 1 is characterized in that: described FDDL software package parameter lambda
1Scope be 0.001 ~ 0.01, λ
2Scope be 0.01 ~ 0.1.
9. described remote sensing images multi-class targets detection and Identification method according to claim 1, it is characterized in that: the span of described S is the integer between 40 ~ 90, the span of b is the integer between 1 ~ 15.
10. described remote sensing images multi-class targets detection and Identification method according to claim 1, it is characterized in that: the span of described threshold tau is 0 ~ 1.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201210300645.8A CN102867195B (en) | 2012-08-22 | 2012-08-22 | Method for detecting and identifying a plurality of types of objects in remote sensing image |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201210300645.8A CN102867195B (en) | 2012-08-22 | 2012-08-22 | Method for detecting and identifying a plurality of types of objects in remote sensing image |
Publications (2)
Publication Number | Publication Date |
---|---|
CN102867195A true CN102867195A (en) | 2013-01-09 |
CN102867195B CN102867195B (en) | 2014-11-26 |
Family
ID=47446059
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201210300645.8A Active CN102867195B (en) | 2012-08-22 | 2012-08-22 | Method for detecting and identifying a plurality of types of objects in remote sensing image |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN102867195B (en) |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103258210A (en) * | 2013-05-27 | 2013-08-21 | 中山大学 | High-definition image classification method based on dictionary learning |
CN103632164A (en) * | 2013-11-25 | 2014-03-12 | 西北工业大学 | Reeling steel state identifying method based on KNN reeling steel picture data of optimized KAP sample |
CN104517121A (en) * | 2014-12-10 | 2015-04-15 | 中国科学院遥感与数字地球研究所 | Spatial big data dictionary learning method based on particle swarm optimization |
CN105740422A (en) * | 2016-01-29 | 2016-07-06 | 北京大学 | Pedestrian retrieval method and apparatus |
CN106067041A (en) * | 2016-06-03 | 2016-11-02 | 河海大学 | A kind of multi-target detection method of based on rarefaction representation of improvement |
CN107451595A (en) * | 2017-08-04 | 2017-12-08 | 河海大学 | Infrared image salient region detection method based on hybrid algorithm |
CN109190457A (en) * | 2018-07-19 | 2019-01-11 | 北京市遥感信息研究所 | A kind of oil depot complex target rapid detection method based on large format remote sensing images |
CN109946076A (en) * | 2019-01-25 | 2019-06-28 | 西安交通大学 | A kind of planet wheel bearing fault identification method of weighted multiscale dictionary learning frame |
CN110189328A (en) * | 2019-06-11 | 2019-08-30 | 北华航天工业学院 | A kind of Remote sensing image processing system and its processing method |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100124377A1 (en) * | 2008-11-19 | 2010-05-20 | Nec Laboratories America, Inc. | Linear spatial pyramid matching using sparse coding |
CN102129573A (en) * | 2011-03-10 | 2011-07-20 | 西安电子科技大学 | SAR (Synthetic Aperture Radar) image segmentation method based on dictionary learning and sparse representation |
CN102324047A (en) * | 2011-09-05 | 2012-01-18 | 西安电子科技大学 | High spectrum image atural object recognition methods based on sparse nuclear coding SKR |
-
2012
- 2012-08-22 CN CN201210300645.8A patent/CN102867195B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100124377A1 (en) * | 2008-11-19 | 2010-05-20 | Nec Laboratories America, Inc. | Linear spatial pyramid matching using sparse coding |
CN102129573A (en) * | 2011-03-10 | 2011-07-20 | 西安电子科技大学 | SAR (Synthetic Aperture Radar) image segmentation method based on dictionary learning and sparse representation |
CN102324047A (en) * | 2011-09-05 | 2012-01-18 | 西安电子科技大学 | High spectrum image atural object recognition methods based on sparse nuclear coding SKR |
Non-Patent Citations (1)
Title |
---|
梁天一 等: "基于稀疏编码的图像语义分类器模型", 《华东理工大学学报(自然科学报)》, vol. 33, no. 6, 31 December 2007 (2007-12-31), pages 827 - 892 * |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103258210A (en) * | 2013-05-27 | 2013-08-21 | 中山大学 | High-definition image classification method based on dictionary learning |
CN103258210B (en) * | 2013-05-27 | 2016-09-14 | 中山大学 | A kind of high-definition image classification method based on dictionary learning |
CN103632164B (en) * | 2013-11-25 | 2017-03-01 | 西北工业大学 | The volume firm state classification recognition methodss of the KNN coil image data based on KAP sample optimization |
CN103632164A (en) * | 2013-11-25 | 2014-03-12 | 西北工业大学 | Reeling steel state identifying method based on KNN reeling steel picture data of optimized KAP sample |
CN104517121A (en) * | 2014-12-10 | 2015-04-15 | 中国科学院遥感与数字地球研究所 | Spatial big data dictionary learning method based on particle swarm optimization |
CN105740422A (en) * | 2016-01-29 | 2016-07-06 | 北京大学 | Pedestrian retrieval method and apparatus |
CN106067041A (en) * | 2016-06-03 | 2016-11-02 | 河海大学 | A kind of multi-target detection method of based on rarefaction representation of improvement |
CN106067041B (en) * | 2016-06-03 | 2019-05-31 | 河海大学 | A kind of improved multi-target detection method based on rarefaction representation |
CN107451595A (en) * | 2017-08-04 | 2017-12-08 | 河海大学 | Infrared image salient region detection method based on hybrid algorithm |
CN109190457A (en) * | 2018-07-19 | 2019-01-11 | 北京市遥感信息研究所 | A kind of oil depot complex target rapid detection method based on large format remote sensing images |
CN109190457B (en) * | 2018-07-19 | 2021-12-03 | 北京市遥感信息研究所 | Oil depot cluster target rapid detection method based on large-format remote sensing image |
CN109946076A (en) * | 2019-01-25 | 2019-06-28 | 西安交通大学 | A kind of planet wheel bearing fault identification method of weighted multiscale dictionary learning frame |
CN109946076B (en) * | 2019-01-25 | 2020-04-28 | 西安交通大学 | Planetary wheel bearing fault identification method of weighted multi-scale dictionary learning framework |
CN110189328A (en) * | 2019-06-11 | 2019-08-30 | 北华航天工业学院 | A kind of Remote sensing image processing system and its processing method |
CN110189328B (en) * | 2019-06-11 | 2021-02-23 | 北华航天工业学院 | Satellite remote sensing image processing system and processing method thereof |
Also Published As
Publication number | Publication date |
---|---|
CN102867195B (en) | 2014-11-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN102867195B (en) | Method for detecting and identifying a plurality of types of objects in remote sensing image | |
CN110348319B (en) | Face anti-counterfeiting method based on face depth information and edge image fusion | |
CN107346436B (en) | Visual saliency detection method fusing image classification | |
CN105608456B (en) | A kind of multi-direction Method for text detection based on full convolutional network | |
Sirmacek et al. | Urban area detection using local feature points and spatial voting | |
Yuan et al. | Large-scale solar panel mapping from aerial images using deep convolutional networks | |
CN103390164B (en) | Method for checking object based on depth image and its realize device | |
CN102542295B (en) | Method for detecting landslip from remotely sensed image by adopting image classification technology | |
CN102324047B (en) | Hyper-spectral image ground object recognition method based on sparse kernel representation (SKR) | |
CN102819740B (en) | A kind of Single Infrared Image Frame Dim targets detection and localization method | |
CN110555475A (en) | few-sample target detection method based on semantic information fusion | |
CN106023220A (en) | Vehicle exterior part image segmentation method based on deep learning | |
CN101930549B (en) | Second generation curvelet transform-based static human detection method | |
CN104008370A (en) | Video face identifying method | |
CN103942540A (en) | False fingerprint detection algorithm based on curvelet texture analysis and SVM-KNN classification | |
CN104182985A (en) | Remote sensing image change detection method | |
CN106845341A (en) | A kind of unlicensed vehicle identification method based on virtual number plate | |
CN102663724B (en) | Method for detecting remote sensing image change based on adaptive difference images | |
CN103745233B (en) | The hyperspectral image classification method migrated based on spatial information | |
CN103646256A (en) | Image characteristic sparse reconstruction based image classification method | |
CN104809471A (en) | Hyperspectral image residual error fusion classification method based on space spectrum information | |
CN111275070B (en) | Signature verification method and device based on local feature matching | |
CN106446854A (en) | High-resolution optical remote sensing image target detection method based on rotation invariant HOG feature | |
CN108073940A (en) | A kind of method of 3D object instance object detections in unstructured moving grids | |
CN103455826A (en) | Efficient matching kernel body detection method based on rapid robustness characteristics |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant |