CN105184320A - Non-negative sparse coding image classification method based on structural similarity - Google Patents

Non-negative sparse coding image classification method based on structural similarity Download PDF

Info

Publication number
CN105184320A
CN105184320A CN201510566662.XA CN201510566662A CN105184320A CN 105184320 A CN105184320 A CN 105184320A CN 201510566662 A CN201510566662 A CN 201510566662A CN 105184320 A CN105184320 A CN 105184320A
Authority
CN
China
Prior art keywords
matrix
code book
image
sparse coding
sift feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510566662.XA
Other languages
Chinese (zh)
Other versions
CN105184320B (en
Inventor
石伟伟
王进军
龚怡宏
张世周
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xian Jiaotong University
Original Assignee
Xian Jiaotong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xian Jiaotong University filed Critical Xian Jiaotong University
Priority to CN201510566662.XA priority Critical patent/CN105184320B/en
Publication of CN105184320A publication Critical patent/CN105184320A/en
Application granted granted Critical
Publication of CN105184320B publication Critical patent/CN105184320B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

The invention provides a non-negative sparse coding image classification method based on structural similarity. The method comprises the steps that dense SIFT feature extracting is carried out on all images in an image data set to be processed; a number of SIFT features are randomly selected to acquire the codebook of the image data set to be processed; a non-negative sparse coding model based on structural similarity is established; the SIFT features which are randomly selected to solve the codebook of the image data set; the codebook is fixed, and all SIFT features are coded; a spatial pyramid maximum pooling method is used to integrate the coding of each image in the image data set, so as to acquire the feature vector of each image; the image data set is divided into a training set and a test set, and the image feature vectors after spatial pyramid maximum pooling in the training set and corresponding labels of the images are used to train a classifier; and for any image, the feature vector after spatial pyramid maximum pooling is input into the trained classifier to acquire the prediction category of the image.

Description

The image classification method of the non-negative sparse coding of structure based similarity
Technical field:
The invention belongs to computer visual image sorting technique field, be specifically related to a kind of image classification method of non-negative sparse coding of structure based similarity.
Background technology:
A critical function in biological vision system primary stage is exactly the statistical redundancy removing input stimulus as far as possible.The response that primary visual cortex stimulates to external world meets openness, namely only has the neuron of minority to be activated, is encoded to sparse coding accordingly.Sparse coding generally, is exactly a combination signal being expressed as one group of base, and requires only to need several bases of minority just can by signal reconstruction out.Sparse coding has been widely applied to the field such as computer vision, picture signal process, such as, and the application such as signal reconstruction, signal denoising, image characteristics extraction and classification.
Structural information is defined as independent of brightness, contrast by structural similarity index, the attribute of reflected signal structure, and is the combination of brightness, contrast and structure three Different factor by distortion modeling.With the estimation of average as brightness, the estimation that standard deviation is spent as a comparison, covariance is as the tolerance of structure similarity degree.
Traditional sparse coding method is based on the reconstruct under least mean-square error meaning, namely make the quadratic sum of reconstructed error little as far as possible, meanwhile, make rarefaction of encoding accordingly as far as possible, the element of the proper vector that sparse table is encoded in encoding now is exactly as much as possible is zero.The current image classification method based on sparse coding, major part is all do based on the encoding model of minimum reconstructed quadratic sum, and error sum of squares does not meet the visual characteristic of human eye as the judgment criteria of distortion.Nearest research shows, the major function of human visual system is the structured message extracted from visual zone image and video, and error sum of squares does not consider the visual characteristic of human eye fully, therefore traditional sparse coding reconstruct can not evaluate the structural similarity of reconstructed image and original image well.
Summary of the invention:
The object of the invention is to for the deficiencies in the prior art, provide a kind of image classification method of non-negative sparse coding of structure based similarity.
For achieving the above object, the present invention adopts following technical scheme:
The image classification method of the non-negative sparse coding of structure based similarity, comprises the following steps:
1) all images are concentrated densely to extract SIFT feature respectively to pending view data;
2), after concentrating the complete SIFT feature of all image zooming-out to pending view data, the SIFT feature choosing 5 ten thousand to 50 ten thousand is randomly for asking the code book of pending image data set;
3) the non-negative sparse coding model of structure based similarity is set up;
4) according to step 2) and step 3), the code book of this image data set is solved by the SIFT feature of random selecting;
5), after the code book of this image data set solves out, fixed code book, encodes to all SIFT feature;
6) concentrate the coding often opening image to carry out spatial pyramid maximum pond method to view data to integrate, often opened the proper vector of image;
7) image data set is divided into training set and test set, with image feature vector and the corresponding label of image in the maximum pond of the spatial pyramid of training set, trains a sorter;
8) for any image, proper vector after maximum for its spatial pyramid pondization is input in the sorter trained, obtains this image prediction classification.
The present invention further improves and is, step 1) in, concentrate all images according to the block of pixels of 16 to 32 and the sliding step of 6 to 10 to pending view data, densely extract the SIFT feature of each image.
The present invention further improves and is, step 3) in, if code book is A=[a 1, a 2..., a k], a base vector is shown in each list of A, and the columns of code book is k, SIFT feature vector x iunder code book A, corresponding sparse coding is s i, definition encoder matrix S=[s 1, s 2..., s n], namely each row of encoder matrix S are structural similarity sparse codings of corresponding SIFT feature; The objective function of non-negative sparse coding model is as follows:
m i n S ≥ 0 , ∀ i , || a i || = 1 F ( A , S ) = Σ i = 1 n ( 1 - S S I M ( x i , As i ) ) + λ Σ i = 1 n Σ j = 1 k s j i
Wherein, i=1,2 ..., n, n are the number of the SIFT feature of random choose, j=1,2 ..., k, k are the columns of code book A, s jifor sparse coding is s ia jth component;
Write as matrix form:
m i n S ≥ 0 , ∀ i , || a i || = 1 F ( A , S ) = Σ i = 1 n ( 1 - S S I M ( x i , As i ) ) + λ || S || m 1
Wherein, || a i||=1, namely each mould length arranged of code book or L-2 norm are 1, the m of representing matrix S 1norm, m 1norm equal the absolute value of all elements of matrix and;
In the presentation code matrix S of S >=0, each element is non-negative;
λ is adjustment structure distortion level and openness weight coefficient of encoding, and λ is larger, and corresponding coding is sparse, and has 0.05≤λ≤0.5;
SSIM () is structural similarity target function.
The present invention further improves and is, step 4) concrete methods of realizing as follows:
401) according to step 2) in choose randomly 5 ten thousand to 50 ten thousand SIFT feature, the value of given weight coefficient λ;
402) initialization code book A:
Random initializtion code book A (0), by be normalized into 1 by long for the mould of each row of code book A, put t=1; A (t)represent the t time iterative value of code book, A (0)represent code book initial value;
403) the corresponding encoder matrix S of random initializtion (0), each element in matrix is initialized as a random number between 0 to 1, S (t)the t time iterative value of presentation code matrix S;
404) encoder matrix S is upgraded, specific as follows:
Wherein: symbol ← expression assignment, the value by the right variable is assigned to the variable on the left side; Symbol with represent that Hardmard takes advantage of and removes respectively, namely the dot product of matrix and point remove; Evolution function sqrt () acts on each component elements of matrix representation to corresponding matrix and does open computing; B +=max (B, 0), B -=-min (B, 0), with evolution function sqrt (), max () and min () makes comparisons each element and 0 of matrix here respectively, gets maximum or least member accordingly; h be with the element of matrix S same order be entirely 1 matrix;
405) code book A is upgraded, specific as follows:
A ( t ) ← A ( t - 1 ) - σ ▿ A E
Wherein, σ is the Gradient Descent step-length optimizing code book, and the span of h is 0.01 ~ 0.1, molecule max{|A ij| represent the maximal value of element absolute value in code book A, denominator represent the gradient matrix of E about code book A the mean value of all elements absolute value, the transposition of subscript T representing matrix or vector, ρ represents penalty coefficient, and F is the objective function of non-negative sparse coding model;
406) if this variable relative of twice is less than 10 before and after target function value -6, or reach predetermined iterations, just stop performing, obtain code book matrix A; Otherwise, put t=t+1, forward step 403 to).
The present invention further improves and is, wherein, namely D is a diagonal matrix, and its diagonal element is respectively i is the unit matrix with D same order.
The present invention further improves and is, step 5) in, in objective function, code book is fixed, according to calculate the non-negative sparse coding s of the structure based similarity of SIFT feature x.
Relative to prior art, the present invention has following advantage:
The present invention is by being incorporated into inside sparse coding method by structural similarity index, change traditional coded system, improve the quality of coding, make to encode and more meet the coded system of human visual system, then the mode that corresponding coding carries out the maximum pond of spatial pyramid is integrated, the proper vector of image can be obtained, and then the proper vector behind maximum for spatial pyramid pond is applied to Images Classification.
The present invention eliminates the reconstruct distortion metrics based on least mean-square error in traditional sparse coding method, by the tolerance of structural similarity as reconstruct distortion, propose the non-negative sparse coding model of structure based similarity, the party can encode to the feature of any vectorization, is a kind of coding method meeting human-eye visual characteristic.
The present invention densely extracts SIFT feature to image, the similar non-negative sparse coding of structure based is carried out to SIFT feature, the proper vector that the maximum pondization of spatial pyramid just obtains entire image is carried out to coding, the proper vector obtained is used for Images Classification.
Accompanying drawing illustrates:
Fig. 1 is the process flow diagram of the image classification method of the non-negative sparse coding that the present invention is based on structural similarity.
Fig. 2 is spatial pyramid maximum pond schematic diagram.
Embodiment:
Below in conjunction with drawings and Examples, the present invention is described in further detail.
The present invention attempts from the angle of structural similarity to find corresponding sparse coding, the present invention introduces structural similarity keeps degree important measurement index as information, then add that non-negative sparse retrains, give the non-negative sparse coding model of structure based similarity.The nonnegativity why requirement is encoded is because the coding of non-negative has better stability in the application.
Setting signal x and y, x, y ∈ R n, structural similarity is defined as follows:
S S I M ( x , y ) = ( 2 μ x μ y + C 1 ) ( 2 σ x , y + C 2 ) ( μ x 2 + μ y 2 + C 1 ) ( σ x 2 + σ y 2 + C 2 ) ,
Wherein, x ifor i-th component elements of signal x, y ifor i-th component elements of signal y, σ xand σ yrepresent the standard deviation of signal x and y respectively, &sigma; y = ( 1 N - 1 &Sigma; i = 1 N ( y i - &mu; y ) 2 ) 1 2 , σ x,yrepresent the covariance of signal x and y, &sigma; x , y = 1 N - 1 &Sigma; i = 1 N ( x i - &mu; x ) ( y i - &mu; y ) , 0<C 1, C 2< < 1 is 0 and ad hoc two minimum normal numbers for avoiding denominator, and structural similarity, more close to 1, illustrates that these two features are more close.Therefore, 1-SSIM (x, y) can as the degree weighing Characteristic Distortion.
As shown in Figure 1, the present invention is based on the image classification method of the non-negative sparse coding of structural similarity, comprise the following steps:
1): to each width picture of data set according to a certain size block of pixels (such as the image block of 16 × 16 pixel sizes) and pre-determined sliding step up and down (such as the sliding step of 6 pixels), densely extract SIFT feature.
2): from extracted all SIFT feature, choose n (such as, 200,000) individual SIFT feature randomly, selected whole features form a matrix, are designated as X, note X=[x 1, x 2..., x n]; Each arranges x i∈ R p × 1(i=1,2 ..., n) represent a SIFT feature vector, p represents the dimension of extracted SIFT feature, if when SIFT feature vector is 128 dimension, so p=128, whole SIFT feature selected are here for asking the code book of this data set.
3): set code book as A=[a 1, a 2..., a k], a base vector is shown in each list of A, and the columns of code book is k, SIFT feature vector x i(i=1,2 ..., sparse coding n) under code book A is s i(i=1,2 ..., n), definition encoder matrix S=[s 1, s 2..., s n].Provide the objective function of non-negative sparse coding model below:
m i n S &GreaterEqual; 0 , &ForAll; i , || a i || = 1 F ( A , S ) = &Sigma; i = 1 n ( 1 - S S I M ( x i , As i ) ) + &lambda; &Sigma; i = 1 n &Sigma; j = 1 k s j i
Wherein, i=1,2 ..., n, n are the number of the SIFT feature of random choose, j=1,2 ..., k, k are the columns of code book, s jifor sparse coding is s ia jth component;
Write as matrix form:
m i n S &GreaterEqual; 0 , &ForAll; i , || a i || = 1 F ( A , S ) = &Sigma; i = 1 n ( 1 - S S I M ( x i , As i ) ) + &lambda; || S || m 1
Wherein, || a i||=1, namely each mould length arranged of code book or L-2 norm are 1 (the L-2 norm of vector equals its all elements square root sum square), and requiring the mould of each base in code book length to be 1 is to prevent trivial solution, the m of representing matrix S 1norm, m 1norm equal the absolute value of all elements of matrix and.In the representing matrix S of S >=0, each element is non-negative.In objective function, the effect of Section 1 is that structure is similar as much as possible to former feature in order to ensure the feature of reconstruct, that the one of reconstruct distortion is weighed, in traditional sparse coding model, reconstruct distortion weighs by the quadratic sum of error, and the present invention's structural similarity index is weighed.Section 2 in objective function is to ensure the openness of coding.λ regulates this weight coefficient of two, and λ is larger, and corresponding coding is sparse, and in method implementation procedure, λ is taken as the constant being greater than zero, and the size of λ value can regulate according to different data sets.Optimum code book sought by the present invention's needs and corresponding structural similarity sparse coding makes objective function obtain minimum value.
4) SIFT feature selected: according to step 2) and step 3) in objective function, solve the code book of the optimum of pending image data set.Adopt the strategy of alternative optimization to optimize the objective function of the non-negative sparse coding model of structure based similarity, concrete scheme is:
The first step: given selected SIFT feature, forms matrix X; The value of given weight coefficient λ.
Second step, initialization code book A;
3rd step: fixing A, optimizes S, adopts the convex optimization method of nonnegative matrix;
4th step: fixing S, optimizes A, adopts gradient descent method;
Repeat the 3rd step and the 4th step, until algorithm convergence.
When objective function converges, the code book obtained is exactly the code book of this data set.Judge that the method for convergence can with of following two kinds of criterions: before and after (a) target function value, this variable relative of twice is less than 10 -6; This variable relative of b F-norm that the front and back two of () code book are taken second place poor is less than 10 -6; When restraining, A is required code book.Here the F-norm of matrix equals square root sum square of all elements in matrix.
Concrete analysis is as follows with the process solving scheme:
By F ( A , S ) = &Sigma; i = 1 n ( 1 - S S I M ( x i , As i ) ) + &lambda; || S || m 1 Can obtain
&part; F &part; S = &dtri; S F = - B + &lambda; H
Wherein, h for the element of S same order (namely line number and columns are all equal) be entirely 1 matrix.
Because mould length of each row of bound for objective function requirement code book is 1, the restricted problem about code book is converted into unconstrained problem by Means of Penalty Function Methods by us.
That is, make here, the transposition of subscript T representing matrix or vector, ρ represents penalty coefficient, ρ=1 to 1000.
&part; E &part; A = &dtri; A E = &part; F &part; A + 4 &rho; A ( D - I ) , Wherein, D = d i a g ( a 1 T a 1 , a 2 T a 2 , ... , a k T a k ) , Namely be a diagonal matrix, its diagonal element is respectively i is the unit matrix with D same order.
Order for optimizing the Gradient Descent step-length of code book, the span of s is 0.01 ~ 0.1, generally gets s=0.05, molecule max{|A ij| representing the element value of maximum absolute value in code book, denominator represents that E is about code book A gradient matrix the mean value of all elements absolute value.
Optimized algorithm:
Step 1: random initializtion code book A (0), by be normalized into 1 by long for the mould of each row of code book, put t=1; A (t)represent the t time iterative value of code book, A (0)represent code book initial value.
Step 2: the corresponding encoder matrix S of random initializtion (0)(each element in matrix is initialized as a random number between 0 to 1), A (t)represent the t time iterative value of S.
Step 3: upgrade encoder matrix S,
Step 4: upgrade code book (code book) A
A ( t ) &LeftArrow; A ( t - 1 ) - &sigma; &dtri; A E
Step 5: if algorithm convergence, or reach predetermined iterations, just stop performing, obtain A, S; Otherwise, put t=t+1, forward step 3 to.
Note: here, symbol ← expression assignment, the value by the right variable is assigned to the variable on the left side; Symbol with represent that Hardmard takes advantage of and removes respectively, namely the dot product of matrix removes (each component elements does corresponding computing respectively) with point; Evolution function sqrt () acts on each component elements of matrix representation to corresponding matrix and does open computing; For any one matrix M, M +=max (M, 0), M -=-min (M, 0), with evolution function sqrt (), max () and min () makes comparisons each element and 0 of matrix here respectively, gets maximum or least member accordingly.
5): fixed code book, then according to the non-negative sparse coding model of structure based similarity, corresponding sparse coding is calculated to the SIFT feature of each width of data centralization figure.Namely in objective function, code book is fixed, according to calculate the non-negative sparse coding s of the structure based similarity of SIFT feature x, calculate SIFT feature x sparse coding adopt Methods and steps 4) in fixed code book A go the method for Optimized Coding Based matrix S identical.
6): for the every piece image of pending image data set, by the corresponding sparse coding of its all SIFT feature, integrate according to the mode in the maximum pond of spatial pyramid (SPMmaxpooling), can obtain the proper vector of a higher-dimension, the proper vector of the higher-dimension obtained is exactly the proper vector for visual tasks such as classifying of this image.
As shown in Figure 2, the maximum pond of spatial pyramid (SPMmaxpooling): spatial pyramid is exactly the grid be divided into respectively by former picture, generally be divided into the grid of 1 × 1,2 × 2,4 × 4, each grid can regard a larger image block as.In each image block, carry out maximum pond to the sparse coding of its all SIFT feature on every one dimension, namely in each dimension, the result in pond is the maximal value of absolute value in all sparse coding respective dimension.The proper vector on this image block can be obtained behind each grid pond, proper vector after all grids or image block upper storage reservoir is stitched together, the feature of the higher-dimension obtained is exactly the proper vector of this picture in its entirety, and this operating process is just called the maximum pond of spatial pyramid.Figure of description 1 is shown in by the schematic diagram in the maximum pond of spatial pyramid.
7): in each class of pending image data set, several pictures of random selecting, or determine some images in advance, as training set, with label training sorter that the characteristic sum in the maximum pond of the spatial pyramid of training set image is corresponding, after sorter trains, the parameter of sorter just determines, and generally just uses SVM classifier.
8) proper vector behind maximum for the spatial pyramid of test set picture pond is input to sorter, just can obtains corresponding classification prediction label.

Claims (6)

1. the image classification method of the non-negative sparse coding of structure based similarity, is characterized in that, comprises the following steps:
1) all images are concentrated densely to extract SIFT feature respectively to pending view data;
2), after concentrating the complete SIFT feature of all image zooming-out to pending view data, the SIFT feature choosing 5 ten thousand to 50 ten thousand is randomly for asking the code book of pending image data set;
3) the non-negative sparse coding model of structure based similarity is set up;
4) according to step 2) and step 3), the code book of this image data set is solved by the SIFT feature of random selecting;
5), after the code book of this image data set solves out, fixed code book, encodes to all SIFT feature;
6) concentrate the coding often opening image to carry out spatial pyramid maximum pond method to view data to integrate, often opened the proper vector of image;
7) image data set is divided into training set and test set, with image feature vector and the corresponding label of image in the maximum pond of the spatial pyramid of training set, trains a sorter;
8) for any image, proper vector after maximum for its spatial pyramid pondization is input in the sorter trained, obtains this image prediction classification.
2. the image classification method of the non-negative sparse coding of structure based similarity according to claim 1, it is characterized in that, step 1) in, concentrate all images according to the block of pixels of 16 to 32 and the sliding step of 6 to 10 to pending view data, densely extract the SIFT feature of each image.
3. the image classification method of the non-negative sparse coding of structure based similarity according to claim 1, is characterized in that, step 3) in, if code book is A=[a 1, a 2..., a k], a base vector is shown in each list of A, and the columns of code book is k, SIFT feature vector x iunder code book A, corresponding sparse coding is s i, definition encoder matrix S=[s 1, s 2..., s n], namely each row of encoder matrix S are structural similarity sparse codings of corresponding SIFT feature; The objective function of non-negative sparse coding model is as follows:
m i n S &GreaterEqual; 0 , &ForAll; i , || a i || = 1 F ( A , S ) = &Sigma; i = 1 n ( 1 - S S I M ( x i , As i ) ) + &lambda; &Sigma; i = 1 n &Sigma; j = 1 k s j i
Wherein, i=1,2 ..., n, n are the number of the SIFT feature of random choose, j=1,2 ..., k, k are the columns of code book A, s jifor sparse coding is s ia jth component;
Write as matrix form:
m i n S &GreaterEqual; 0 , &ForAll; i , || a i || = 1 F ( A , S ) = &Sigma; i = 1 n ( 1 - S S I M ( x i , As i ) ) + &lambda; || S || m 1
Wherein, || a i||=1, namely each mould length arranged of code book or L-2 norm are 1, the m of representing matrix S 1norm, m 1norm equal the absolute value of all elements of matrix and;
In the presentation code matrix S of S >=0, each element is non-negative;
λ is adjustment structure distortion level and openness weight coefficient of encoding, and λ is larger, and corresponding coding is sparse, and has 0.05≤λ≤0.5;
SSIM () is structural similarity target function.
4. the image classification method of the non-negative sparse coding of structure based similarity according to claim 3, is characterized in that, step 4) concrete methods of realizing as follows:
401) according to step 2) in choose randomly 5 ten thousand to 50 ten thousand SIFT feature, the value of given weight coefficient λ;
402) initialization code book A:
Random initializtion code book A (0), by be normalized into 1 by long for the mould of each row of code book A, put t=1; A (t)represent the t time iterative value of code book, A (0)represent code book initial value;
403) the corresponding encoder matrix S of random initializtion (0), each element in matrix is initialized as a random number between 0 to 1, S (t)the t time iterative value of presentation code matrix S;
404) encoder matrix S is upgraded, specific as follows:
Wherein: symbol ← expression assignment, the value by the right variable is assigned to the variable on the left side; Symbol with represent that Hardmard takes advantage of and removes respectively, namely the dot product of matrix and point remove; Evolution function sqrt () acts on each component elements of matrix representation to corresponding matrix and does open computing; B +=max (B, 0), B -=-min (B, 0), with evolution function sqrt (), max () and min () makes comparisons each element and 0 of matrix here respectively, gets maximum or least member accordingly; B=[b 1, b 2..., b n], h be with the element of matrix S same order be entirely 1 matrix;
405) code book A is upgraded, specific as follows:
A ( t ) &LeftArrow; A ( t - 1 ) - &sigma; &dtri; A E
Wherein, σ is the Gradient Descent step-length optimizing code book, and the span of h is 0.01 ~ 0.1, molecule max{|A ij| represent the maximal value of element absolute value in code book A, denominator represent the gradient matrix of E about code book A the mean value of all elements absolute value, the transposition of subscript T representing matrix or vector, ρ represents penalty coefficient, and F is the objective function of non-negative sparse coding model;
406) if this variable relative of twice is less than 10 before and after target function value -6, or reach predetermined iterations, just stop performing, obtain code book matrix A; Otherwise, put t=t+1, forward step 403 to).
5. the image classification method of the non-negative sparse coding of structure based similarity according to claim 4, is characterized in that, &dtri; A E = &part; F &part; A + 4 &rho; A ( D - I ) , Wherein, D = d i a g ( a 1 T a 1 , a 2 T a 2 , ... , a k T a k ) , Namely D is a diagonal matrix, and its diagonal element is respectively i is the unit matrix with D same order.
6. the image classification method of the non-negative sparse coding of structure based similarity according to claim 4, is characterized in that, step 5) in, in objective function, code book is fixed, according to calculate the non-negative sparse coding s of the structure based similarity of SIFT feature x.
CN201510566662.XA 2015-09-08 2015-09-08 The image classification method of non-negative sparse coding based on structural similarity Active CN105184320B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510566662.XA CN105184320B (en) 2015-09-08 2015-09-08 The image classification method of non-negative sparse coding based on structural similarity

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510566662.XA CN105184320B (en) 2015-09-08 2015-09-08 The image classification method of non-negative sparse coding based on structural similarity

Publications (2)

Publication Number Publication Date
CN105184320A true CN105184320A (en) 2015-12-23
CN105184320B CN105184320B (en) 2019-01-15

Family

ID=54906384

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510566662.XA Active CN105184320B (en) 2015-09-08 2015-09-08 The image classification method of non-negative sparse coding based on structural similarity

Country Status (1)

Country Link
CN (1) CN105184320B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106096658A (en) * 2016-06-16 2016-11-09 华北理工大学 Based on the Aerial Images sorting technique without supervision deep space feature coding
CN106408018A (en) * 2016-09-13 2017-02-15 大连理工大学 Image classification method based on amplitude-frequency characteristic sparse filtering
CN116842030A (en) * 2023-09-01 2023-10-03 广州尚航信息科技股份有限公司 Data synchronous updating method and system of server

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103020647A (en) * 2013-01-08 2013-04-03 西安电子科技大学 Image classification method based on hierarchical SIFT (scale-invariant feature transform) features and sparse coding

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103020647A (en) * 2013-01-08 2013-04-03 西安电子科技大学 Image classification method based on hierarchical SIFT (scale-invariant feature transform) features and sparse coding

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
JIANCHAO YANG ET AL: "Linear spatial pyramid matching using sparse coding for image classification", 《COMPUTER VISION AND PATTERN RECOGNITION》 *
李志清: "基于结构相似度的稀疏编码模型", 《软件学报》 *
李钱钱: "基于改进的非负稀疏编码图像分类研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106096658A (en) * 2016-06-16 2016-11-09 华北理工大学 Based on the Aerial Images sorting technique without supervision deep space feature coding
CN106096658B (en) * 2016-06-16 2019-05-24 华北理工大学 Aerial Images classification method based on unsupervised deep space feature coding
CN106408018A (en) * 2016-09-13 2017-02-15 大连理工大学 Image classification method based on amplitude-frequency characteristic sparse filtering
CN106408018B (en) * 2016-09-13 2019-05-14 大连理工大学 A kind of image classification method based on amplitude-frequency characteristic sparseness filtering
CN116842030A (en) * 2023-09-01 2023-10-03 广州尚航信息科技股份有限公司 Data synchronous updating method and system of server
CN116842030B (en) * 2023-09-01 2023-11-17 广州尚航信息科技股份有限公司 Data synchronous updating method and system of server

Also Published As

Publication number Publication date
CN105184320B (en) 2019-01-15

Similar Documents

Publication Publication Date Title
CN110414377B (en) Remote sensing image scene classification method based on scale attention network
Yang et al. Learning face age progression: A pyramid architecture of gans
Wang et al. Modality and component aware feature fusion for RGB-D scene classification
JP6244059B2 (en) Face image verification method and face image verification system based on reference image
CN109840560B (en) Image classification method based on clustering in capsule network
CN107977932A (en) It is a kind of based on can differentiate attribute constraint generation confrontation network face image super-resolution reconstruction method
CN104463209A (en) Method for recognizing digital code on PCB based on BP neural network
CN109670576B (en) Multi-scale visual attention image description method
CN106203487A (en) A kind of image classification method based on Multiple Kernel Learning Multiple Classifier Fusion and device
CN106920243A (en) The ceramic material part method for sequence image segmentation of improved full convolutional neural networks
CN108764195A (en) Handwriting model training method, hand-written character recognizing method, device, equipment and medium
CN106529447A (en) Small-sample face recognition method
CN106485259B (en) A kind of image classification method based on high constraint high dispersive principal component analysis network
CN105701502A (en) Image automatic marking method based on Monte Carlo data balance
CN105574548A (en) Hyperspectral data dimensionality-reduction method based on sparse and low-rank representation graph
CN104484886B (en) A kind of dividing method and device of MR images
CN106250811B (en) Unconstrained face identification method based on HOG feature rarefaction representation
CN107729312A (en) More granularity segmenting methods and system based on sequence labelling modeling
CN106529586A (en) Image classification method based on supplemented text characteristic
CN108446589A (en) Face identification method based on low-rank decomposition and auxiliary dictionary under complex environment
CN105631469A (en) Bird image recognition method by multilayer sparse coding features
CN105868711B (en) Sparse low-rank-based human behavior identification method
CN109978074A (en) Image aesthetic feeling and emotion joint classification method and system based on depth multi-task learning
CN105184320A (en) Non-negative sparse coding image classification method based on structural similarity
CN105260736A (en) Fast image feature representing method based on normalized nonnegative sparse encoder

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant