CN102982165A - Large-scale human face image searching method - Google Patents

Large-scale human face image searching method Download PDF

Info

Publication number
CN102982165A
CN102982165A CN2012105278368A CN201210527836A CN102982165A CN 102982165 A CN102982165 A CN 102982165A CN 2012105278368 A CN2012105278368 A CN 2012105278368A CN 201210527836 A CN201210527836 A CN 201210527836A CN 102982165 A CN102982165 A CN 102982165A
Authority
CN
China
Prior art keywords
facial image
face
image
people
local feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN2012105278368A
Other languages
Chinese (zh)
Other versions
CN102982165B (en
Inventor
杨育彬
毛晓蛟
钱洪波
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University
Original Assignee
Nanjing University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University filed Critical Nanjing University
Priority to CN201210527836.8A priority Critical patent/CN102982165B/en
Publication of CN102982165A publication Critical patent/CN102982165A/en
Application granted granted Critical
Publication of CN102982165B publication Critical patent/CN102982165B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Processing Or Creating Images (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a large-scale human face image searching method. The method comprises the following steps of preprocessing human face images; extracting local characteristics from the human face images; extracting overall geometrical characteristics from the human face images; quantifying the local characteristics; quantifying the overall geometrical characteristics; establishing a reverse index; searching a candidate human face image set; and re-arranging the candidate human face image set. By the method, an index for a large-scale human face image database can be established, quick human face research is realized, and the research efficiency is realized. In addition, the accuracy of human face research is improved by embedding an auxiliary information characteristic quantifying and candidate human face image set re-arranging algorithm. Effective and accurate large-scale human face image search is realized by the method, so that the method has higher use value.

Description

A kind of extensive Research on face image retrieval
Technical field
The invention belongs to the facial image searching field, particularly a kind of extensive Research on face image retrieval.
Background technology
In recent years, along with the rise of microblogging and social network sites and the demand of public safety, the data of facial image increase the scale of magnanimity rapidly, in the face database of such magnanimity, retrieving own interested a part of facial image has become a urgent demand, and extensive face retrieval also so gradually becomes the focus of research.Extensive facial image retrieval requires algorithm that the data scale is had good extensibility, and in addition, recall precision, recall rate and accuracy rate etc. are to estimate the common index of retrieval performance, not only require recall precision high, also will guarantee the accuracy of retrieving.
For extensive face retrieval problem, if directly use the method for traditional recognition of face, being the feature of higher-dimension complexity and wanting the whole face database of linear sweep to seek the most similar people's face of extraction do not have extensibility.Simply utilize the word bag model (Bag-of-Words, BoW) in CBIR field, owing to do not utilize the special construction information of people's face itself, the error of quantification is larger, thereby retrieval rate is lower.Some solutions of present main flow, the method based on the ID quantization encoding such as partial block, come the lower quantization loss by this priori of positional information of utilizing people's face partial block, the method that reorders with many reference sets simultaneously improves accuracy rate, obtained preferably effect, but, these solutions still exist people's face geometry information under-utilized, quantize loss and error is large, reference set generates problems such as age etc. being changed inadequate robust, extensive face retrieval remains further to be improved in performance.
Summary of the invention
Goal of the invention: the present invention has proposed a kind of extensive Research on face image retrieval in order to solve the problems of the prior art, thereby effectively solves under the large-scale data, the quick and precisely search problem of facial image.
Summary of the invention: the invention discloses a kind of extensive Research on face image retrieval, comprise following steps:
Face images in facial image to be retrieved and the face database is carried out respectively the processing of following steps 1 ~ step 4;
Step 1, the facial image pre-service: the key point in the facial image of location, and with facial image and the standard faces image alignment of presetting;
Step 2, extract local feature: according to the key point position facial image is divided into the p piece, each piece is called a localized mass, extracts the feature of each localized mass;
Step 3 is extracted overall geometric properties: according to the key point position, extract facial image overall situation geometric properties, comprise distance feature, angle character and curvature feature;
Step 4 is quantified as the vision word with facial image local feature and overall geometric properties: utilize training set to obtain local feature dictionary and global characteristics dictionary, facial image local feature and overall geometric properties are quantified as the vision word;
Step 5 is set up inverted index to the face images in the face database;
Step 6 obtains the candidate face image collection according to the vision word of facial image to be retrieved from the inverted index of step 5;
Step 7 utilizes the algorithm that reorders that the candidate face image collection is reordered, and obtains the facial image reference set, and the facial image reference set is final orderly result for retrieval.
The pre-service of step 1 facial image specifically comprises the steps:
At first, utilize the key point in the facial image of active shape model (Active Shape Model, ASM) location, people's face shape is regarded in the set of all key points of people's face as, and the vector that these key points form is called people's face shape vector.According to two people's face shape vectors, i.e. the set of the key point of two facial images, i people's face shape vector is designated as:
X i=(x i1,y i1,x i2,y i2,…x in,y in) T
(x wherein Ij, y Ij) coordinate of expression i facial image j key point, wherein j is the arbitrary numerical value among 1 ~ n, n is the key point number, preferred 76 of the present invention, the transposition of T representing matrix; General key point is chosen for the natural number more than 10.
Given facial image a and facial image b, its shape is respectively X a=(x A1, y A1, x A2, y A2... x An, y An) TAnd X b=(x B1, y B1, x B2, y B2... x Bn, y Bn) T
Find parameter group (θ, s, t), make parameter E=(X a-M (s, θ) [X b]-t) TW (X a-M (s, θ) [X bThe value of]-t) is minimum; The conversion of people's face shape vector [X to its function M (s, θ) i]+t represents, wherein:
M ( s , θ ) x ij y ij = ( s cos θ ) x ij - ( s sin θ ) y ij ( s sin θ ) x ij + ( s cos θ ) y ij , t=(t x,t y,…,t x,t y) T
W is the weights (w of each key point 1, w 2..., w n) diagonal matrix that forms, each key point weights is the same, n key point, and the weights of each key point are exactly 1/n, can certainly specifically set as required the weights at different facial images position, if each weights more than or equal to 0 and all weights sums be 1 just passable.t x, t yBe respectively key point x, the displacement that the y coordinate is corresponding.In minimizing the process of E, each parameter differentiated can obtain the result, according to the parameter that obtains facial image is changed, and makes two people's faces keep maximum position corresponding, and process is as follows: make scos θ=a x, ssin θ=a y, to a x, a y, t x, t yDifferentiate separately, make E minimum, obtain following four equatioies:
B x a x - B y a y + Wt x = A x B y a x + B x a y + Wt y = A y Za x + B x t x + B y t y = C 1 Za y - B y t x + B x t y = C 2 ,
Wherein A x = Σ k = 1 n w k x ak , B x = Σ k = 1 n w k x bk , A y = Σ k = 1 n w k x ak , B y = Σ k = 1 n w k x bk ,
Z = Σ k = 1 n w k ( x bk 2 + y bk 2 ) , W = Σ k = 1 n w k ,
C 1 = Σ k = 0 n w k ( x ak x bk + y ak y bk ) ,
C 2 = Σ k = 0 n w k ( y ak x bk + x ak y bk ) .
Calculating can obtain conversion parameter group (θ by solving an equation; S; T), again image is changed.
Step 2 is extracted local feature and is comprised:
According to the key point in the facial image of step 1 location, people's face is carried out piecemeal.At first according to the organ site in the facial image, mainly comprise eyes, nose, cheek, people's face is divided into five bulks, method is as follows: with all key points of left eye as a set, with 4 key points of upper and lower, left and right as the border, form a rectangle, this rectangle is considered to piece corresponding to left eye.The acquisition methods of all the other four pieces is identical with the method for obtaining the left eye piece.Then with every a bulk of 4 * 5 fritters that are equally divided into, between these 20 the overlapping region is arranged.People's face can be divided into 100 fritters, and each fritter has independently numbering, and Serial Number Range is 1-100.At last, each piece is extracted local binary (Local Binary Pattern, LBP) and yardstick invariant features conversion (Scale Invariant Feature Transform, SIFT) two kinds of features.
Step 3 is extracted overall geometric properties and is comprised:
According to the key point in the facial image of step 1 location, extract the overall geometric properties between the key point, overall geometric properties is described the space geometry relation between the key point, comprises distance, angle, curvature.According to the key point of location, calculate in twos distance, will be spliced into vector apart from linearity, be called distance feature; Calculate the angle that forms angle between the key point, the angle linearity is spliced into vector, be called the angle vector.Key point to the chin position is calculated curvature, and the curvature linearity is spliced into vector, is called curvature vector.With distance vector, angle vector sum curvature vector linear mosaic, it is exactly the proper vector of overall geometric properties.
Definition vision word comprises in the step 4:
In a kind of extensive face retrieval method, definition vision word is<Name ID Age ID, Gender ID, Position ID 〉, each part or overall geometric properties all are quantified as one or more vision words.In training set, each facial image has its name, age bracket, sex information separately.Wherein Name ID represents the name of people's face in the corresponding facial image training set of this feature, Age ID represents the age bracket under people's face in the corresponding facial image training set of this feature, Gender ID represents the sex of people's face in the corresponding facial image training set of this feature, and Position ID represents block number corresponding to facial image in the corresponding facial image training set of this feature.Especially, for local feature, because people's face is divided into 100 fritters, so the Position ID of the vision word that is quantized into of local feature is the numbering of its corresponding people's face piece, and for overall geometric properties, Position ID is unified to be set to 101.
Local feature quantizes to comprise in the step 4:
Divide block number with everyone face of training set by step 2, every is extracted respectively local feature, comprises the conversion of local binary and yardstick invariant features.All proper vectors of same characteristic features kind (namely being all the conversion of local binary or yardstick invariant features), same age, identical sex, identical numbering are classified as one group.Every stack features utilizes one group of base of sparse coding model training.
min D , V Σ i = 1 n | | x i - Dv i | | 2 + λ | | v i | | 1
s . t . | | v i | | ≤ 1 , ∀ i ,
Wherein, the dictionary that D arrives for study, x iBe the proper vector of every group of training set, v iBe x iCoefficient when carrying out linear reconstruction with D.λ is Sparse parameter.Can calculate dictionary D by training set, can train for every group and obtain a dictionary D.When retrieval, the age bracket of a given proper vector x' and its affiliated people's face, sex and minute block number are chosen corresponding dictionary D', with dictionary x' are reconstructed, and obtain reconstruction coefficients v', and at this moment, local feature x' just is encoded as v'.The name of people's face under the corresponding base of non-zero entry is the Name ID after x' quantizes among the v'.The age bracket of x', sex and piece label are respectively Age ID, Gender ID and the Position ID after the quantification.
Overall geometric properties quantizes to comprise in the step 4:
To all training set face extraction overall situation geometric properties, with people face age bracket, the sex grouping of overall geometric properties according to correspondence, other overall geometric properties of same age bracket homogeneity is divided into one group.
A given overall geometric properties F gAnd age bracket, sex information, from training set, select the feature group of corresponding age bracket, sex, find F by neighbor search gThe neighbour, adjacent identity, age, sex are F gQuantification after Name ID, Age ID, Gender ID, the non-Position ID of overall geometric properties is unified to be set to 101.
Step 5 is set up inverted index and is comprised:
Facial images all in the database is set up inverted index, be used for retrieval.For a facial image, quantize according to step 5 and step 6 respectively, obtain representing the vision word of this people's face, this facial image is indexed on the vision word.For the facial image that does not have age and sex information in the database, can utilize existing age algorithm for estimating and sex algorithm for estimating to obtain.Because the age information that this algorithm needs is an age bracket but not therefore exact value, has certain fault-tolerance to the age algorithm for estimating.Everyone face of database is quantized, index respectively vision word separately.Index structure is the set of vision word, by given vision word, can retrieve the facial image set corresponding with it.
Step 6 obtains the candidate face image collection and comprises:
A given facial image to be retrieved obtains its vision word by above-mentioned quantization method, for the facial image that does not have age and sex information, can utilize existing age algorithm for estimating and sex algorithm for estimating to obtain.Because the age information that this algorithm needs is an age bracket but not therefore exact value, has certain fault-tolerance to the age algorithm for estimating.Find identical vision word from inverted index, facial image that will be corresponding with it takes out, and these facial images have formed the candidate face image collection, and the candidate face image collection is the preliminary search result.
Step 7 algorithm that reorders comprises:
At first utilize a part of facial image training set, training of human face sorter, this sorter are used for two facial images of prediction and whether belong to same person.Concrete steps are as follows:
(1) the sorter training set is some facial images pair, and wherein a part is the facial image pair of same person all ages and classes, and a part is the facial image pair of different people all ages and classes in addition.
(2) to the gradient face characteristic (GradientFaces) of an overall situation of each facial image extraction, the gradient face characteristic of every a pair of people's face asks poor, obtains a difference characteristic.
(3) facial image of same person all ages and classes to the difference characteristic that obtains as positive example, the facial image of different people all ages and classes to the difference characteristic that obtains as counter-example.Utilize support vector machine (Support Vector Machine, SVM) train classification models.
Train after the sorter, to the hamming code feature of all face extraction overall situations of candidate face image collection.
In the algorithm that reorders, by the mode of iteration, from the candidate face image collection, choose people's face at every turn, add the facial image reference set, method is as follows:
Each takes turns iteration, each facial image in the candidate face image collection is calculated the overall distance D of itself and facial image to be retrieved and facial image reference set facial image, the facial image that the D value is minimum is removed from the candidate face image collection, and adds the facial image reference set.D is expressed as:
D = - β · sign ( f ( Q , I ) ) + d ( Q , I ) + α · 1 | R | Σ i d ( R i , I ) ,
Wherein, Q is facial image to be retrieved, and I is any one facial image in the candidate face image collection, the distance of-β sign (f (Q, I))+d (Q, I) expression candidate face image collection image I and facial image Q to be retrieved,
Figure BDA00002558207700062
The distance of expression candidate face image collection image I and facial image reference set facial image, overall distance D is both sums; α, β is scale-up factor, the size of these two coefficients represents that algorithm stresses the distance of classification results or people's face hamming code of semantic classification item, as α during greater than β, algorithm stresses the distance of people's face hamming code, so that the impact of hamming code distance is greater than the classification results of semantic classification item, otherwise as β during greater than α, the classification results influence power of semantic classification item is greater than the hamming code distance, and actual value can be determined according to using needs, generally can get simultaneously 0.5.Sign (f (Q, I)) be the semantic classification item, f is the sorter of above-mentioned training, Q and I are extracted respectively gradient face characteristic and calculated difference feature, utilize the sorter judged result, sign (f (Q, I)) returns 1 expression Q and I is predicted as same person, and returning-1 expression is not same person.D (Q, I) is the distance of calculating the hamming code feature of Q and I.D (R i, I) the hamming code distance of i people's face in calculating I and the facial image reference set.| R| represents facial image reference set size.Above-mentioned every iteration of taking turns is to choose a facial image I from the candidate face image collection, the D value that it calculates is minimum in the candidate face image collection face images, this facial image is removed from the candidate face image collection, added the facial image reference set.
The present invention is the method that proposes for extensive face retrieval specially.The present invention has following characteristics: 1) preprocessing process positions the key point in the facial image, and according to key point alignment people face, alignment procedure makes the different people appearance clearer and more definite with the position corresponding relation of part; 2) utilize simultaneously part and the overall geometric properties of people's face, and designed quantization method separately, utilized the vision word list face of leting others have a look at; 3) utilize the vision word that face database is set up inverted index, inverted index can retrieve people's face candidate face image collection fast, avoids the linear search to database, is conducive to the expansion of extensive face database; 4) designed the algorithm that reorders, the algorithm that reorders has been considered the distance of distance, candidate face image collection and the facial image reference set of the semantic classification of people's face to be retrieved and candidate face image collection, people's face to be retrieved and candidate face image collection simultaneously, and comprehensive like this optimization makes the algorithm that reorders can obtain accurately ranking results.The present invention can be directly used in the quick-searching of extensive people's face data set of has age, attitude variation.
Beneficial effect: the present invention can carry out the alignment of facial image, set up the index structure of large scale database, improve the extendability of retrieval performance and database, in addition, the algorithm that reorders among the present invention has higher accuracy rate, can and obtain an accurate rank of successively decreasing with human face similarity degree to be retrieved to the ordering of candidate face image collection people face.Therefore extensive Research on face image retrieval has higher use value.
Description of drawings
Fig. 1 is process flow diagram of the present invention.
Fig. 2 is for utilizing training set to obtain local feature dictionary process flow diagram.
Fig. 3 is for utilizing training set to obtain overall geometric properties dictionary process flow diagram.
Fig. 4 face piecemeal synoptic diagram of behaving.
Fig. 5 face overall situation geometric properties synoptic diagram of behaving.
Fig. 6 is embodiment 2 synoptic diagram.
Embodiment:
As shown in Figure 1, the invention discloses a kind of extensive Research on face image retrieval, comprise following steps:
Face images in facial image to be retrieved and the face database is carried out respectively the processing of following steps 1 ~ step 4;
Step 1, the facial image pre-service: the key point in the facial image of location, and with facial image and the standard faces image alignment of presetting;
Step 2, extract local feature: according to the key point position facial image is divided into the p piece, each piece is called a localized mass, extracts the feature of each localized mass; The present invention generally is divided into 100.
Step 3 is extracted overall geometric properties: according to the key point position, extract facial image overall situation geometric properties, comprise distance feature, angle character and curvature feature;
Step 4 is quantified as the vision word with facial image local feature and overall geometric properties: utilize training set to obtain local feature dictionary and global characteristics dictionary, facial image local feature and overall geometric properties are quantified as the vision word;
Step 5 is set up inverted index to the face images in the face database;
Step 6 obtains the candidate face image collection according to the vision word of facial image to be retrieved from the inverted index of step 5;
Step 7 utilizes the algorithm that reorders that the candidate face image collection is reordered, and obtains the facial image reference set, and the facial image reference set is final orderly result for retrieval.
Step 1 specifically comprises the steps:
The locator key point positions the key point in the facial image, and the set of key point is as people's face shape;
Facial image and the standard faces image alignment of presetting are comprised for standard faces image a and any facial image b, its people's face shape is respectively X aAnd X b, alignment procedure is calculating parameter group (θ, s, t), makes E=(X a-M (s, θ) [X b]-t) TW (X a-M (s, θ) [X bThe minimum of]-t), wherein, M (s, θ) [X b]+t represents that M represents alignment function to the conversion of facial image b people face shape vector, and θ represents rotation parameter, and s represents zooming parameter, and t represents displacement parameter, and W is the diagonal matrix that the weights of each key point form;
According to parameter group (θ, s, t), facial image b is alignd with standard faces image a.
People's face method of partition as shown in Figure 4 in the step 2 extraction local feature.At first according to organ site, mainly comprise eyes, nose, cheek, people's face is divided into five bulks, Fig. 4 is (owing to the invention relates to the invention of picture, therefore must use gray scale pictures as synoptic diagram) people's five bulks in regional corresponding the method for grid on the face among rear five people's face figure, method is as follows: with all key points of left eye as a set, with 4 key points of upper and lower, left and right as the border, form a rectangle, this rectangle is considered to piece corresponding to left eye.The acquisition methods of all the other four pieces is identical with the method for obtaining the left eye piece.Then with every a bulk of 4 * 5 fritters that are equally divided into, each bulk is divided into 20 fritters again by further gridding among Fig. 4, between these 20 the overlapping region is arranged.People's face can be divided into 100 fritters, and each fritter has independently numbering, and Serial Number Range is 1 ~ 100.The feature of extracting each localized mass in the step 2 comprises local binary and two kinds of features of yardstick invariant features conversion.
Extract overall geometric properties in the step 3 as shown in Figure 5 (owing to the invention relates to the invention of picture, therefore must use gray scale pictures as synoptic diagram), step 3 is extracted overall geometric properties and is comprised that specifically the space geometry between the key point concerns, comprises distance feature, angle character and curvature feature.The 2nd people's face figure expression distance feature among Fig. 5, its middle conductor represents the distance between the key point.The 3rd subgraph represents angle character among Fig. 5, and wherein the angle of two line segment compositions is the angle in the algorithm.The 4th subgraph represents the curvature feature among Fig. 5, and curve is the curve after key point connects, and the curvature feature in the algorithm is this curvature of a curve.According to the key point of location, calculate in twos distance, will be spliced into vector apart from linearity, be called distance feature; Calculate the angle that forms angle between the key point, the angle linearity is spliced into vector, be called the angle vector.Key point to the chin position is calculated curvature, and the curvature linearity is spliced into vector, is called curvature vector.With distance vector, angle vector sum curvature vector linear mosaic, it is exactly the proper vector of overall geometric properties.The result who returns according to the key point location algorithm in the facial image, select 5 key points as left eye region geometric properties point, select 5 key points as right eye region geometric properties point, select 5 key points as nasal area geometric properties point, select 4 key points as face region geometry unique point, select 7 key points as chin area geometric properties point, 26 geometric properties points represent with S set altogether.With coupling together with line between these 26 geometric properties points, obtain altogether 39 line segments, calculate the length of these line segments, i.e. distance feature between the key point.According to each above-mentioned line segment distance, use the triangle cosine law, calculate the angle character between the line segment.Extract the contour feature of chin area, the radius-of-curvature that 5 geometric properties points of use chin area are linked to be curve correspondence position point incorporates overall geometric properties as contour feature, comes the differential of this point of approximate representation with the difference of discrete point.
Utilize training set to obtain the local feature dictionary in the step 4 and the global characteristics dictionary comprises the steps:
Definition vision word: the vision word comprises Name ID, Age ID, Gender ID, Position ID; Wherein, for local feature, Name ID represents the name of the affiliated people's face of localized mass, and Age ID represents the age bracket of the affiliated people's face of localized mass, and Gender ID represents the sex of the affiliated people's face of localized mass, and Position ID represents the numbering of localized mass in affiliated people's face; For global characteristics, Name ID represents the name of the corresponding people's face of global characteristics, and Age ID represents the age bracket of the corresponding people's face of global characteristics, and Gender ID represents the sex of the corresponding people's face of global characteristics, Position ID is made as fixed value p+q, any positive integer of q value;
Utilize training set to generate the local feature dictionary as shown in Figure 2, its process is as follows: to facial images all in the training set carry out step 1 ~ 2, and local binary, the yardstick invariant features of training set face images are changed two kinds of local features be included into separately a set, obtain two set, each set is called the local feature set; With the local feature of same age bracket, same sex and same numbering in the set of each local feature as one group, every group is utilized respectively the training of sparse coding model, obtain a local feature dictionary, this dictionary is comprised of a stack features vector, a local feature in the corresponding group of each proper vector is called base; Any one proper vector all is reconstructed by the linear combination of base, and the coefficient vector of linear combination is that the sparse coding of proper vector represents;
Utilize training set to generate global characteristics dictionary such as Fig. 3, its process is as follows: facial images all in the training set is carried out step 1 and step 3, as one group, each group is considered as a global characteristics dictionary with the overall geometric properties of same age bracket, same sex in the training set.
In the step 4 facial image local feature and overall geometric properties being quantified as the vision word comprises the steps:
To everyone face image local feature F lUtilize the baseline of the facial image local feature dictionary of its same characteristic features kind (local binary or the conversion of yardstick invariant features), same age section, identical sex and identical numbering to be combined into line reconstruction, obtain base corresponding to nonzero element in the coefficient vector; The name of the facial image of base is as facial image local feature F lThe Name ID of corresponding vision word, age bracket is as facial image local feature F lThe Age ID of corresponding vision word, sex is as facial image local feature F lThe Gender ID of corresponding vision word, numbering is as facial image local feature F lThe Position ID of corresponding vision word;
For people's face image overall geometric properties F g, select its same age section, other global characteristics dictionary of homogeny, with facial image overall situation geometric properties F gCarry out nearest neighbor search with the global characteristics in the dictionary, obtain the name of the facial image of arest neighbors in the dictionary as facial image overall situation geometric properties F gThe Name ID of corresponding vision word, age bracket is as facial image overall situation geometric properties F gThe Age ID of corresponding vision word, sex is as facial image overall situation geometric properties F gThe Gender ID of corresponding vision word, facial image overall situation geometric properties F gPosition ID be made as p+q, any positive integer of q value.
Step 5 comprises the steps: all unduplicated vision words are formed a vision set of letters, and each vision word is as an index entry in the vision set of letters, and each index entry and a facial image set that comprises this vision word are bound; The facial image that is quantized into the vision word in the face database is related with the vision word in the vision set of letters, form inverted index, after being about to the facial image quantification in the face database, facial image is joined in the index entry tabulation of its corresponding vision word.
Step 6 comprises the steps:
The vision word corresponding according to facial image to be retrieved, the described vision word of search from the vision set of letters of inverted index, the facial image of face database that will be related with it takes out, and the set that the face images of taking-up forms is called the candidate face image collection.
Step 7 comprises the steps:
According to facial image to be retrieved, from the candidate face image collection, select lineup's face image by process of iteration, add the facial image reference set, after iteration finished, final facial image reference set was the result for retrieval after reordering;
Each takes turns iteration, each facial image in the candidate face image collection is calculated the overall distance D of itself and facial image to be retrieved and facial image reference set facial image, the facial image that the D value is minimum is removed from the candidate face image collection, and adds the facial image reference set.Overall distance D adopts following formula to calculate:
D = - β · sign ( f ( Q , I ) ) + d ( Q , I ) + α · 1 | R | Σ i d ( R i , I ) ,
Wherein, Q is facial image to be retrieved, and I is any one facial image in the candidate face image collection, the distance of-β sign (f (Q, I))+d (Q, I) expression candidate face image collection image I and facial image Q to be retrieved,
Figure BDA00002558207700112
The distance of expression candidate face image collection image I and facial image reference set facial image, overall distance D is both sums.α, β are scale-up factor, and the span of α, β is 0 ~ 1, sign (f (Q, I)) be the semantic classification item, wherein f (Q, I) expression is carried out semantic classification to the facial image I in facial image Q to be retrieved and the candidate face image collection, judge whether it belongs to same class, f is classification function, and sign gets sign function, d (Q, I) distance of image I in facial image Q to be retrieved and the candidate face image collection is calculated in expression
Figure BDA00002558207700113
The mean distance of facial image I and facial image reference set in the set of expression calculated candidate facial image, | R| represents the size of facial image reference set, R iIn the expression current iteration in the facial image reference set i open facial image.
Embodiment 1
Present embodiment comprises following part:
1. the facial image key point is located:
Utilize active shape model (Active Shape Models, ASM) to the location of the key point in the facial image, this module mainly is divided into two steps: model training and key point search.
In the training pattern process, used principal component analysis (PCA) (principal component analysis, PCA).
Principal component analysis (PCA) is to focus on exploratory statistic analysis method on certain several overall target (major component) with being dispersed in one group of information on the variable, utilize the inner structure of major component descriptor data set, in fact play a part Data Dimensionality Reduction, several eigenvalue of maximum characteristic of correspondence vectors consisted of one group of base before the method was selected former data covariance matrix, to reach the best purpose that characterizes former data.Consider the vector x in the n-dimensional space, for dimensionality reduction, need to be with the vector x of a m dimension ' come approximate simulation, wherein m<n namely seeks a conversion f:R n→ R mThe conversion of the following form of PCA utilization:
y=W T(x-μ)
Wherein μ is expectation μ=E[x of random vector x], W=(w 1, w 2... w m) be the transformation matrix of a n * m.Suppose that x is zero-mean, then y i=w i TX.The target of PCA conversion is under above-mentioned constraint condition, seek so that | y i| 2Each maximum w i, namely wish farthest to keep the original difference of x.The objective function of above optimization problem is:
E i = Σ i = 1 D | y i | 2 D = Σ i = 1 D y i y i D = Σ i = 1 D w i T xx T w i D = w i T Cw i
Wherein D is number of training,
Figure BDA00002558207700122
Be x covariance matrix based on x 1, x 2..., x DValue.In order to ask the maximal value of objective function, use method of Lagrange multipliers,
0 = ∂ ( E i + λ i ( 1 - w i T w i ) ) ∂ w i = 2 Cw i - 2 λ i w i
Work as Cw i=λ w iThe time objective function get maximal value.Therefore, w iShould be taken as the proper vector of C, λ iFor with this proper vector characteristic of correspondence value.Be λ according to the eigenwert ordering 1〉=λ 2〉=... 〉=λ n, they are respectively w by the characteristic of correspondence vector 1, w 2... w n, then get top m vectorial w 1, w 2... w mForm the PCA transformation matrix.For the consideration of numerical evaluation aspect, usually use svd (SVD, Singular Value Decomposition) to calculate a front m major component.
Training pattern: key point is located main training of human face shape model and local gray-scale statistical model.Training set is people's face data set of mark key point,, all key points of people's face form people's face shape vector x i, it is a 2 * n-dimensional vector:
x i=(x i1,y i1,x i2,y i2,…,x in,y in) T,i=1,2,…,N
(x wherein Ij, y Ij) be the coordinate of j unique point on i the training sample.N is counting of people's face shape model, and N is the number of training sample.With the some combination of shapes in the training set, and by principal component analysis (PCA), training obtains shape subspace, and then people's face shape vector can be expressed as arbitrarily:
x = x ‾ + Pb
Wherein: P=(p 1, p 2... p t) be the matrix that front t proper vector that PCA obtains forms, be called the subspace, b=(b 1, b 2... b t) TBe any people's face at the projection coefficient of subspace, can obtain different people's face shape vectors by changing b, so far finished the foundation of people's face shape model.
For certain specific key point of people's face, it be similar that the feature around it distributes.In each key point of the every width of cloth facial image of people's face data centralization of mark key point, through this point, along getting the gray scale differential of k pixel on its point of proximity vertical direction and to its standardization, with this vector as local gray level information.When the local gray-scale statistical model of training, j key point of all N sample image extracted gray feature, be designated as g Ij, then to its standardization:
g ij ′ = g ij Σ t = 1 k | g ij ( k ) |
G wherein Ij(k) expression g IjK pixel grey scale differential.To j key point normalized feature of every people's face, obtain the training set of j key point, estimate average and covariance matrix with sample average and sample covariance.Sample average
Figure BDA00002558207700132
With the sample covariance matrix ∑ jBe calculated as follows:
l j ‾ = 1 N Σ i = 1 N g ij ′
Σ j = 1 N Σ i = 1 N ( g ij ′ - l j ‾ ) ( g ij ′ - l j ‾ ) T
Each key point is calculated such statistical model, and so far, the training of local gray level statistical model is finished.
Key point search: at first find human face region by the fast face detection algorithm, then in human face region, provide the initial position of key point, it is original shape, the center of general employment face shape model, namely the mean value of people's face data centralization people face shape vector of mark key point is as initial position.Behind the given initial position, find the accurate location of each key point by iterative algorithm, process is as follows:
1) searches for each crucial neighborhood of a point, find the optimum point of epicycle search.Calculate local gray level feature g at certain neighborhood point, by calculating mahalanobis distance:
f ( g ) = ( g - l j ‾ ) Σ j - 1 ( g - l j ‾ ) T
∑ wherein j -1It is ∑ jInverse matrix, the epicycle of thinking of mahalanobis distance minimum is searched for the position of optimum key point;
2) according to the displacement of all key points, by new shape of people's face shape model generation, replace original shape;
3) return step 1), until people's face shape vector changes less than threshold value T 1
Obtain the position of the key point in the facial image after search finishes, finished the key point location.
2. people's face alignment:
The corresponding relation of the key point position in the facial image is namely confirmed in alignment, and therefore, the present invention utilizes the people's face that aligns of the key point in the facial image of location, makes people's face position relationship keep farthest corresponding.Given two people's face shapes vector, i.e. the set of key point, the shape of i people's face is designated as:
X i=(x i1,y i1,x i2,y i2,…x in,y in) T
(x wherein Ij, y Ij) coordinate of expression i people's face j key point, j is the arbitrary numerical value among 1 ~ n, n is the key point number, the transposition of T representing matrix.Given facial image a and facial image b, its shape is respectively X a=(x A1, y A1, x A2, y A2... x An, y An) TAnd X b=(x B1, y B1, x B2, y B2... x Bn, y Bn) T, the purpose of alignment is to find one group of parameter (θ; S; T), make
E=(X a-M(s,θ)[X b]-t) TW(X a-M(s,θ)[X b]-t)
Minimum.Wherein θ represents the anglec of rotation, and s represents scaling, and t represents the displacement of translation, to conversion M (s, the θ) [X of people's face shape vector i]+t represents, wherein
M ( s , θ ) x ij y ij = ( s cos θ ) x ij - ( s sin θ ) y ij ( s sin θ ) x ij + ( s cos θ ) y ij
t=(t x,t y,…,t x,t y) T
W is the weights (w of each key point 1, w 2..., w n) diagonal matrix that forms, t x, t yBe respectively key point x, the displacement that the y coordinate is corresponding.In minimizing the process of E, each parameter differentiated can obtain the result, according to the parameter that obtains facial image is changed, and makes two people's faces keep maximum position corresponding, and process is as follows:
Make scos θ=a x, ssin θ=a y, to a x, a y, t x, t yDifferentiate separately, make E minimum, obtain following four equatioies:
B x a x - B y a y + Wt x = A x B y a x + B x a y + Wt y = A y Za x + B x t x + B y t y = C 1 Za y - B y t x + B x t y = C 2
Wherein A x = Σ k = 1 n w k x ak , B x = Σ k = 1 n w k x bk , A y = Σ k = 1 n w k y ak , B y = Σ k = 1 n w k x bk ,
Z = Σ k = 1 n w k ( x bk 2 + y bk 2 ) , W = Σ k = 1 n w k , C 1 = Σ k = 0 n w k ( x ak x bk + y ak y bk ) , C 2 = Σ k = 0 n w k ( y ak x bk - x ak y bk ) . Calculate the transformation parameter group (θ that to obtain aliging by solving an equation; S; T), again image is changed.
3. extraction local feature
According to the key point in the facial image of step 1 location, people's face is carried out piecemeal.At first according to organ site, mainly comprise eyes, nose, cheek, people's face is divided into five bulks.Then with every a bulk of 4 * 5 fritters that are equally divided into, between these 20 the overlapping region is arranged.People's face can be divided into 100 fritters, and each fritter has independently numbering, and Serial Number Range is 1-100.At last, each piece is extracted local binary (Local Binary Pattern, LBP) and yardstick invariant features conversion (Scale Invariant Feature Transform, SIFT) two kinds of features.
4. extract overall geometric properties
According to the key point in the facial image of step 1 location, extract the overall geometric properties between the key point, overall geometric properties is described the space geometry relation between the key point, comprises distance feature, angle character, curvature feature.According to the key point of location, calculate in twos distance, will be spliced into vector apart from linearity, be called distance feature; Calculate the angle that forms angle between the key point, the angle linearity is spliced into vector, be called the angle vector.Key point to the chin position is calculated curvature, and the curvature linearity is spliced into vector, is called curvature vector.With distance vector, angle vector sum curvature vector linear mosaic, it is exactly the proper vector of overall geometric properties.Particularly:
1) result who returns according to the key point location algorithm in the facial image, select 5 key points as left eye region geometric properties point, select 5 key points as right eye region geometric properties point, select 5 key points as nasal area geometric properties point, select 4 key points as face region geometry unique point, select 7 key points as chin area geometric properties point, 26 geometric properties points represent with S set altogether.To couple together with line between these 26 geometric properties points, obtain altogether 39 line segments, calculate the length of these line segments, it is distance between the key point, here with the distance between two as gauged distance d', the impact that the distance of other calculating causes with the removal of images yardstick divided by this gauged distance, distance adopts European tolerance, as shown in the formula:
D = { d ij | d ij = | | p i - p j | | 2 d ′ , p i , p j ∈ S }
Wherein, p iI point among the expression S, d IjThe distance that represents i point and j point.Have comparability during for the quantification treatment of follow-up overall geometric properties, the eigenwert of adjusting the distance is carried out normalization:
D norm = d n | D | , | D | = Σ n = 1 39 d n 2 , d n ∈ D dn∈D
D wherein NormBe the vector of the distance feature after the normalization.
2) according to each above-mentioned line segment distance, use the triangle cosine law, calculate the angle between the line segment:
θ = { θ ikj | θ ikj = arccos ( d ik 2 + d kj 2 - d ij 2 2 d ik · d kj ) , p i , p k , p j ∈ S }
θ wherein IkjRepresent i, k, a j angle angle that point forms.Then, these angle character values are carried out normalization:
θ norm = θ n | θ | , | θ | = Σ n = 1 15 θ n 2 , θ n ∈ θ θn∈θ
3) extract the contour feature of chin area, the radius-of-curvature of using 5 geometric properties points of chin area to be linked to be curve correspondence position point here incorporates overall geometric properties as contour feature, comes the differential of this point of approximate representation with the difference of discrete point, that is:
ρ = { ρ i | ρ i = | ( 1 + y i ′ 2 ) 3 2 y i ′ ′ | , p i ∈ S }
Y wherein i' be the first order derivative at unique point i place, use the first order difference approximate representation, y i" be the second derivative at unique point i place, use the second order difference approximate representation.Equally these 5 radius-of-curvature eigenwerts are carried out normalization, that is:
ρ norm = ρ n | ρ | , | ρ | = Σ n = 1 5 ρ n 2 , ρ n ∈ ρ ρn∈ρ
5. define the vision word
In a kind of extensive face retrieval method, definition vision word is<Name ID Age ID, Gender ID, Position ID 〉, each part or overall geometric properties all are quantified as one or more vision words.In training set, each facial image has its name, age, sex information separately.Wherein Name ID represents the name of people's face in the corresponding facial image training set of this feature, Age ID represents the age bracket of people's face in the corresponding facial image training set of this feature, age bracket is divided, be divided into 8 sections, be respectively 1-15,15-22,23-30,31-38,39-46,47-54, more than the 55-62 and 62, everyone face is divided into an age bracket.Gender ID represents the sex of people's face in the corresponding facial image training set of this feature, and Position ID represents block number corresponding to facial image in the corresponding facial image training set of this feature.Especially, for local feature, because people's face is divided into 100 fritters, so the Position ID of the vision word that is quantized into of local feature is the numbering of its corresponding people's face piece, and for overall geometric properties, Position ID is unified to be set to 101.
6. local feature quantizes
Divide block number with everyone face of training set by step 2, every is extracted respectively local feature.All proper vectors of same feature kind (local binary or the conversion of yardstick invariant features), age-grade, same sex, same numbering are classified as one group.Every stack features utilizes one group of base of sparse coding model training.
min D , V Σ i = 1 n | | x i - Dv i | | 2 + λ | | v i | | 1
s . t . | | v i | | ≤ 1 , ∀ i
Wherein, the dictionary that D arrives for study, x iBe the proper vector of every group of training set, v iBe x iCoefficient when carrying out linear reconstruction with D.λ is Sparse parameter.Can calculate dictionary D by training set, can train for every group and obtain a dictionary D.When retrieval, the age bracket of a given proper vector x' and its affiliated people's face, sex and minute block number are chosen corresponding dictionary D', with dictionary x' are reconstructed, and obtain reconstruction coefficients v', and at this moment, local feature x' just is encoded as v'.The name of people's face under the corresponding base of non-zero entry is the Name ID after x' quantizes among the v'.The age bracket of x', sex and piece label are respectively Age ID, Gender ID and the Position ID after the quantification.In practice, be not to choose all non-zero entry to obtain the vision word, but select 3 elements of coefficient maximum to obtain the vision word.
7. overall geometric properties quantizes
To all training set face extraction overall situation geometric properties, with people face age bracket, the sex grouping of overall geometric properties according to correspondence, other overall geometric properties of same age bracket homogeneity is divided into one group.
A given overall geometric properties F gAnd age bracket, sex information, from training set, select the feature group of corresponding age bracket, sex, find F by neighbor search g3 arest neighbors, neighbour's identity, age bracket, sex are F gQuantification after Name ID, Age ID, Gender ID, the non-Position ID of overall geometric properties is unified to be set to 101.
Step 7 is set up inverted index and is comprised:
Facial images all in the database is set up inverted index, be used for retrieval.For a facial image, quantize according to above-mentioned process, obtain representing the vision word of this people's face, this facial image is indexed on the vision word.Everyone face of database is quantized, index respectively vision word separately.Index structure is the set of vision word, by given vision word, can retrieve the facial image set corresponding with it.
8. obtain the candidate face image collection
A given facial image to be retrieved, obtain its vision word by above-mentioned quantization method, find identical vision word from inverted index, facial image that will be corresponding with it takes out, these facial images have formed the candidate face image collection, and the candidate face image collection is the preliminary search result.
9. reorder
At first utilize a part of facial image training set, training of human face sorter, this sorter are used for two facial images of prediction and whether belong to same person.Concrete steps are as follows:
(1) the sorter training set is some facial images pair, and wherein a part is the facial image of same person all ages and classes, and a part is the facial image of different people all ages and classes in addition.
(2) to the gradient face characteristic (GradientFaces) of an overall situation of each facial image extraction, the gradient face characteristic of every a pair of people's face asks poor, obtains a difference characteristic.
(3) facial image of same person all ages and classes to the difference characteristic that obtains as positive example, the facial image of different people all ages and classes to the difference characteristic that obtains as counter-example.Utilize support vector machine (SupportVector Machine, SVM) train classification models.
Train after the sorter, to the hamming code feature of all face extraction overall situations of candidate face image collection.
In the algorithm that reorders, by the mode of iteration, from the candidate face image collection, choose people's face at every turn, add the facial image reference set, method is as follows:
Each takes turns iteration, each facial image in the candidate face image collection is calculated the overall distance D of itself and facial image to be retrieved and facial image reference set facial image, the facial image that the D value is minimum is removed from the candidate face image collection, and adds the facial image reference set.D is expressed as:
D = - β · sign ( f ( Q , I ) ) + d ( Q , I ) + α · 1 | R | Σ i d ( R i , I ) ,
Wherein, Q is facial image to be retrieved, and I is any one facial image in the candidate face image collection, the distance of-β sign (f (Q, I))+d (Q, I) expression candidate face image collection image I and facial image Q to be retrieved,
Figure BDA00002558207700182
The distance of expression candidate face image collection image I and facial image reference set facial image, overall distance D is both sums; α, β is scale-up factor, sign (f (Q, I)) be the semantic classification item, f is the sorter of above-mentioned training, and Q and I are extracted respectively gradient face characteristic and calculated difference feature, utilize sorter to predict the outcome, sign (f (Q, I)) returns 1 expression Q and I is predicted as same person, and returning-1 expression is not same person.D (Q, I) is the distance of calculating the hamming code feature of Q and I.D (R i, I) the hamming code distance of i people's face in calculating I and the facial image reference set.| R| represents facial image reference set size.
Each is taken turns in the iteration, at first candidate face image collection face images is calculated its corresponding D value, selects that facial image of D value minimum, he is removed from the candidate face image collection and adds the facial image reference set.After taking turns iteration through 100, choose altogether 100 candidate face image collection people faces and added the facial image reference set, and these 100 people's faces are to sort according to the sequencing of selecting in the facial image reference set, the facial image reference set is as final result for retrieval, be 100 facial images that retrieve in the database, these 100 images sort from big to small by database images and image similarity to be retrieved.Can revise the final size of facial image reference set according to the real needs of return results number in the reality.
Embodiment 2
Fig. 6 is embodiment 2 retrieval synoptic diagram, and the character image source is public LFW database among the figure.1 is original facial image among the figure, frame table among the figure face piecemeal of leting others have a look at, the feature that 2 expressions are extracted comprises that local feature and the overall situation are extraordinary, the dictionary of 3 expression training, the 4th, the preliminary search result who obtains according to dictionary, i.e. candidate face image collection, the 5th, facial image reference set, according to the facial image reference set, the result reorders to preliminary search, obtains 6, i.e. final result for retrieval.People's face in people's face and 1 in 6 is same person, and expression is retrieved successfully.
The invention provides a kind of extensive Research on face image retrieval; method and the approach of this technical scheme of specific implementation are a lot; the above only is preferred implementation of the present invention; should be understood that; for those skilled in the art; under the prerequisite that does not break away from the principle of the invention, can also make some improvements and modifications, these improvements and modifications also should be considered as protection scope of the present invention.In the present embodiment not clear and definite each ingredient all available prior art realized.

Claims (9)

1. an extensive Research on face image retrieval is characterized in that, comprises following steps:
Face images in facial image to be retrieved and the face database is carried out respectively the processing of following steps 1 ~ step 4;
Step 1, the facial image pre-service: the key point in the facial image of location, and with facial image and the standard faces image alignment of presetting;
Step 2, extract local feature: according to the key point position facial image is divided into the p piece, each piece is called a localized mass, extracts the feature of each localized mass;
Step 3 is extracted overall geometric properties: according to the key point position, extract facial image overall situation geometric properties, comprise distance feature, angle character and curvature feature;
Step 4 is quantified as the vision word with facial image local feature and overall geometric properties: utilize training set to obtain local feature dictionary and global characteristics dictionary, facial image local feature and overall geometric properties are quantified as the vision word;
Step 5 is set up inverted index to the face images in the face database;
Step 6 obtains the candidate face image collection according to the vision word of facial image to be retrieved from the inverted index of step 5;
Step 7 utilizes the algorithm that reorders that the candidate face image collection is reordered, and obtains the facial image reference set, and the facial image reference set is final orderly result for retrieval.
2. a kind of extensive face retrieval method according to claim 1 is characterized in that step 1 specifically comprises the steps:
The locator key point positions the key point in the facial image, and the set of key point is as people's face shape;
Facial image and the standard faces image alignment of presetting are comprised for standard faces image a and any facial image b, its people's face shape is respectively X aAnd X b, alignment procedure is calculating parameter group (θ, s, t), makes E=(X a-M (s, θ) [X b]-t) TW (X a-M (s, θ) [X bThe minimum of]-t), wherein, M (s, θ) [X b]+t represents that M represents alignment function to the conversion of facial image b people face shape vector, and θ represents rotation parameter, and s represents zooming parameter, and t represents displacement parameter, and W is the diagonal matrix that the weights of each key point form;
According to parameter group (θ, s, t), facial image b is alignd with standard faces image a.
3. a kind of extensive face retrieval method according to claim 2 is characterized in that, the feature of extracting each localized mass in the step 2 comprises local binary and two kinds of features of yardstick invariant features conversion.
4. a kind of extensive face retrieval method according to claim 3 is characterized in that, step 3 is extracted overall geometric properties and comprised that specifically the space geometry between the key point concerns, comprises distance feature, angle character and curvature feature.
5. a kind of extensive face retrieval method according to claim 4 is characterized in that, utilizes training set to obtain the local feature dictionary in the step 4 and the global characteristics dictionary comprises the steps:
Definition vision word: the vision word comprises Name ID, Age ID, Gender ID, Position ID; Wherein, for local feature, Name ID represents the name of the affiliated people's face of localized mass, and Age ID represents the age bracket of the affiliated people's face of localized mass, and Gender ID represents the sex of the affiliated people's face of localized mass, and Position ID represents the numbering of localized mass in affiliated people's face; For global characteristics, Name ID represents the name of the corresponding people's face of global characteristics, and Age ID represents the age bracket of the corresponding people's face of global characteristics, and Gender ID represents the sex of the corresponding people's face of global characteristics, Position ID is made as fixed value p+q, any positive integer of q value;
It is as follows to utilize training set to generate local feature dictionary process: to facial images all in the training set carry out step 1 ~ 2, and local binary, the yardstick invariant features of training set face images are changed two kinds of features be included into separately a set, each set is called the local feature set; With the local feature of same age bracket, same sex and same numbering in the set of each local feature as one group, every group is utilized respectively the training of sparse coding model, obtain a local feature dictionary, this dictionary is comprised of a stack features vector, a local feature in the corresponding group of each proper vector is called base; Any one proper vector all is reconstructed by the linear combination of base, and the coefficient vector of linear combination is that the sparse coding of proper vector represents;
It is as follows to utilize training set to generate global characteristics dictionary process: facial images all in the training set is carried out step 1 and step 3, as one group, each group is considered as a global characteristics dictionary with the overall geometric properties of same age bracket, same sex in the training set.
6. a kind of extensive face retrieval method according to claim 5 is characterized in that, in the step 4 facial image local feature and overall geometric properties is quantified as the vision word and comprises the steps:
To everyone face image local feature F l, be combined into line reconstruction according to the baseline of the facial image local feature dictionary of same characteristic features kind, same age section, identical sex and identical numbering, obtain base corresponding to nonzero element in the coefficient vector; The name of the facial image of base is as facial image local feature F lThe Name ID of corresponding vision word, age bracket is as facial image local feature F lThe Age ID of corresponding vision word, sex is as facial image local feature F lThe Gender ID of corresponding vision word, numbering is as facial image local feature F lThe Position ID of corresponding vision word;
For people's face image overall geometric properties F g, select its same age section, other global characteristics dictionary of homogeny, with facial image overall situation geometric properties F gCarry out nearest neighbor search with the global characteristics in the dictionary, obtain the name of the facial image of arest neighbors in the dictionary as facial image overall situation geometric properties F gThe Name ID of corresponding vision word, age bracket is as facial image overall situation geometric properties F gThe Age ID of corresponding vision word, sex is as facial image overall situation geometric properties F gThe Gender ID of corresponding vision word, facial image overall situation geometric properties F gPosition ID be made as p+q, any positive integer of q value.
7. a kind of extensive face retrieval method according to claim 6, it is characterized in that, step 5 comprises the steps: all unduplicated vision words are formed a vision set of letters, each vision word is as an index entry in the vision set of letters, and each index entry and a facial image set that comprises this vision word are bound; The facial image that is quantized into the vision word in the face database is related with the vision word in the vision set of letters, form inverted index, after being about to the facial image quantification in the face database, facial image is joined in the index entry tabulation of its corresponding vision word.
8. a kind of extensive face retrieval method according to claim 7 is characterized in that step 6 comprises the steps:
The vision word corresponding according to facial image to be retrieved, the described vision word of search from the vision set of letters of inverted index, the facial image of face database that will be related with it takes out, and the set that the face images of taking-up forms is called the candidate face image collection.
9. a kind of extensive face retrieval method according to claim 8 is characterized in that step 7 comprises the steps:
According to facial image to be retrieved, from the candidate face image collection, select lineup's face image by process of iteration, add the facial image reference set, after iteration finished, final facial image reference set was the result for retrieval after reordering;
Each takes turns iteration, each facial image in the candidate face image collection is calculated the overall distance D of facial image in itself and facial image to be retrieved and the facial image reference set, the facial image that overall distance D value is minimum is removed from the candidate face image collection, and add the facial image reference set, overall distance D adopts following formula to calculate:
D = - β · sign ( f ( Q , I ) ) + d ( Q , I ) + α · 1 | R | Σ i d ( R i , I ) ,
Wherein, Q is facial image to be retrieved, and I is any one facial image in the candidate face image collection, the distance of facial image I and facial image Q to be retrieved in-β sign (f (Q, I))+d (Q, I) expression candidate face image collection,
Figure FDA00002558207600041
The distance of facial image in facial image I and the facial image reference set in the expression candidate face image collection, overall distance D is both sums; α, β are scale-up factor, and the span of α, β is 0 ~ 1, sign (f (Q, I)) be the semantic classification item, wherein f (Q, I) expression is carried out semantic classification to the facial image I in facial image Q to be retrieved and the candidate face image collection, judge whether it belongs to same class, f is classification function, and sign gets sign function, d (Q, I) distance of image I in facial image Q to be retrieved and the candidate face image collection is calculated in expression The mean distance of facial image I and facial image reference set in the set of expression calculated candidate facial image, | R| represents the size of facial image reference set, R iIn the expression current iteration in the facial image reference set i open facial image.
CN201210527836.8A 2012-12-10 2012-12-10 Large-scale human face image searching method Expired - Fee Related CN102982165B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201210527836.8A CN102982165B (en) 2012-12-10 2012-12-10 Large-scale human face image searching method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201210527836.8A CN102982165B (en) 2012-12-10 2012-12-10 Large-scale human face image searching method

Publications (2)

Publication Number Publication Date
CN102982165A true CN102982165A (en) 2013-03-20
CN102982165B CN102982165B (en) 2015-05-13

Family

ID=47856182

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201210527836.8A Expired - Fee Related CN102982165B (en) 2012-12-10 2012-12-10 Large-scale human face image searching method

Country Status (1)

Country Link
CN (1) CN102982165B (en)

Cited By (38)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103294813A (en) * 2013-06-07 2013-09-11 北京捷成世纪科技股份有限公司 Sensitive image search method and device
CN103514254A (en) * 2013-07-04 2014-01-15 李文博 Image set ordering method for mining hidden operation behavior
CN104036012A (en) * 2014-06-24 2014-09-10 中国科学院计算技术研究所 Dictionary learning method, visual word bag characteristic extracting method and retrieval system
CN104166685A (en) * 2014-07-24 2014-11-26 北京捷成世纪科技股份有限公司 Video clip detecting method and device
CN104574299A (en) * 2014-12-25 2015-04-29 小米科技有限责任公司 Face picture processing method and device
CN104714633A (en) * 2013-12-12 2015-06-17 华为技术有限公司 Method and terminal for terminal configuration
CN104765764A (en) * 2015-02-06 2015-07-08 南京理工大学 Indexing method based on large-scale image
CN104915351A (en) * 2014-03-12 2015-09-16 华为技术有限公司 Picture sorting method and terminal
CN105930834A (en) * 2016-07-01 2016-09-07 北京邮电大学 Face identification method and apparatus based on spherical hashing binary coding
CN106126579A (en) * 2016-06-17 2016-11-16 北京市商汤科技开发有限公司 Object identification method and device, data processing equipment and terminal unit
CN106250818A (en) * 2016-07-19 2016-12-21 南京大学 A kind of total order keeps the face age estimation method of projection
CN103810303B (en) * 2014-03-18 2017-01-18 苏州大学 Image search method and system based on focus object recognition and theme semantics
CN106777167A (en) * 2016-12-21 2017-05-31 中国科学院上海高等研究院 Magnanimity Face Image Retrieval System and search method based on Spark frameworks
CN107423294A (en) * 2016-02-25 2017-12-01 北京联合大学 A kind of community image search method and system
CN107506559A (en) * 2017-09-08 2017-12-22 廖海斌 Star's face shaping based on human face similarity degree analysis, which is made up, recommends method and apparatus
CN107679235A (en) * 2017-10-25 2018-02-09 成都尽知致远科技有限公司 searching system based on cloud platform
CN107798093A (en) * 2017-10-25 2018-03-13 成都尽知致远科技有限公司 Image search method
CN108009532A (en) * 2017-12-28 2018-05-08 盎锐(上海)信息科技有限公司 Personal identification method and terminal based on 3D imagings
CN108352068A (en) * 2015-11-18 2018-07-31 皇家飞利浦有限公司 Method and apparatus for the absolute dimension size for estimating test object
CN108400972A (en) * 2018-01-30 2018-08-14 北京兰云科技有限公司 A kind of method for detecting abnormality and device
CN108491773A (en) * 2018-03-12 2018-09-04 中国工商银行股份有限公司 A kind of recognition methods and system
CN108595927A (en) * 2018-04-04 2018-09-28 北京市商汤科技开发有限公司 Authentication, unlock and method of payment, device, storage medium, product and equipment
CN108921065A (en) * 2018-06-21 2018-11-30 北京陌上花科技有限公司 The method and apparatus for establishing property data base
CN109241327A (en) * 2017-07-03 2019-01-18 北大方正集团有限公司 Image search method and device
CN109284729A (en) * 2018-10-08 2019-01-29 北京影谱科技股份有限公司 Method, apparatus and medium based on video acquisition human face recognition model training data
CN109376741A (en) * 2018-09-10 2019-02-22 平安科技(深圳)有限公司 Recognition methods, device, computer equipment and the storage medium of trademark infringement
CN109878290A (en) * 2019-03-13 2019-06-14 的卢技术有限公司 A kind of intelligent vehicle-carried air conditioning control method and system based on AI technology
CN109902190A (en) * 2019-03-04 2019-06-18 京东方科技集团股份有限公司 Image encrypting algorithm optimization method, search method, device, system and medium
CN109993039A (en) * 2018-01-02 2019-07-09 上海银晨智能识别科技有限公司 Portrait identification method and device, computer readable storage medium
CN110160528A (en) * 2019-05-30 2019-08-23 华中科技大学 A kind of mobile device pose localization method based on angle character identification
CN110458224A (en) * 2019-08-06 2019-11-15 北京字节跳动网络技术有限公司 Image processing method, device, electronic equipment and computer-readable medium
CN111008620A (en) * 2020-03-05 2020-04-14 支付宝(杭州)信息技术有限公司 Target user identification method and device, storage medium and electronic equipment
CN111738194A (en) * 2020-06-29 2020-10-02 深圳力维智联技术有限公司 Evaluation method and device for similarity of face images
WO2021056440A1 (en) * 2019-09-27 2021-04-01 京东方科技集团股份有限公司 Method and apparatus for image search, and storage medium
CN112634125A (en) * 2020-12-16 2021-04-09 西安理工大学 Automatic face replacement method based on off-line face database
CN113159211A (en) * 2021-04-30 2021-07-23 杭州好安供应链管理有限公司 Method, computing device and computer storage medium for similar image retrieval
CN113254909A (en) * 2021-06-04 2021-08-13 宁波均联智行科技股份有限公司 Vehicle-mounted user account login method and system
CN114241559A (en) * 2021-12-16 2022-03-25 中国农业银行股份有限公司 Face recognition method, device, equipment and storage medium

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110299743A1 (en) * 2010-06-03 2011-12-08 Microsoft Corporation Scalable face image retrieval

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110299743A1 (en) * 2010-06-03 2011-12-08 Microsoft Corporation Scalable face image retrieval

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
T.F.COOTES等: "Active Shape Models-Their Training and Application", 《COMPUTER VISION AND IMAGE UNDERSTANDING》, vol. 61, no. 1, 31 January 1995 (1995-01-31), pages 38 - 59 *
ZHONG WU 等: ""Scalable Face Image Retrieval with Identity-Based Quantization and Multireference Reranking"", 《IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE》, vol. 33, no. 10, 31 October 2011 (2011-10-31), pages 1991 - 2001, XP011409102, DOI: 10.1109/TPAMI.2011.111 *
吴中: ""大规模图像搜索中的图像表示模型"", 《中国博士学位论文全文数据库》, 31 May 2012 (2012-05-31) *
钱洪波: ""基于局部系数量化的大规模人脸检索关键技术研究"", 《中国优秀硕士学位论文全文数据库》, 31 December 2014 (2014-12-31), pages 138 - 292 *

Cited By (55)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103294813A (en) * 2013-06-07 2013-09-11 北京捷成世纪科技股份有限公司 Sensitive image search method and device
CN103514254A (en) * 2013-07-04 2014-01-15 李文博 Image set ordering method for mining hidden operation behavior
CN104714633A (en) * 2013-12-12 2015-06-17 华为技术有限公司 Method and terminal for terminal configuration
US9977951B2 (en) 2014-03-12 2018-05-22 Huawei Technologies Co., Ltd. Picture ranking method, and terminal
CN104915351B (en) * 2014-03-12 2018-08-21 华为技术有限公司 Picture sort method and terminal
CN104915351A (en) * 2014-03-12 2015-09-16 华为技术有限公司 Picture sorting method and terminal
US10242250B2 (en) 2014-03-12 2019-03-26 Huawei Technologies Co., Ltd. Picture ranking method, and terminal
US10755086B2 (en) 2014-03-12 2020-08-25 Huawei Technologies Co., Ltd. Picture ranking method, and terminal
CN103810303B (en) * 2014-03-18 2017-01-18 苏州大学 Image search method and system based on focus object recognition and theme semantics
CN104036012A (en) * 2014-06-24 2014-09-10 中国科学院计算技术研究所 Dictionary learning method, visual word bag characteristic extracting method and retrieval system
CN104036012B (en) * 2014-06-24 2017-06-30 中国科学院计算技术研究所 Dictionary learning, vision bag of words feature extracting method and searching system
CN104166685A (en) * 2014-07-24 2014-11-26 北京捷成世纪科技股份有限公司 Video clip detecting method and device
CN104166685B (en) * 2014-07-24 2017-07-11 北京捷成世纪科技股份有限公司 A kind of method and apparatus for detecting video segment
CN104574299A (en) * 2014-12-25 2015-04-29 小米科技有限责任公司 Face picture processing method and device
CN104765764A (en) * 2015-02-06 2015-07-08 南京理工大学 Indexing method based on large-scale image
CN108352068A (en) * 2015-11-18 2018-07-31 皇家飞利浦有限公司 Method and apparatus for the absolute dimension size for estimating test object
CN107423294A (en) * 2016-02-25 2017-12-01 北京联合大学 A kind of community image search method and system
CN106126579A (en) * 2016-06-17 2016-11-16 北京市商汤科技开发有限公司 Object identification method and device, data processing equipment and terminal unit
CN105930834B (en) * 2016-07-01 2019-04-09 北京邮电大学 Face identification method and device based on ball Hash binary-coding
CN105930834A (en) * 2016-07-01 2016-09-07 北京邮电大学 Face identification method and apparatus based on spherical hashing binary coding
CN106250818A (en) * 2016-07-19 2016-12-21 南京大学 A kind of total order keeps the face age estimation method of projection
CN106250818B (en) * 2016-07-19 2019-03-26 南京大学 A kind of total order keeps the face age estimation method of projection
CN106777167B (en) * 2016-12-21 2020-05-12 中国科学院上海高等研究院 Massive human face image retrieval system and retrieval method based on Spark framework
CN106777167A (en) * 2016-12-21 2017-05-31 中国科学院上海高等研究院 Magnanimity Face Image Retrieval System and search method based on Spark frameworks
CN109241327A (en) * 2017-07-03 2019-01-18 北大方正集团有限公司 Image search method and device
CN107506559B (en) * 2017-09-08 2021-03-23 廖海斌 Star face shaping makeup recommendation method and device based on face similarity analysis
CN107506559A (en) * 2017-09-08 2017-12-22 廖海斌 Star's face shaping based on human face similarity degree analysis, which is made up, recommends method and apparatus
CN107798093A (en) * 2017-10-25 2018-03-13 成都尽知致远科技有限公司 Image search method
CN107679235A (en) * 2017-10-25 2018-02-09 成都尽知致远科技有限公司 searching system based on cloud platform
CN107798093B (en) * 2017-10-25 2022-05-03 成都尽知致远科技有限公司 Image retrieval method
CN107679235B (en) * 2017-10-25 2021-09-21 成都尽知致远科技有限公司 Retrieval system based on cloud platform
CN108009532A (en) * 2017-12-28 2018-05-08 盎锐(上海)信息科技有限公司 Personal identification method and terminal based on 3D imagings
CN109993039A (en) * 2018-01-02 2019-07-09 上海银晨智能识别科技有限公司 Portrait identification method and device, computer readable storage medium
CN108400972A (en) * 2018-01-30 2018-08-14 北京兰云科技有限公司 A kind of method for detecting abnormality and device
CN108491773A (en) * 2018-03-12 2018-09-04 中国工商银行股份有限公司 A kind of recognition methods and system
CN108595927A (en) * 2018-04-04 2018-09-28 北京市商汤科技开发有限公司 Authentication, unlock and method of payment, device, storage medium, product and equipment
CN108595927B (en) * 2018-04-04 2023-09-19 北京市商汤科技开发有限公司 Identity authentication, unlocking and payment method and device, storage medium, product and equipment
CN108921065A (en) * 2018-06-21 2018-11-30 北京陌上花科技有限公司 The method and apparatus for establishing property data base
CN109376741A (en) * 2018-09-10 2019-02-22 平安科技(深圳)有限公司 Recognition methods, device, computer equipment and the storage medium of trademark infringement
CN109284729A (en) * 2018-10-08 2019-01-29 北京影谱科技股份有限公司 Method, apparatus and medium based on video acquisition human face recognition model training data
CN109902190A (en) * 2019-03-04 2019-06-18 京东方科技集团股份有限公司 Image encrypting algorithm optimization method, search method, device, system and medium
CN109902190B (en) * 2019-03-04 2021-04-27 京东方科技集团股份有限公司 Image retrieval model optimization method, retrieval method, device, system and medium
CN109878290A (en) * 2019-03-13 2019-06-14 的卢技术有限公司 A kind of intelligent vehicle-carried air conditioning control method and system based on AI technology
CN110160528A (en) * 2019-05-30 2019-08-23 华中科技大学 A kind of mobile device pose localization method based on angle character identification
CN110458224A (en) * 2019-08-06 2019-11-15 北京字节跳动网络技术有限公司 Image processing method, device, electronic equipment and computer-readable medium
WO2021056440A1 (en) * 2019-09-27 2021-04-01 京东方科技集团股份有限公司 Method and apparatus for image search, and storage medium
US11803585B2 (en) 2019-09-27 2023-10-31 Boe Technology Group Co., Ltd. Method and apparatus for searching for an image and related storage medium
CN111008620A (en) * 2020-03-05 2020-04-14 支付宝(杭州)信息技术有限公司 Target user identification method and device, storage medium and electronic equipment
CN111738194A (en) * 2020-06-29 2020-10-02 深圳力维智联技术有限公司 Evaluation method and device for similarity of face images
CN111738194B (en) * 2020-06-29 2024-02-02 深圳力维智联技术有限公司 Method and device for evaluating similarity of face images
CN112634125A (en) * 2020-12-16 2021-04-09 西安理工大学 Automatic face replacement method based on off-line face database
CN112634125B (en) * 2020-12-16 2023-07-25 西安理工大学 Automatic face replacement method based on off-line face database
CN113159211A (en) * 2021-04-30 2021-07-23 杭州好安供应链管理有限公司 Method, computing device and computer storage medium for similar image retrieval
CN113254909A (en) * 2021-06-04 2021-08-13 宁波均联智行科技股份有限公司 Vehicle-mounted user account login method and system
CN114241559A (en) * 2021-12-16 2022-03-25 中国农业银行股份有限公司 Face recognition method, device, equipment and storage medium

Also Published As

Publication number Publication date
CN102982165B (en) 2015-05-13

Similar Documents

Publication Publication Date Title
CN102982165B (en) Large-scale human face image searching method
Zhang et al. Actively learning human gaze shifting paths for semantics-aware photo cropping
Xie et al. Deep multimetric learning for shape-based 3D model retrieval
CN105138972A (en) Face authentication method and device
CN105718960A (en) Image ordering model based on convolutional neural network and spatial pyramid matching
CN101551809A (en) Search method of SAR images classified based on Gauss hybrid model
CN105184298A (en) Image classification method through fast and locality-constrained low-rank coding process
CN104915658B (en) A kind of emotion component analyzing method and its system based on emotion Distributed learning
CN105574475A (en) Common vector dictionary based sparse representation classification method
JP7373624B2 (en) Method and apparatus for fine-grained image classification based on scores of image blocks
CN104077742B (en) Human face sketch synthetic method and system based on Gabor characteristic
Kastaniotis et al. HEp-2 cell classification with vector of hierarchically aggregated residuals
CN104281835A (en) Face recognition method based on local sensitive kernel sparse representation
Xie et al. Deep nonlinear metric learning for 3-D shape retrieval
CN116630100B (en) Travel data processing method, device, equipment and storage medium
CN103927554A (en) Image sparse representation facial expression feature extraction system and method based on topological structure
Zhang et al. Second-and high-order graph matching for correspondence problems
Li et al. Image gradient orientations embedded structural error coding for face recognition with occlusion
Wiggers et al. Document image retrieval using deep features
CN109034213A (en) Hyperspectral image classification method and system based on joint entropy principle
Sajid et al. Facial asymmetry-based feature extraction for different applications: a review complemented by new advances
CN104573726B (en) Facial image recognition method based on the quartering and each ingredient reconstructed error optimum combination
CN112966629A (en) Remote sensing image scene classification method based on image transformation and BoF model
CN105550642A (en) Gender identification method and system based on multi-scale linear difference characteristic low-rank expression
Shu et al. Local and global regularized sparse coding for data representation

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
CB03 Change of inventor or designer information

Inventor after: Yang Yubin

Inventor after: Mao Xiaojiao

Inventor after: Xu Jing

Inventor after: Qian Hongbo

Inventor before: Yang Yubin

Inventor before: Mao Xiaojiao

Inventor before: Qian Hongbo

COR Change of bibliographic data

Free format text: CORRECT: INVENTOR; FROM: YANG YUBIN MAO XIAOJIAO QIAN HONGBO TO: YANG YUBIN MAO XIAOJIAO XU JING QIAN HONGBO

C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20150513

Termination date: 20211210