CN105469018B - Method and equipment for positioning human eyes - Google Patents

Method and equipment for positioning human eyes Download PDF

Info

Publication number
CN105469018B
CN105469018B CN201410388103.XA CN201410388103A CN105469018B CN 105469018 B CN105469018 B CN 105469018B CN 201410388103 A CN201410388103 A CN 201410388103A CN 105469018 B CN105469018 B CN 105469018B
Authority
CN
China
Prior art keywords
matrix
image
eye
ith
human
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201410388103.XA
Other languages
Chinese (zh)
Other versions
CN105469018A (en
Inventor
王勃飞
邓伟洪
张殿凯
雷晨雨
瞿广财
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
ZTE Corp
Original Assignee
ZTE Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ZTE Corp filed Critical ZTE Corp
Priority to CN201410388103.XA priority Critical patent/CN105469018B/en
Priority to PCT/CN2015/072055 priority patent/WO2016019715A1/en
Publication of CN105469018A publication Critical patent/CN105469018A/en
Application granted granted Critical
Publication of CN105469018B publication Critical patent/CN105469018B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/10Image acquisition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The embodiment of the invention discloses a method and equipment for positioning human eyes, wherein the method comprises the following steps: preprocessing an original face image to obtain an image to be processed; filtering each pixel point of the image to be processed according to a pre-trained average synthesis correlation filter (ASEF) template to obtain a corresponding filtering response value of each pixel point of the image to be processed; sequentially selecting a preset number of human eye candidate points from all pixel points of the image to be processed according to the sequence of the filtering response values corresponding to each pixel point of the image to be processed from large to small; and determining human eye position points from the preset number of human eye candidate points through parameters of a pre-trained fast local linear Support Vector Machine (SVM).

Description

Method and equipment for positioning human eyes
Technical Field
The present invention relates to image processing technologies, and in particular, to a method and an apparatus for positioning human eyes.
Background
In the related art of facial image processing such as face tracking, face recognition, expression analysis, eye control, etc., eye localization is a very critical step because the eyes are the most stable and prominent features of the face and there is enough gradient information to facilitate accurate localization.
Currently, among many methods for locating human eyes, a classifier method based on a Support Vector Machine (SVM) has high precision, but the time complexity in the calculation process is o (n)2) Moreover, because the movement speed of human eyes is high, the requirement of quickly positioning the human eyes in practical application is difficult to meet with high time complexity, and the positioning efficiency is low.
Disclosure of Invention
In order to solve the above technical problems, embodiments of the present invention are expected to provide a method and an apparatus for positioning human eyes, which can reduce the time complexity of human eye positioning under a high-precision condition, so as to achieve higher positioning efficiency while maintaining high-precision human eye positioning.
The technical scheme of the invention is realized as follows:
in a first aspect, an embodiment of the present invention provides a method for positioning a human eye, where the method may include:
preprocessing an original face image to obtain an image to be processed;
filtering each pixel point of the image to be processed according to a pre-trained average synthesis correlation filter (ASEF) template to obtain a corresponding filtering response value of each pixel point of the image to be processed;
sequentially selecting a preset number of human eye candidate points from all pixel points of the image to be processed according to the sequence of the filtering response values corresponding to each pixel point of the image to be processed from large to small;
and determining human eye position points from the preset number of human eye candidate points through parameters of a pre-trained fast local linear Support Vector Machine (SVM).
Further, the method further comprises:
carrying out face detection on a preset number of sample images, normalizing the detected face images to a preset size, and carrying out Gaussian smoothing to obtain a preset number of face sample images;
and training the parameters of the ASEF template and the rapid local linear SVM through a preset number of face sample images.
Further, the training of the ASEF template by a preset number of face sample images includes:
the ith human face sample image Im _ orign (x)i,yi) After windowing, performing two-dimensional fast Fourier transform on the ith windowed human face sample image to obtain the ith human faceFrequency domain image Im _ orign _ FFT (u) of sample imagei,vi) Wherein i represents the number of the face sample image, and the value range of i is a positive integer greater than zero and less than N, where N is the preset number, (x)i,yi) The ith human face sample image representing the image domain, (u)i,vi) An ith human face sample image representing a frequency domain;
according to the ith human face sample image Im _ orign (x)i,yi) Actual position of the human eye (x)0,y0) Acquiring a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Corresponding Gaussian impulse response after passing through the filter template
Figure BDA0000550633480000021
And responding to the Gaussian pulse with a Response _ im (x)i,yi) Performing two-dimensional fast Fourier transform to obtain the Gaussian pulse Response _ im (x)i,yi) Frequency domain Response _ im _ FFT (u)i,vi) Wherein e is a natural base number, and σ is the Gaussian impulse Response _ im (x)i,yi) The variance of (a);
acquiring the ith human face sample image Im _ orign (x) according to the following formulai,yi) Corresponding Filter template Filter _ FFT (ui,vi),
Figure BDA0000550633480000022
Carrying out averaging operation on Filter templates corresponding to all face sample images according to the following formula to obtain an ASEF template Filter _ FFT (u, v),
Figure BDA0000550633480000031
where the symbol Σ is the sum operator.
Further, according to the ith human face sample image Im _ orign (x)i,yi) Middle-aged peopleActual position of eye (x)0,y0) Acquiring a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Corresponding Gaussian impulse response after passing through the filter template
Figure BDA0000550633480000032
And responding to the Gaussian pulse with a Response _ im (x)i,yi) Performing two-dimensional fast Fourier transform to obtain the Gaussian pulse Response _ im (x)i,yi) Frequency domain Response _ im _ FFT (u)i,vi) The method comprises the following steps:
according to the ith human face sample image Im _ orign (x)i,yi) Actual position of the left eye of the middle human (x)0L,y0L) Acquiring a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Corresponding left-eye Gaussian impulse response after passing through the filter template
Figure BDA0000550633480000033
And responding to the left-eye Gaussian pulse by L _ Response _ im (x)i,yi) Performing two-dimensional fast Fourier transform to obtain the left-eye Gaussian pulse Response L _ Response _ im (x)i,yi) Frequency domain Response of L _ Response _ im _ FFT (u)i,vi);
According to the ith human face sample image Im _ orign (x)i,yi) Actual position of the right eye of the middle person (x)0R,y0R) Acquiring a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Corresponding right-eye Gaussian impulse response after passing through the filter template
Figure BDA0000550633480000034
And responding to the Gaussian pulse of the right eye R _ Response _ im (x)i,yi) Performing two-dimensional fast Fourier transform to obtain the Gaussian impulse Response R _ Response _ im (x) of the right eyei,yi) Frequency domain Response of R _ Response _ im _ FFT (u)i,vi);
Accordingly, the ith human face sample image Im _ orign (x)i,yi) Corresponding Filter template Filter _ FFT (ui,vi) The method comprises the following steps:
the ith human face sample image Im _ orign (x)i,yi) Corresponding left eye Filter template L _ Filter _ FFT (u)i,vi) And the ith human face sample image Im _ orign (x)i,yi) Corresponding right eye Filter template R _ Filter _ FFT (u)i,vi) Wherein, in the step (A),
Figure BDA0000550633480000035
Figure BDA0000550633480000043
the ASEF template Filter _ FFT (u, v) comprises: left eye ASEF template L _ Filter _ FFT (u, v) and right eye ASEF template R _ Filter _ FFT (u, v), where,
Figure BDA0000550633480000041
Figure BDA0000550633480000042
further, training the parameters of the fast local linear SVM through a preset number of face sample images, including:
generating corresponding human eye sample images from a preset number of human face sample images, and forming a sample matrix X from all the human eye sample images through dimensionality reduction;
obtaining a correlation matrix K between any two human eye sample images according to the sample matrix X;
acquiring a support vector of the sample matrix X and a weight corresponding to the support vector through the correlation matrix K and an SVM optimization algorithm;
and calculating parameters of the fast local linear SVM according to the support vector of the sample matrix X and the weight corresponding to the support vector.
Further, generating corresponding human eye sample images from a preset number of human face sample images, and forming a sample matrix X from all the human eye sample images through dimensionality reduction, wherein the method comprises the following steps:
setting the ith human face sample image Im _ orign (x)i,yi) Wherein i represents the image number of the human face sample, the value range of i is a positive integer greater than zero and less than N, N is the preset number, (x)i,yi) The ith human face sample image representing the image domain;
after acquiring human eye positive samples and human eye negative samples of all human face sample images, performing mean value removing and modular normalization processing on all the human eye positive samples and the human eye negative samples to generate human eye sample images corresponding to each human face sample image;
stretching the ith human eye sample image into a column vector corresponding to the ith human eye sample image, and taking the column vector corresponding to the ith human eye sample image as the ith column vector in the sample matrix X.
Further, the obtaining a correlation matrix K between any two face sample images according to the sample matrix X includes:
performing singular value SVD decomposition on the sample matrix X to obtain a result shown as the following formula:
X=V∑UH
the first decomposition matrix V and the third decomposition matrix U are unitary matrixes, the second decomposition matrix sigma is a semi-positive determined diagonal matrix, and H represents a conjugate transpose;
a sub-matrix of a first decomposition matrix V consisting of the first t column vectors of the first decomposition matrix V
Figure BDA0000550633480000053
And a sub-matrix of a second decomposition matrix sigma consisting of the first t rows and the first t columns of said second decomposition matrix sigma
Figure BDA0000550633480000054
Obtaining code word byForming a matrix G:
Figure BDA0000550633480000051
wherein the superscript one 1 denotes the sub-matrix to the second decomposition matrix sigma
Figure BDA0000550633480000055
Carrying out inversion operation;
generating a matrix G according to the codes and a column vector x corresponding to the ith human eye sample imageiThe column vector x corresponding to the jth human eye sample imagejObtaining the element K of the ith row and the jth column of the correlation matrix K by the following formulaij
Kij=(GTxi)T(GTxj)xi Txj
Wherein, KijA column vector x representing the i-th human eye sample imageiThe column vector x corresponding to the jth human eye sample imagejThe degree of correlation between them, T, represents the transposition operation.
Further, the obtaining the parameters of the fast local linear SVM according to the support vector of the sample matrix X and the weight corresponding to the support vector includes:
obtaining an intermediate matrix A according to the support vector of the sample matrix X, the weight corresponding to the support vector and the code generation matrix G by the following formula:
Figure BDA0000550633480000052
wherein M is the number of support vectors of the sample matrix X, sjThe jth support vector, α, representing the sample matrix XjRepresents the weight corresponding to the jth support vector, yjRepresenting a training sample label corresponding to the jth support vector;
by A' ═ A + A according to the intermediate matrix AT) A symmetric matrix A' is obtained;
performing eigenvalue decomposition on the symmetric matrix A 'to obtain eigenvalues of the symmetric matrix A' and eigenvectors corresponding to the eigenvalues;
according to the sequence of the eigenvalues of the symmetric matrix A 'from large to small, P selected eigenvalues and eigenvectors corresponding to the P selected eigenvalues are sequentially selected from the eigenvalues of the symmetric matrix A', and the P selected eigenvalues and the eigenvectors corresponding to the P selected eigenvalues are used as the parameters of the fast local linear SVM.
Further, determining a human eye position point from the preset number of human eye candidate points through a pre-trained parameter of the fast local linear support vector machine SVM, comprising:
stretching pixel block with k-th personal eye candidate point as neighborhood into vector z to be processedk
According to the vector z to be processedkAnd the candidate decision value V1 of the kth personal eye candidate point is obtained through the following formula by the P selected feature values and the feature vectors corresponding to the P selected feature values:
Figure BDA0000550633480000061
wherein λ isiRepresenting the ith selected characteristic value, qiRepresenting the feature vector corresponding to the ith selected feature value;
adding the candidate decision value V1 of the kth personal eye candidate point and the filter response value corresponding to the kth personal eye candidate point to obtain a final decision value of the kth personal eye candidate point;
and selecting the human eye candidate point with the highest final judgment value from the preset number of human eye candidate points as the human eye position point.
In a second aspect, an embodiment of the present invention provides an apparatus for positioning a human eye, where the apparatus includes: a preprocessing unit, a filtering unit, a selecting unit and a determining unit, wherein,
the preprocessing unit is used for preprocessing the original face image to obtain an image to be processed;
the filtering unit is used for filtering each pixel point of the image to be processed according to a pre-trained average synthesis correlation filter (ASEF) template to obtain a filtering response value corresponding to each pixel point of the image to be processed;
the selecting unit is used for sequentially selecting a preset number of human eye candidate points from all pixel points of the image to be processed according to the sequence that the filtering response value corresponding to each pixel point of the image to be processed is from large to small;
and the determining unit is used for determining the human eye position points from the preset number of human eye candidate points through the parameters of the pre-trained fast local linear support vector machine SVM.
Further, the apparatus further comprises: a detection unit, a first training unit and a second training unit, wherein,
the detection unit is used for carrying out face detection on a preset number of sample images, normalizing the detected face images to a preset size, and then carrying out Gaussian smoothing to obtain a preset number of face sample images;
the first training unit is used for training the ASEF template through a preset number of face sample images;
the second training unit is used for training the parameters of the fast local linear SVM through a preset number of face sample images.
Further, the first training unit is configured to:
the ith human face sample image Im _ orign (x)i,yi) After windowing, performing two-dimensional fast Fourier transform on the ith windowed human face sample image to obtain a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Wherein i represents the number of the face sample image, and the value range of i is a positive integer greater than zero and less than N, where N is the preset number, (x)i,yi) The ith human face sample image representing the image domain, (u)i,vi) Ith sheet showing frequency domainA face sample image;
and according to the ith human face sample image Im _ orign (x)i,yi) Actual position of the human eye (x)0,y0) Acquiring a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Corresponding Gaussian impulse response after passing through the filter template
Figure BDA0000550633480000071
And responding to the Gaussian pulse with a Response _ im (x)i,yi) Performing two-dimensional fast Fourier transform to obtain the Gaussian pulse Response _ im (x)i,yi) Frequency domain Response _ im _ FFT (u)i,vi) Wherein e is a natural base number, and σ is the Gaussian impulse Response _ im (x)i,yi) The variance of (a);
and acquiring the ith human face sample image Im _ orign (x) according to the following formulai,yi) Corresponding Filter template Filter _ FFT (ui,vi),
Figure BDA0000550633480000081
Carrying out averaging operation on Filter templates corresponding to all face sample images according to the following formula to obtain an ASEF template Filter _ FFT (u, v),
Figure BDA0000550633480000082
where the symbol Σ is the sum operator.
Further, the first training unit is configured to train the face image according to the ith face sample image Im _ orign (x)i,yi) Actual position of the left eye of the middle human (x)0L,y0L) Acquiring a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Corresponding left-eye Gaussian impulse response after passing through the filter template
Figure BDA0000550633480000083
And responding to the left-eye Gaussian pulse by L _ Response _ im (x)i,yi) Performing two-dimensional fast Fourier transform to obtain the left-eye Gaussian pulse Response L _ Response _ im (x)i,yi) Frequency domain Response of L _ Response _ im _ FFT (u)i,vi);
And according to the ith human face sample image Im _ orign (x)i,yi) Actual position of the right eye of the middle person (x)0R,y0R) Acquiring a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Corresponding right-eye Gaussian impulse response after passing through the filter template
Figure BDA0000550633480000084
And responding to the Gaussian pulse of the right eye R _ Response _ im (x)i,yi) Performing two-dimensional fast Fourier transform to obtain the Gaussian impulse Response R _ Response _ im (x) of the right eyei,yi) Frequency domain Response of R _ Response _ im _ FFT (u)i,vi);
Accordingly, the ith human face sample image Im _ orign (x)i,yi) Corresponding Filter template Filter _ FFT (ui,vi) The method comprises the following steps:
the ith human face sample image Im _ orign (x)i,yi) Corresponding left eye Filter template L _ Filter _ FFT (u)i,vi) And the ith human face sample image Im _ orign (x)i,yi) Corresponding right eye Filter template R _ Filter _ FFT (u)i,vi) Wherein, in the step (A),
Figure BDA0000550633480000091
Figure BDA0000550633480000092
the ASEF template Filter _ FFT (u, v) comprises: left eye ASEF template L _ Filter _ FFT (u, v) and right eye ASEF template R _ Filter _ FFT (u, v), where,
Figure BDA0000550633480000093
Figure BDA0000550633480000094
further, the second training unit includes: a dimension reduction subunit, an acquisition subunit, and a computation subunit, wherein,
the dimension reduction subunit is used for generating corresponding human eye sample images from a preset number of human face sample images and forming a sample matrix X from all the human eye sample images through dimension reduction;
the obtaining subunit is configured to obtain a correlation matrix K between any two human eye sample images according to the sample matrix X; and the number of the first and second groups,
acquiring a support vector of the sample matrix X and a weight corresponding to the support vector through the correlation matrix K and an SVM optimization algorithm;
and the calculating subunit is configured to calculate a parameter of the fast local linear SVM according to the support vector of the sample matrix X and the weight corresponding to the support vector.
Further, the dimension reduction subunit is configured to:
setting the ith human face sample image Im _ orign (x)i,yi) Wherein i represents the image number of the human face sample, the value range of i is a positive integer greater than zero and less than N, N is the preset number, (x)i,yi) The ith human face sample image representing the image domain;
after acquiring human eye positive samples and human eye negative samples of all human face sample images, performing mean value removing and modulo two normalization processing on all human eye positive samples and human eye negative samples to generate human eye sample images corresponding to each human face sample image;
and stretching the ith human eye sample image into a column vector corresponding to the ith human eye sample image, and taking the column vector corresponding to the ith human eye sample image as the ith column vector in the sample matrix X.
Further, the obtaining subunit is configured to:
performing singular value SVD decomposition on the sample matrix X to obtain a result shown as the following formula:
X=V∑UH
the first decomposition matrix V and the third decomposition matrix U are unitary matrixes, the second decomposition matrix sigma is a semi-positive determined diagonal matrix, and H represents a conjugate transpose;
and a sub-matrix of the first decomposition matrix V composed of the first t column vectors of the first decomposition matrix V
Figure BDA0000550633480000101
And a sub-matrix of a second decomposition matrix sigma consisting of the first t rows and the first t columns of said second decomposition matrix sigma
Figure BDA0000550633480000102
The code generation matrix G is obtained by:
Figure BDA0000550633480000103
where the superscript-1 denotes the sub-matrix to the second decomposition matrix sigma
Figure BDA0000550633480000104
Carrying out inversion operation;
and generating a matrix G according to the codes and a column vector x corresponding to the ith human eye sample imageiThe column vector x corresponding to the jth human eye sample imagejObtaining the element K of the ith row and the jth column of the correlation matrix K by the following formulaij
Kij=(GTxi)T(GTxj)xi Txj
Wherein, KijA column vector x representing the i-th human eye sample imageiAnd the jth human eye sample imageLike corresponding column vector xjThe degree of correlation between them, T, represents the transposition operation.
Further, the calculating subunit is configured to:
obtaining an intermediate matrix A according to the support vector of the sample matrix X, the weight corresponding to the support vector and the code generation matrix G by the following formula:
Figure BDA0000550633480000111
wherein M is the number of support vectors of the sample matrix X, sjThe jth support vector, α, representing the sample matrix XjRepresents the weight corresponding to the jth support vector, yjRepresenting a training sample label corresponding to the jth support vector;
and, according to said intermediate matrix a, by a' ═ a + aT) A symmetric matrix A' is obtained;
performing eigenvalue decomposition on the symmetric matrix A 'to obtain eigenvalues of the symmetric matrix A' and eigenvectors corresponding to the eigenvalues;
and sequentially selecting P selected eigenvalues and eigenvectors corresponding to the P selected eigenvalues from the eigenvalues of the symmetric matrix A 'according to the descending order of the eigenvalues of the symmetric matrix A', and taking the P selected eigenvalues and the eigenvectors corresponding to the P selected eigenvalues as the parameters of the fast local linear SVM.
Further, the determining unit is configured to:
stretching pixel block with k-th personal eye candidate point as neighborhood into vector z to be processedk
And according to the vector z to be processedkAnd the candidate decision value V1 of the kth personal eye candidate point is obtained through the following formula by the P selected feature values and the feature vectors corresponding to the P selected feature values:
Figure BDA0000550633480000112
wherein λ isiRepresenting the ith selected characteristic value, qiRepresenting the feature vector corresponding to the ith selected feature value;
adding the candidate judgment value V1 of the kth personal eye candidate point and the filter response value corresponding to the kth personal eye candidate point to obtain a final judgment value of the kth personal eye candidate point;
and selecting the human eye candidate point with the highest final judgment value from the preset number of human eye candidate points as the human eye position point.
The embodiment of the invention provides a method and equipment for positioning human eyes, wherein the position of the human eyes is determined by an ASEF template obtained by pre-training an image to be positioned and parameters of a fast local linear SVM; the time complexity of human eye positioning can be reduced under the high-precision condition, so that high positioning efficiency is realized while high-precision human eye positioning is kept.
Drawings
Fig. 1 is a schematic flow chart of a method for positioning human eyes according to an embodiment of the present invention;
fig. 2 is a schematic flow chart illustrating filtering performed on each pixel point of an image to be processed according to a pre-trained ASEF template according to an embodiment of the present invention;
fig. 3 is a schematic flowchart of determining a position point of a human eye according to an embodiment of the present invention;
fig. 4 is a schematic flow chart of a method for training parameters of an ASEF template and a fast local linear SVM according to an embodiment of the present invention;
FIG. 5 is a schematic diagram of a Gaussian impulse response provided by an embodiment of the present invention;
fig. 6 is a schematic structural diagram of an apparatus for positioning human eyes according to an embodiment of the present invention;
fig. 7 is a schematic structural diagram of another human eye positioning device according to an embodiment of the present invention.
Detailed Description
The technical solution in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention.
Referring to fig. 1, a flow chart of a method for positioning a human eye according to an embodiment of the present invention is shown, where the method may include:
s101: preprocessing an original face image to obtain an image to be processed;
exemplarily, since the face image is obtained under natural conditions, the original face image needs to be preprocessed before being processed, so as to remove interference caused by uneven illumination, noise and other factors in the original face image;
in this embodiment, the process of preprocessing the original face image may include:
firstly, detecting an original face image from an original obtained image by a common face detector;
then, normalizing the original face image into an image with a preset size, wherein the preset size is selected to be 100 × 100 in the embodiment, and the unit is a pixel;
and then, performing Gaussian smoothing on the normalized image, and dividing the pixel value of the position corresponding to the normalized image by the pixel value of the position corresponding to the image after the Gaussian smoothing, thereby completing the preprocessing of the original face image and obtaining the image to be processed without the influences of uneven illumination, noise and the like.
It should be noted that, because the preprocessing is used to remove the influence of uneven illumination, noise, etc. in the original image, the embodiment of the present invention is not limited to use only one preprocessing method, and other preprocessing methods that can achieve corresponding effects may also be applied to the embodiment of the present invention.
S102: filtering each pixel point of the image to be processed according to a pre-trained ASEF template to obtain a corresponding filtering response value of each pixel point of the image to be processed;
it should be noted that the pre-trained ASEF template essentially belongs to a filter, and therefore, a specific process of filtering each pixel point of the image to be processed according to the pre-trained ASEF template may be as shown in fig. 2: taking each pixel point of the image to be processed as an input value IN of the pre-trained ASEF template, and obtaining an output value OUT which is a filter response value corresponding to each pixel point of the image to be processed after the filtering processing of the pre-trained ASEF template;
as for the filtering process shown in fig. 2, it can be understood that, in the image domain, the convolution operation is performed on the image to be processed and the pre-trained ASEF template; in the frequency domain, the frequency spectrum of the image to be processed is multiplied by the frequency spectrum of the ASEF template which is trained in advance.
S103: sequentially selecting a preset number of human eye candidate points from all pixel points of the image to be processed according to the sequence of the filtering response values corresponding to each pixel point of the image to be processed from large to small;
s104: determining human eye position points from the preset number of human eye candidate points through parameters of a pre-trained fast local linear SVM;
for example, in step S104, a corresponding determination value may be determined for each eye candidate point through a pre-trained parameter of the fast local linear SVM, so as to determine a human eye position point in the human eye candidate points; however, in order to improve the accuracy of determining the eye position points, the determination value corresponding to each eye candidate point may be considered in combination with the filter response value corresponding to the eye candidate point obtained in step S102 to obtain a final determination value corresponding to each eye candidate point, and the eye candidate point with the highest final determination value is used as the eye position point;
specifically, referring to fig. 3, determining the human eye position point from the preset number of human eye candidate points through the parameters of the pre-trained fast local linear support vector machine SVM may include S1041 to S1044:
s1041: stretching pixel block with k-th personal eye candidate point as neighborhood into vector z to be processedk
S1042: according to the vector z to be processedkP selected eigenvalues and eigenvectors corresponding to the P selected eigenvalues, and obtaining the kth personal eye candidate by the following formulaCandidate determination value of point V1:
Figure BDA0000550633480000141
further, the P selected feature values and the feature vectors corresponding to the P selected feature values are parameters of the pre-trained fast local linear support vector machine SVM, λiRepresenting the ith selected characteristic value, qiRepresenting the eigenvector corresponding to the ith selected eigenvalue, wherein T is transposition operation;
s1043: adding the candidate decision value V1 of the kth personal eye candidate point and the filter response value corresponding to the kth personal eye candidate point to obtain a final decision value of the kth personal eye candidate point;
s1044: and selecting the human eye candidate point with the highest final judgment value from the preset number of human eye candidate points as the human eye position point.
It can be understood that, in the prior art, usually, the method for locating the human eye directly by the SVM needs to calculate each pixel point of the face image, and therefore, the time complexity of the prior art for the method for locating the human eye is o (n)2) (ii) a In the human eye positioning method provided in this embodiment, because the positioning is performed only in the preset number of human eye candidate points through the fast local linear SVM, the time complexity of the method provided in this embodiment is o (n), so that the time complexity in the human eye positioning process is reduced, and the efficiency is improved.
It should be noted that, since the pre-trained ASEF template and the pre-trained fast local linear SVM parameters are used in step S102 and step S104, respectively, the parameters of the ASEF template and the fast local linear SVM need to be trained before the method shown in fig. 1 locates the human eye. Therefore, on the basis of the embodiment shown in fig. 1, referring to fig. 4, it shows a flow of a training method for parameters of an ASEF template and a fast local linear SVM provided in an embodiment of the present invention, where the training method may be obtained by training a preset number of existing sample images, and the specific process may include:
s401: carrying out face detection on a preset number of sample images, normalizing the detected face images to a preset size, and carrying out Gaussian smoothing to obtain a preset number of face sample images;
optionally, the specific process of S401 may be the same as the specific process of S101, and is not described herein again.
S402: training parameters of the ASEF template and the fast local linear SVM through a preset number of face sample images;
exemplarily, the parameters of the ASEF template and the fast local linear SVM may be trained through a preset number of face sample images, respectively; therefore, the following describes in detail the training process of the parameters of the ASEF template and the fast local linear SVM respectively through a preset number of face sample images according to the embodiment of the present invention.
Preferably, the training of the ASEF template by a preset number of face sample images may include:
first, the ith human face sample image Im _ orign (x)i,yi) After windowing, performing two-dimensional fast Fourier transform on the ith windowed human face sample image to obtain a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi);
Wherein i represents the number of the face sample image, and the value range of i is a positive integer greater than zero and less than N, where N is the preset number, (x)i,yi) The ith human face sample image representing the image domain, (u)i,vi) An ith human face sample image representing a frequency domain; in this embodiment, the ith human face sample image Im _ orign (x) may be obtained firsti,yi) And after windowing is carried out through the cosine window, two-dimensional fast Fourier transform is carried out.
Secondly, according to the ith human face sample image Im _ orign (x)i,yi) Actual position of the human eye (x)0,y0) Acquiring a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Warp beamCorresponding Gaussian impulse response after passing through the filter template
Figure BDA0000550633480000151
And responding to the Gaussian pulse with a Response _ im (x)i,yi) Performing two-dimensional fast Fourier transform to obtain the Gaussian pulse Response _ im (x)i,yi) Frequency domain Response _ im _ FFT (u)i,vi) Wherein e is a natural base number, and σ is the Gaussian impulse Response _ im (x)i,yi) The variance of (a); and the gaussian impulse response may be as shown in fig. 5;
as will be appreciated, since the human eye can be divided into a left eye and a right eye, this step may specifically comprise:
according to the ith human face sample image Im _ orign (x)i,yi) Actual position of the left eye of the middle human (x)0L,y0L) Acquiring a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Corresponding left-eye Gaussian impulse response after passing through the filter template
Figure BDA0000550633480000152
And responding to the left-eye Gaussian pulse by L _ Response _ im (x)i,yi) Performing two-dimensional fast Fourier transform to obtain the left-eye Gaussian pulse Response L _ Response _ im (x)i,yi) Frequency domain Response of L _ Response _ im _ FFT (u)i,vi) (ii) a And the number of the first and second groups,
according to the ith human face sample image Im _ orign (x)i,yi) Actual position of the right eye of the middle person (x)0R,y0R) Acquiring a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Corresponding right-eye Gaussian impulse response after passing through the filter template
Figure BDA0000550633480000161
And responding to the Gaussian pulse of the right eye R _ Response _ im (x)i,yi) A two-dimensional fast fourier transform is performed,obtaining the Gaussian impulse Response R _ Response _ im (x) of the right eyei,yi) Frequency domain Response of R _ Response _ im _ FFT (u)i,vi);
Then, the ith human face sample image Im _ orign (x) is obtained according to the following formulai,yi) Corresponding Filter template Filter _ FFT (ui,vi),
Figure BDA0000550633480000162
Corresponding to the previous step, the Gaussian impulse Response L _ Response _ im _ FFT (u) of the left eyei,vi) And a Gaussian impulse Response R _ Response _ im _ FFT (u) for the right eyei,vi) The ith human face sample image Im _ orign (x) obtained in the stepi,yi) Corresponding Filter template Filter _ FFT (ui,vi) The method comprises the following steps: the ith human face sample image Im _ orign (x)i,yi) Corresponding left eye Filter template L _ Filter _ FFT (u)i,vi) And the ith human face sample image Im _ orign (x)i,yi) Corresponding right eye Filter template R _ Filter _ FFT (u)i,vi) Wherein, in the step (A),
Figure BDA0000550633480000163
Figure BDA0000550633480000164
finally, according to the following formula, averaging operation is carried out on Filter templates corresponding to all face sample images to obtain the ASEF template Filter _ FFT (u, v),
Figure BDA0000550633480000165
where the symbol Σ is the sum operator.
Corresponding to the ith human face sample image Im \uobtained in the last steporign(xi,yi) Left eye Filter template L _ Filter _ FFT (u)i,vi) And the ith face sample image Im _ orign (x)i,yi) Corresponding right eye Filter template R _ Filter _ FFT (u)i,vi) The Filter _ FFT (u, v) of the ASEF template obtained in this step may include: left eye ASEF template L _ Filter _ FFT (u, v) and right eye ASEF template R _ Filter _ FFT (u, v), where,
Figure BDA0000550633480000171
Figure BDA0000550633480000172
preferably, the training of the parameters of the fast local linear SVM through a preset number of face sample images may include:
firstly, generating corresponding human eye sample images from a preset number of human face sample images, and forming a sample matrix X from all the human eye sample images through dimensionality reduction;
further, the step may include:
setting the ith human face sample image Im _ orign (x)i,yi) Wherein i represents a face sample image number, a value range of i is a positive integer greater than zero and smaller than N, N is the preset number, in this embodiment, N is 10000, (x) isi,yi) The ith human face sample image representing the image domain; it should be noted that, in this embodiment, a rectangular pixel block of a fixed size centered on the human eye may be used as a positive sample of the human eye; taking a rectangular pixel block of a human eye region except for a fixed pixel from the center of the human eye as a human eye negative sample, it can be understood that a common region exists between the human eye positive sample and the human eye negative sample.
After acquiring human eye positive samples and human eye negative samples of all human face sample images, performing mean value removing and modulo two normalization processing on all human eye positive samples and human eye negative samples to generate human eye sample images corresponding to each human face sample image; in this embodiment, the size of the human eye sample image corresponding to each human face sample image may be 31 pixels × 31 pixels;
stretching the ith human eye sample image into a column vector corresponding to the ith human eye sample image, and taking the column vector corresponding to the ith human eye sample image as the ith column vector in the sample matrix X; in this embodiment, the dimension of the column vector corresponding to the i-th human eye sample image is 31 × 31 ═ 961;
and, the above-mentioned processing is performed on each human eye sample image, and finally the sample matrix X is obtained, so that it can be known that, in the present embodiment, the sample matrix is a 961 × 10000 matrix.
Secondly, obtaining a correlation matrix K between any two human eye sample images according to the sample matrix X;
further, the step may include:
performing singular value SVD decomposition on the sample matrix X to obtain a result shown as the following formula:
X=V∑UH
the first decomposition matrix V and the third decomposition matrix U are unitary matrixes, the second decomposition matrix sigma is a semi-positive determined diagonal matrix, and H represents a conjugate transpose;
in the present embodiment, the first decomposition matrix V is a 961 × 961 matrix, the second decomposition matrix Σ is a 961 × 10000 matrix, and the third decomposition matrix U is 10000 × 10000 matrix.
And a sub-matrix of the first decomposition matrix V composed of the first t column vectors of the first decomposition matrix V
Figure BDA0000550633480000183
And a sub-matrix of a second decomposition matrix sigma consisting of the first t rows and the first t columns of said second decomposition matrix sigma
Figure BDA0000550633480000184
The code generation matrix G is obtained by the following equation:
Figure BDA0000550633480000181
where the superscript-1 denotes the sub-matrix to the second decomposition matrix sigma
Figure BDA0000550633480000182
Carrying out inversion operation;
in this embodiment, the value of t may be set to 150, and thus, the sub-matrix of the first decomposition matrix V
Figure BDA0000550633480000185
A sub-matrix of 961 × 150 matrix, a second decomposition matrix ∑
Figure BDA0000550633480000186
The code generation matrix G is a 961 × 150 matrix.
Generating a matrix G according to the codes and a column vector x corresponding to the ith human eye sample imageiThe column vector x corresponding to the jth human eye sample imagejObtaining the element K of the ith row and the jth column of the correlation matrix K by the following formulaij
Kij=(GTxi)T(GTxj)xi Txj
Wherein, KijA column vector x representing the i-th human eye sample imageiThe column vector x corresponding to the jth human eye sample imagejThe correlation degree between the two elements, T, represents a transposition operation, and it can be understood that all the elements of the correlation matrix K can be obtained through this step, and the correlation matrix K is a symmetric matrix.
Then, acquiring a support vector of the sample matrix X and a weight corresponding to the support vector through the correlation matrix K and an SVM optimization algorithm; in this embodiment, the support vector of the sample matrix X may be represented by a labeled column vector, where an element in the labeled column vector is a column label of the support vector of the sample matrix X in the sample matrix X; the weight corresponding to the support vector can also be represented by a weight column vector, the elements in the weight column vector are the weights corresponding to the support vectors of the sample matrix X represented by the corresponding positions in the label column vector,
finally, calculating parameters of the fast local linear SVM according to the support vector of the sample matrix X and the weight corresponding to the support vector;
further, the step may include:
obtaining an intermediate matrix A according to the support vector of the sample matrix X, the weight corresponding to the support vector and the code generation matrix G by the following formula:
Figure BDA0000550633480000191
wherein M is the number of support vectors of the sample matrix X, sjThe jth support vector, α, representing the sample matrix XjRepresents the weight corresponding to the jth support vector, yjRepresenting a training sample label corresponding to the jth support vector;
and, according to said intermediate matrix a, by a' ═ a + aT) A symmetric matrix A' is obtained;
performing eigenvalue decomposition on the symmetric matrix A 'to obtain eigenvalues of the symmetric matrix A' and eigenvectors corresponding to the eigenvalues;
and sequentially selecting P selected eigenvalues and eigenvectors corresponding to the P selected eigenvalues from the eigenvalues of the symmetric matrix A 'according to the descending order of the eigenvalues of the symmetric matrix A', and taking the P selected eigenvalues and the eigenvectors corresponding to the P selected eigenvalues as the parameters of the fast local linear SVM.
Through the two preferred training methods, the parameters of the ASEF template and the fast local linear SVM can be obtained, and then the human eyes in the image can be located through the embodiment shown in fig. 1 according to the parameters of the ASEF template and the fast local linear SVM obtained through training.
The embodiment of the invention provides a method for positioning human eyes, which is characterized in that the positions of the human eyes are determined by an ASEF template obtained by pre-training a face image and parameters of a fast local linear SVM; the time complexity of human eye positioning can be reduced under the high-precision condition, so that high positioning efficiency is realized while high-precision human eye positioning is kept.
Based on the same technical concept of the foregoing embodiment, referring to fig. 6, which illustrates an apparatus 60 for positioning human eyes provided by an embodiment of the present invention, the apparatus may include: a preprocessing unit 601, a filtering unit 602, a selecting unit 603, and a determining unit 604, wherein,
the preprocessing unit 601 is configured to preprocess an original face image to obtain an image to be processed;
the filtering unit 602 is configured to filter each pixel point of the to-be-processed image obtained by the preprocessing unit 601 according to a pre-trained average synthesis correlation filter ASEF template, and obtain a filtering response value corresponding to each pixel point of the to-be-processed image;
the selecting unit 603 is configured to sequentially select a preset number of human eye candidate points from all pixel points of the image to be processed according to a descending order of the filtering response values corresponding to each pixel point of the image to be processed, which are obtained by the filtering unit 602;
the determining unit 604 is configured to determine a human eye position point from a preset number of human eye candidate points selected by the selecting unit 603 according to a parameter of a pre-trained fast local linear support vector machine SVM.
Exemplarily, referring to fig. 7, the apparatus further comprises: a detection unit 605, a first training unit 606 and a second training unit 607, wherein,
the detection unit 605 is configured to perform face detection on a preset number of sample images, normalize the detected face images to a preset size, and perform gaussian smoothing to obtain a preset number of face sample images;
the first training unit 606 is configured to train the ASEF template through a preset number of face sample images obtained by the detection unit 605;
the second training unit 607 is configured to train the parameters of the fast local linear SVM through the preset number of face sample images obtained by the detection unit 605.
Preferably, the first training unit 606 may be configured to:
the ith human face sample image Im _ orign (x)i,yi) After windowing, performing two-dimensional fast Fourier transform on the ith windowed human face sample image to obtain a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Wherein i represents the number of the face sample image, and the value range of i is a positive integer greater than zero and less than N, where N is the preset number, (x)i,yi) The ith human face sample image representing the image domain, (u)i,vi) An ith human face sample image representing a frequency domain;
and according to the ith human face sample image Im _ orign (x)i,yi) Actual position of the human eye (x)0,y0) Acquiring a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Corresponding Gaussian impulse response after passing through the filter template
Figure BDA0000550633480000211
And responding to the Gaussian pulse with a Response _ im (x)i,yi) Performing two-dimensional fast Fourier transform to obtain the Gaussian pulse Response _ im (x)i,yi) Frequency domain Response _ im _ FFT (u)i,vi) Wherein e is a natural base number, and σ is the Gaussian impulse Response _ im (x)i,yi) The variance of (a);
and acquiring the ith human face sample image Im _ orign (x) according to the following formulai,yi) Corresponding Filter template Filter _ FFT (ui,vi),
Figure BDA0000550633480000212
Carrying out averaging operation on Filter templates corresponding to all face sample images according to the following formula to obtain an ASEF template Filter _ FFT (u, v),
Figure BDA0000550633480000213
where the symbol Σ is the sum operator.
Specifically, the first training unit 606 is configured to:
according to the ith human face sample image Im _ orign (x)i,yi) Actual position of the left eye of the middle human (x)0L,y0L) Acquiring a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Corresponding left-eye Gaussian impulse response after passing through the filter template
Figure BDA0000550633480000221
And responding to the left-eye Gaussian pulse by L _ Response _ im (x)i,yi) Performing two-dimensional fast Fourier transform to obtain the left-eye Gaussian pulse Response L _ Response _ im (x)i,yi) Frequency domain Response of L _ Response _ im _ FFT (u)i,vi);
And according to the ith human face sample image Im _ orign (x)i,yi) Actual position of the right eye of the middle person (x)0R,y0R) Acquiring a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Corresponding right-eye Gaussian impulse response after passing through the filter template
Figure BDA0000550633480000222
And responding to the Gaussian pulse of the right eye R _ Response _ im (x)i,yi) Performing two-dimensional fast Fourier transform to obtain the Gaussian impulse Response R _ Response _ im (x) of the right eyei,yi) Frequency domain Response of R _ Response _ im _ FFT (u)i,vi);
Accordingly, the number of the first and second electrodes,the ith human face sample image Im _ orign (x)i,yi) Corresponding Filter template Filter _ FFT (ui,vi) The method comprises the following steps:
the ith human face sample image Im _ orign (x)i,yi) Corresponding left eye Filter template L _ Filter _ FFT (u)i,vi) And the ith human face sample image Im _ orign (x)i,yi) Corresponding right eye Filter template R _ Filter _ FFT (u)i,vi) Wherein, in the step (A),
Figure BDA0000550633480000223
Figure BDA0000550633480000224
the ASEF template Filter _ FFT (u, v) comprises: left eye ASEF template L _ Filter _ FFT (u, v) and right eye ASEF template R _ Filter _ FFT (u, v), where,
Figure BDA0000550633480000225
Figure BDA0000550633480000226
preferably, referring to fig. 7, the second training unit 607 includes: a dimension reduction sub-unit 6071, an acquisition sub-unit 6072 and a calculation sub-unit 6073, wherein,
the dimension reduction subunit 6071 is configured to generate corresponding human eye sample images from a preset number of human face sample images, and form a sample matrix X from all the human eye sample images through dimension reduction;
the obtaining subunit 6072 is configured to obtain a correlation matrix K between any two human eye sample images according to a sample matrix X formed by the dimension reduction subunit 6071; and the number of the first and second groups,
acquiring a support vector of the sample matrix X and a weight corresponding to the support vector through the correlation matrix K and an SVM optimization algorithm;
the calculating subunit 6073 is configured to calculate a parameter of the fast local linear SVM according to the support vector of the sample matrix X obtained by the obtaining subunit 6072 and a weight corresponding to the support vector.
Specifically, the dimension reduction subunit 6071 is configured to:
setting the ith human face sample image Im _ orign (x)i,yi) Wherein i represents the image number of the human face sample, the value range of i is a positive integer greater than zero and less than N, N is the preset number, (x)i,yi) The ith human face sample image representing the image domain;
after acquiring human eye positive samples and human eye negative samples of all human face sample images, performing mean value removing and modulo two normalization processing on all human eye positive samples and human eye negative samples to generate human eye sample images corresponding to each human face sample image;
and stretching the ith human eye sample image into a column vector corresponding to the ith human eye sample image, and taking the column vector corresponding to the ith human eye sample image as the ith column vector in the sample matrix X.
Specifically, the acquiring subunit 6072 is configured to:
performing singular value SVD decomposition on the sample matrix X to obtain a result shown as the following formula:
X=V∑UH
the first decomposition matrix V and the third decomposition matrix U are unitary matrixes, the second decomposition matrix sigma is a semi-positive determined diagonal matrix, and H represents a conjugate transpose;
and a sub-matrix of the first decomposition matrix V composed of the first t column vectors of the first decomposition matrix V
Figure BDA0000550633480000231
And a sub-matrix of a second decomposition matrix sigma consisting of the first t rows and the first t columns of said second decomposition matrix sigma
Figure BDA0000550633480000241
Is obtained byCode generation matrix G:
Figure BDA0000550633480000242
where the superscript-1 denotes the sub-matrix to the second decomposition matrix sigma
Figure BDA0000550633480000244
Carrying out inversion operation;
and generating a matrix G according to the codes and a column vector x corresponding to the ith human eye sample imageiThe column vector x corresponding to the jth human eye sample imagejObtaining the element K of the ith row and the jth column of the correlation matrix K by the following formulaij
Kij=(GTxi)T(GTxj)xi Txj
Wherein, KijA column vector x representing the i-th human eye sample imageiThe column vector x corresponding to the jth human eye sample imagejThe degree of correlation between them, T, represents the transposition operation.
In particular, the computing subunit 6073 is configured to:
obtaining an intermediate matrix A according to the support vector of the sample matrix X, the weight corresponding to the support vector and the code generation matrix G by the following formula:
Figure BDA0000550633480000243
wherein M is the number of support vectors of the sample matrix X, sjThe jth support vector, α, representing the sample matrix XjRepresents the weight corresponding to the jth support vector, yjRepresenting a training sample label corresponding to the jth support vector;
and, according to said intermediate matrix a, by a' ═ a + aT) A symmetric matrix A' is obtained;
performing eigenvalue decomposition on the symmetric matrix A 'to obtain eigenvalues of the symmetric matrix A' and eigenvectors corresponding to the eigenvalues;
and sequentially selecting P selected eigenvalues and eigenvectors corresponding to the P selected eigenvalues from the eigenvalues of the symmetric matrix A 'according to the descending order of the eigenvalues of the symmetric matrix A', and taking the P selected eigenvalues and the eigenvectors corresponding to the P selected eigenvalues as the parameters of the fast local linear SVM.
Further, the determining unit 604 is configured to:
stretching pixel block with k-th personal eye candidate point as neighborhood into vector z to be processedk
And according to the vector z to be processedkAnd the candidate decision value V1 of the kth personal eye candidate point is obtained through the following formula by the P selected feature values and the feature vectors corresponding to the P selected feature values:
Figure BDA0000550633480000251
wherein λ isiRepresenting the ith selected characteristic value, qiRepresenting the feature vector corresponding to the ith selected feature value;
adding the candidate judgment value V1 of the kth personal eye candidate point and the filter response value corresponding to the kth personal eye candidate point to obtain a final judgment value of the kth personal eye candidate point;
and selecting the human eye candidate point with the highest final judgment value from the preset number of human eye candidate points as the human eye position point.
The embodiment provides a device 60 for positioning human eyes, which determines the positions of human eyes by an ASEF template obtained by pre-training a human face image and parameters of a fast local linear SVM; the time complexity of human eye positioning can be reduced under the high-precision condition, so that high positioning efficiency is realized while high-precision human eye positioning is kept.
As will be appreciated by one skilled in the art, embodiments of the present invention may be provided as a method, system, or computer program product. Accordingly, the present invention may take the form of a hardware embodiment, a software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present invention may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, optical storage, and the like) having computer-usable program code embodied therein.
The present invention is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the invention. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
The above description is only a preferred embodiment of the present invention, and is not intended to limit the scope of the present invention.

Claims (18)

1. A method of eye localization, the method comprising:
preprocessing an original face image to obtain an image to be processed;
filtering each pixel point of the image to be processed according to a pre-trained average synthesis correlation filter (ASEF) template to obtain a corresponding filtering response value of each pixel point of the image to be processed;
sequentially selecting a preset number of human eye candidate points from all pixel points of the image to be processed according to the sequence of the filtering response values corresponding to each pixel point of the image to be processed from large to small;
and determining human eye position points from the preset number of human eye candidate points through parameters of a pre-trained fast local linear Support Vector Machine (SVM).
2. The method of claim 1, further comprising:
carrying out face detection on a preset number of sample images, normalizing the detected face images to a preset size, and carrying out Gaussian smoothing to obtain a preset number of face sample images;
and training the parameters of the ASEF template and the rapid local linear SVM through a preset number of face sample images.
3. The method of claim 2, wherein the training of the ASEF template by a preset number of face sample images comprises:
the ith human face sample image Im _ orign (x)i,yi) After windowing, performing two-dimensional fast Fourier transform on the ith windowed human face sample image to obtain a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Wherein i represents the number of the face sample image, the value range of i is a positive integer which is larger than zero and smaller than N, wherein,n is the predetermined number, (x)i,yi) The ith human face sample image representing the image domain, (u)i,vi) An ith human face sample image representing a frequency domain;
according to the ith human face sample image Im _ orign (x)i,yi) Actual position of the human eye (x)0,y0) Acquiring a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Corresponding Gaussian impulse response after passing through the filter template
Figure FDA0002304390800000011
And responding to the Gaussian pulse with a Response _ im (x)i,yi) Performing two-dimensional fast Fourier transform to obtain the Gaussian pulse Response _ im (x)i,yi) Frequency domain Response _ im _ FFT (u)i,vi) Wherein e is a natural base number, and σ is the Gaussian impulse Response _ im (x)i,yi) The variance of (a);
acquiring the ith human face sample image Im _ orign (x) according to the following formulai,yi) Corresponding Filter template Filter _ FFT (ui,vi),
Figure FDA0002304390800000021
Carrying out averaging operation on Filter templates corresponding to all face sample images according to the following formula to obtain an ASEF template Filter _ FFT (u, v),
Figure FDA0002304390800000022
where the symbol Σ is the sum operator.
4. The method of claim 3, wherein the image is based on an i-th face sample image Im _ orign (x)i,yi) Actual position of the human eye (x)0,y0) Obtaining the ith human face sample imageFrequency domain image Im _ orign _ FFT (u)i,vi) Corresponding Gaussian impulse response after passing through the filter template
Figure FDA0002304390800000023
And responding to the Gaussian pulse with a Response _ im (x)i,yi) Performing two-dimensional fast Fourier transform to obtain the Gaussian pulse Response _ im (x)i,yi) Frequency domain Response _ im _ FFT (u)i,vi) The method comprises the following steps:
according to the ith human face sample image Im _ orign (x)i,yi) Actual position of the left eye of the middle human (x)0L,y0L) Acquiring a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Corresponding left-eye Gaussian impulse response after passing through the filter template
Figure FDA0002304390800000024
And responding to the left-eye Gaussian pulse by L _ Response _ im (x)i,yi) Performing two-dimensional fast Fourier transform to obtain the left-eye Gaussian pulse Response L _ Response _ im (x)i,yi) Frequency domain Response of L _ Response _ im _ FFT (u)i,vi);
According to the ith human face sample image Im _ orign (x)i,yi) Actual position of the right eye of the middle person (x)0R,y0R) Acquiring a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Corresponding right-eye Gaussian impulse response after passing through the filter template
Figure FDA0002304390800000031
And responding to the Gaussian pulse of the right eye R _ Response _ im (x)i,yi) Performing two-dimensional fast Fourier transform to obtain the Gaussian impulse Response R _ Response _ im (x) of the right eyei,yi) Frequency domain Response of R _ Response _ im _ FFT (u)i,vi);
Accordingly, the ith human face sample image Im _ orign (x)i,yi) Corresponding Filter template Filter _ FFT (ui,vi) The method comprises the following steps:
the ith human face sample image Im _ orign (x)i,yi) Corresponding left eye Filter template L _ Filter _ FFT (u)iv,i) And the ith human face sample image Im _ orign (x)i,yi) Corresponding right eye Filter template R _ Filter _ FFT (u)i,vi) Wherein, in the step (A),
Figure FDA0002304390800000032
Figure FDA0002304390800000033
the ASEF template Filter _ FFT (u, v) comprises: left eye ASEF template L _ Filter _ FFT (u, v) and right eye ASEF template R _ Filter _ FFT (u, v), where,
Figure FDA0002304390800000034
Figure FDA0002304390800000035
5. the method of claim 2, wherein the training of the parameters of the fast local linear SVM is performed by using a preset number of face sample images, and comprises:
generating corresponding human eye sample images from a preset number of human face sample images, and forming a sample matrix X from all the human eye sample images through dimensionality reduction;
obtaining a correlation matrix K between any two human eye sample images according to the sample matrix X;
acquiring a support vector of the sample matrix X and a weight corresponding to the support vector through the correlation matrix K and an SVM optimization algorithm;
and calculating parameters of the fast local linear SVM according to the support vector of the sample matrix X and the weight corresponding to the support vector.
6. The method of claim 5, wherein generating corresponding human eye sample images from a preset number of human face sample images, and forming a sample matrix X from all human eye sample images by dimension reduction comprises:
setting the ith human face sample image Im _ orign (x)i,yi) Wherein i represents the image number of the human face sample, the value range of i is a positive integer greater than zero and less than N, N is the preset number, (x)i,yi) The ith human face sample image representing the image domain;
after acquiring human eye positive samples and human eye negative samples of all human face sample images, performing mean value removing and modular normalization processing on all the human eye positive samples and the human eye negative samples to generate human eye sample images corresponding to each human face sample image;
stretching the ith human eye sample image into a column vector corresponding to the ith human eye sample image, and taking the column vector corresponding to the ith human eye sample image as the ith column vector in the sample matrix X.
7. The method according to claim 6, wherein the obtaining a correlation matrix K between any two face sample images according to the sample matrix X comprises:
performing singular value SVD decomposition on the sample matrix X to obtain a result shown as the following formula:
X=VΣUH
the first decomposition matrix V and the third decomposition matrix U are unitary matrixes, the second decomposition matrix sigma is a semi-positive diagonal matrix, and H represents a conjugate transpose;
a sub-matrix of a first decomposition matrix V consisting of the first t column vectors of the first decomposition matrix V
Figure FDA0002304390800000041
And t rows before the second decomposition matrix ΣSubmatrix of a second decomposition matrix sigma composed of t columns
Figure FDA0002304390800000042
The code generation matrix G is obtained by:
Figure FDA0002304390800000043
wherein the superscript-1 denotes the sub-matrix to the second decomposition matrix Σ
Figure FDA0002304390800000044
Carrying out inversion operation;
generating a matrix G according to the codes and a column vector x corresponding to the ith human eye sample imageiThe column vector x corresponding to the jth human eye sample imagejObtaining the element K of the ith row and the jth column of the correlation matrix K by the following formulaij
Kij=(GTxi)T(GTxj)xi Txj
Wherein, KijA column vector x representing the i-th human eye sample imageiThe column vector x corresponding to the jth human eye sample imagejThe degree of correlation between them, T, represents the transposition operation.
8. The method according to claim 7, wherein the obtaining parameters of the fast local linear SVM according to the support vector of the sample matrix X and the corresponding weight of the support vector comprises:
obtaining an intermediate matrix A according to the support vector of the sample matrix X, the weight corresponding to the support vector and the code generation matrix G by the following formula:
Figure FDA0002304390800000051
wherein M is the number of support vectors of the sample matrix X, sjThe jth support vector, α, representing the sample matrix XjRepresents the weight corresponding to the jth support vector, yjRepresenting a training sample label corresponding to the jth support vector;
by A' ═ A + A according to the intermediate matrix AT) A symmetric matrix A' is obtained;
performing eigenvalue decomposition on the symmetric matrix A 'to obtain eigenvalues of the symmetric matrix A' and eigenvectors corresponding to the eigenvalues;
according to the sequence of the eigenvalues of the symmetric matrix A 'from large to small, P selected eigenvalues and eigenvectors corresponding to the P selected eigenvalues are sequentially selected from the eigenvalues of the symmetric matrix A', and the P selected eigenvalues and the eigenvectors corresponding to the P selected eigenvalues are used as the parameters of the fast local linear SVM.
9. The method of claim 8, wherein determining human eye location points from the preset number of human eye candidate points through parameters of a pre-trained fast local linear Support Vector Machine (SVM) comprises:
stretching pixel block with k-th personal eye candidate point as neighborhood into vector z to be processedk
According to the vector z to be processedkAnd the candidate decision value V1 of the kth personal eye candidate point is obtained through the following formula by the P selected feature values and the feature vectors corresponding to the P selected feature values:
Figure FDA0002304390800000061
wherein λ isiRepresenting the ith selected characteristic value, qiRepresenting the feature vector corresponding to the ith selected feature value;
adding the candidate decision value V1 of the kth personal eye candidate point and the filter response value corresponding to the kth personal eye candidate point to obtain a final decision value of the kth personal eye candidate point;
and selecting the human eye candidate point with the highest final judgment value from the preset number of human eye candidate points as the human eye position point.
10. An apparatus for eye positioning, the apparatus comprising: a preprocessing unit, a filtering unit, a selecting unit and a determining unit, wherein,
the preprocessing unit is used for preprocessing the original face image to obtain an image to be processed;
the filtering unit is used for filtering each pixel point of the image to be processed according to a pre-trained average synthesis correlation filter (ASEF) template to obtain a filtering response value corresponding to each pixel point of the image to be processed;
the selecting unit is used for sequentially selecting a preset number of human eye candidate points from all pixel points of the image to be processed according to the sequence that the filtering response value corresponding to each pixel point of the image to be processed is from large to small;
and the determining unit is used for determining the human eye position points from the preset number of human eye candidate points through the parameters of the pre-trained fast local linear support vector machine SVM.
11. The apparatus of claim 10, further comprising: a detection unit, a first training unit and a second training unit, wherein,
the detection unit is used for carrying out face detection on a preset number of sample images, normalizing the detected face images to a preset size, and then carrying out Gaussian smoothing to obtain a preset number of face sample images;
the first training unit is used for training the ASEF template through a preset number of face sample images;
the second training unit is used for training the parameters of the fast local linear SVM through a preset number of face sample images.
12. The apparatus of claim 11, wherein the first training unit is configured to:
the ith human face sample image Im _ orign (x)i,yi) After windowing, performing two-dimensional fast Fourier transform on the ith windowed human face sample image to obtain a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Wherein i represents the number of the face sample image, and the value range of i is a positive integer greater than zero and less than N, where N is the preset number, (x)i,yi) The ith human face sample image representing the image domain, (u)i,vi) An ith human face sample image representing a frequency domain;
and according to the ith human face sample image Im _ orign (x)i,yi) Actual position of the human eye (x)0,y0) Acquiring a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Corresponding Gaussian impulse response after passing through the filter template
Figure FDA0002304390800000071
And responding to the Gaussian pulse with a Response _ im (x)i,yi) Performing two-dimensional fast Fourier transform to obtain the Gaussian pulse Response _ im (x)i,yi) Frequency domain Response _ im _ FFT (u)i,vi) Wherein e is a natural base number, and σ is the Gaussian impulse Response _ im (x)i,yi) The variance of (a);
and acquiring the ith human face sample image Im _ orign (x) according to the following formulai,yi) Corresponding Filter template Filter _ FFT (ui,vi),
Figure FDA0002304390800000072
Carrying out averaging operation on Filter templates corresponding to all face sample images according to the following formula to obtain an ASEF template Filter _ FFT (u, v),
Figure FDA0002304390800000073
where the symbol Σ is the sum operator.
13. The apparatus according to claim 12, wherein the first training unit is configured to train the face recognition unit according to an ith face sample image Im _ orign (x)i,yi) Actual position of the left eye of the middle human (x)0L,y0L) Acquiring a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Corresponding left-eye Gaussian impulse response after passing through the filter template
Figure FDA0002304390800000074
And responding to the left-eye Gaussian pulse by L _ Response _ im (x)i,yi) Performing two-dimensional fast Fourier transform to obtain the left-eye Gaussian pulse Response L _ Response _ im (x)i,yi) Frequency domain Response of L _ Response _ im _ FFT (u)i,vi);
And according to the ith human face sample image Im _ orign (x)i,yi) Actual position of the right eye of the middle person (x)0R,y0R) Acquiring a frequency domain image Im _ orign _ FFT (u) of the ith human face sample imagei,vi) Corresponding right-eye Gaussian impulse response after passing through the filter template
Figure FDA0002304390800000081
And responding to the Gaussian pulse of the right eye R _ Response _ im (x)i,yi) Performing two-dimensional fast Fourier transform to obtain the Gaussian impulse Response R _ Response i _ m (x) of the right eyeiyi) Frequency domain Response R _ Response i _ m FF _ T (u)ivi);
Accordingly, the ith human face sample image Im _ orign (x)i,yi) Corresponding Filter template Filter _ FFT (ui,vi) The method comprises the following steps:
the ith human face sample image Im _ orign (x)i,yi) Corresponding left eye Filter template L _ Filter _ FFT (u)i,vi) And the ith human face sample image Im _ orign (x)i,yi) Corresponding right eye Filter template R _ Filter _ FFT (u)i,vi) Wherein, in the step (A),
Figure FDA0002304390800000082
Figure FDA0002304390800000083
the ASEF template Filter _ FFT (u, v) comprises: left eye ASEF template L _ Filter _ FFT (u, v) and right eye ASEF template R _ Filter _ FFT (u, v), where,
Figure FDA0002304390800000084
Figure FDA0002304390800000085
14. the apparatus of claim 11, wherein the second training unit comprises: a dimension reduction subunit, an acquisition subunit, and a computation subunit, wherein,
the dimension reduction subunit is used for generating corresponding human eye sample images from a preset number of human face sample images and forming a sample matrix X from all the human eye sample images through dimension reduction;
the obtaining subunit is configured to obtain a correlation matrix K between any two human eye sample images according to the sample matrix X; and the number of the first and second groups,
acquiring a support vector of the sample matrix X and a weight corresponding to the support vector through the correlation matrix K and an SVM optimization algorithm;
and the calculating subunit is configured to calculate a parameter of the fast local linear SVM according to the support vector of the sample matrix X and the weight corresponding to the support vector.
15. The apparatus of claim 14, wherein the dimension reduction subunit is configured to:
setting the ith human face sample image Im _ orign (x)i,yi) Wherein i represents the image number of the human face sample, the value range of i is a positive integer greater than zero and less than N, N is the preset number, (x)i,yi) The ith human face sample image representing the image domain;
after acquiring human eye positive samples and human eye negative samples of all human face sample images, performing mean value removing and modulo two normalization processing on all human eye positive samples and human eye negative samples to generate human eye sample images corresponding to each human face sample image;
and stretching the ith human eye sample image into a column vector corresponding to the ith human eye sample image, and taking the column vector corresponding to the ith human eye sample image as the ith column vector in the sample matrix X.
16. The apparatus of claim 15, wherein the obtaining subunit is configured to:
performing singular value SVD decomposition on the sample matrix X to obtain a result shown as the following formula:
X=VΣUH
the first decomposition matrix V and the third decomposition matrix U are unitary matrixes, the second decomposition matrix sigma is a semi-positive diagonal matrix, and H represents a conjugate transpose;
and a sub-matrix of the first decomposition matrix V composed of the first t column vectors of the first decomposition matrix V
Figure FDA0002304390800000091
A sub-matrix of a second decomposition matrix Σ consisting of t rows before t columns before t rows before the second decomposition matrix Σ
Figure FDA0002304390800000092
The code generation matrix G is obtained by:
Figure FDA0002304390800000093
wherein the superscript-1 denotes the sub-matrix to the second decomposition matrix Σ
Figure FDA0002304390800000094
Carrying out inversion operation;
and generating a matrix G according to the codes and a column vector x corresponding to the ith human eye sample imageiThe column vector x corresponding to the jth human eye sample imagejObtaining the element K of the ith row and the jth column of the correlation matrix K by the following formulaij
Kij=(GTxi)T(GTxj)xi Txj
Wherein, KijA column vector x representing the i-th human eye sample imageiThe column vector x corresponding to the jth human eye sample imagejThe degree of correlation between them, T, represents the transposition operation.
17. The apparatus of claim 16, wherein the computing subunit is configured to:
obtaining an intermediate matrix A according to the support vector of the sample matrix X, the weight corresponding to the support vector and the code generation matrix G by the following formula:
Figure FDA0002304390800000101
wherein M is the number of support vectors of the sample matrix X, sjThe jth support vector, α, representing the sample matrix XjRepresents the weight corresponding to the jth support vector, yjRepresenting a training sample label corresponding to the jth support vector;
and, according to said intermediate matrix a, by a' ═ a + aT) A symmetric matrix A' is obtained;
performing eigenvalue decomposition on the symmetric matrix A 'to obtain eigenvalues of the symmetric matrix A' and eigenvectors corresponding to the eigenvalues;
and sequentially selecting P selected eigenvalues and eigenvectors corresponding to the P selected eigenvalues from the eigenvalues of the symmetric matrix A 'according to the descending order of the eigenvalues of the symmetric matrix A', and taking the P selected eigenvalues and the eigenvectors corresponding to the P selected eigenvalues as the parameters of the fast local linear SVM.
18. The apparatus of claim 17, wherein the determining unit is configured to:
stretching pixel block with k-th personal eye candidate point as neighborhood into vector z to be processedk
And according to the vector z to be processedkAnd the candidate decision value V1 of the kth personal eye candidate point is obtained through the following formula by the P selected feature values and the feature vectors corresponding to the P selected feature values:
Figure FDA0002304390800000111
wherein λ isiRepresenting the ith selected characteristic value, qiRepresenting the feature vector corresponding to the ith selected feature value;
adding the candidate judgment value V1 of the kth personal eye candidate point and the filter response value corresponding to the kth personal eye candidate point to obtain a final judgment value of the kth personal eye candidate point;
and selecting the human eye candidate point with the highest final judgment value from the preset number of human eye candidate points as the human eye position point.
CN201410388103.XA 2014-08-07 2014-08-07 Method and equipment for positioning human eyes Active CN105469018B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201410388103.XA CN105469018B (en) 2014-08-07 2014-08-07 Method and equipment for positioning human eyes
PCT/CN2015/072055 WO2016019715A1 (en) 2014-08-07 2015-01-30 Human eye locating method and device and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410388103.XA CN105469018B (en) 2014-08-07 2014-08-07 Method and equipment for positioning human eyes

Publications (2)

Publication Number Publication Date
CN105469018A CN105469018A (en) 2016-04-06
CN105469018B true CN105469018B (en) 2020-03-13

Family

ID=55263096

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410388103.XA Active CN105469018B (en) 2014-08-07 2014-08-07 Method and equipment for positioning human eyes

Country Status (2)

Country Link
CN (1) CN105469018B (en)
WO (1) WO2016019715A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106548194B (en) * 2016-09-29 2019-10-15 中国科学院自动化研究所 The construction method and localization method of two dimensional image human joint points location model
CN110168599B (en) * 2017-10-13 2021-01-29 华为技术有限公司 Data processing method and terminal

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101533466A (en) * 2009-04-09 2009-09-16 南京壹进制信息技术有限公司 Image processing method for positioning eyes
CN102129553A (en) * 2011-03-16 2011-07-20 上海交通大学 Method for eye detection based on single infrared light supply
CN102314598A (en) * 2011-09-22 2012-01-11 西安电子科技大学 Retinex theory-based method for detecting human eyes under complex illumination
CN102456129A (en) * 2010-10-26 2012-05-16 同方威视技术股份有限公司 Image deviation-rectifying method and system for safety inspection
CN102867172A (en) * 2012-08-27 2013-01-09 Tcl集团股份有限公司 Human eye positioning method, system and electronic equipment
CN102930258A (en) * 2012-11-13 2013-02-13 重庆大学 Face image recognition method
KR20130085316A (en) * 2012-01-19 2013-07-29 한국전자통신연구원 Apparatus and method for acquisition of high quality face image with fixed and ptz camera
CN103632136A (en) * 2013-11-11 2014-03-12 北京天诚盛业科技有限公司 Method and device for locating human eyes
CN103955719A (en) * 2014-05-20 2014-07-30 中国科学院信息工程研究所 Filter bank training method and system and image key point positioning method and system

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8520956B2 (en) * 2009-06-09 2013-08-27 Colorado State University Research Foundation Optimized correlation filters for signal processing

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101533466A (en) * 2009-04-09 2009-09-16 南京壹进制信息技术有限公司 Image processing method for positioning eyes
CN102456129A (en) * 2010-10-26 2012-05-16 同方威视技术股份有限公司 Image deviation-rectifying method and system for safety inspection
CN102129553A (en) * 2011-03-16 2011-07-20 上海交通大学 Method for eye detection based on single infrared light supply
CN102314598A (en) * 2011-09-22 2012-01-11 西安电子科技大学 Retinex theory-based method for detecting human eyes under complex illumination
KR20130085316A (en) * 2012-01-19 2013-07-29 한국전자통신연구원 Apparatus and method for acquisition of high quality face image with fixed and ptz camera
CN102867172A (en) * 2012-08-27 2013-01-09 Tcl集团股份有限公司 Human eye positioning method, system and electronic equipment
CN102930258A (en) * 2012-11-13 2013-02-13 重庆大学 Face image recognition method
CN103632136A (en) * 2013-11-11 2014-03-12 北京天诚盛业科技有限公司 Method and device for locating human eyes
CN103955719A (en) * 2014-05-20 2014-07-30 中国科学院信息工程研究所 Filter bank training method and system and image key point positioning method and system

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
A New Implementation Method of ASEF for Eye Detection;So-Hee Park等;《IEEE》;20121231;全文 *
基于HOG特征的人脸识别***研究;慕春雷;《中国优秀硕士学位论文全文数据库 信息科技辑》;20140115(第01期);第19-21,39,53-54页 *

Also Published As

Publication number Publication date
WO2016019715A1 (en) 2016-02-11
CN105469018A (en) 2016-04-06

Similar Documents

Publication Publication Date Title
CN108182384B (en) Face feature point positioning method and device
CN107871100B (en) Training method and device of face model, and face authentication method and device
Zhou et al. Bayesian low-tubal-rank robust tensor factorization with multi-rank determination
WO2016138838A1 (en) Method and device for recognizing lip-reading based on projection extreme learning machine
AU2014350727B2 (en) Face positioning method and device
CN104517104A (en) Face recognition method and face recognition system based on monitoring scene
CN108154133B (en) Face portrait-photo recognition method based on asymmetric joint learning
CN107871103B (en) Face authentication method and device
CN110110189A (en) Method and apparatus for generating information
KR20080035711A (en) Global feature extraction method for 3d face recognition
US11847821B2 (en) Face recognition network model with face alignment based on knowledge distillation
KR20170024303A (en) System and method for detecting feature points of face
KR20190061538A (en) Method and apparatus of recognizing motion pattern base on combination of multi-model
CN105469018B (en) Method and equipment for positioning human eyes
Roomi et al. Coin detection and recognition using neural networks
CN103927529B (en) The preparation method and application process, system of a kind of final classification device
KR20130067612A (en) Feature vector classifier and recognition device using the same
WO2017129115A1 (en) Orientation-based subject-matching in images
CN104899565B (en) Eye movement recognition methods and device based on textural characteristics
Kaur et al. Illumination invariant face recognition
KR102014093B1 (en) System and method for detecting feature points of face
Angulu et al. Human age estimation using multi-frequency biologically inspired features (MF-BIF)
CN108256405A (en) A kind of face identification method and device
CN113627446A (en) Image matching method and system of feature point description operator based on gradient vector
Widiastuti et al. Scattered object recognition using Hu Moment invariant and backpropagation neural network

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant