CN105224935A - A kind of real-time face key point localization method based on Android platform - Google Patents

A kind of real-time face key point localization method based on Android platform Download PDF

Info

Publication number
CN105224935A
CN105224935A CN201510713055.1A CN201510713055A CN105224935A CN 105224935 A CN105224935 A CN 105224935A CN 201510713055 A CN201510713055 A CN 201510713055A CN 105224935 A CN105224935 A CN 105224935A
Authority
CN
China
Prior art keywords
key point
face
training
sample
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510713055.1A
Other languages
Chinese (zh)
Other versions
CN105224935B (en
Inventor
刘青山
王东
杨静
邓健康
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Information Science and Technology
Original Assignee
Nanjing University of Information Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Information Science and Technology filed Critical Nanjing University of Information Science and Technology
Priority to CN201510713055.1A priority Critical patent/CN105224935B/en
Publication of CN105224935A publication Critical patent/CN105224935A/en
Application granted granted Critical
Publication of CN105224935B publication Critical patent/CN105224935B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • G06V40/171Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses a kind of real-time face key point localization method based on Android platform, belong to technical field of computer vision.The inventive method comprises the following steps: collect face training plan sheet collection, and demarcate key point; N sample in random selecting training set is as the original shape of each training sample; Calculate each training sample standardization target; Extract the shape indexing feature of each key point; Correlation analysis is adopted to select appropriate feature; Adopt two-layer enhancing regressive structure (exterior layer and interior layer); Calculate the recurrence device in each stage; The method of Face datection is used to estimate face window, according to the forecast of regression model face key point position trained.Current existing method computation complexity is higher, runs slow on a mobile platform; And to noise-sensitive, the precision of location is low.The present invention carrys out constraint shapes with the linear combination of sample, applies the precision and the efficiency that improve face key point location based on the method returned.

Description

A kind of real-time face key point localization method based on Android platform
Technical field
The present invention relates to technical field of computer vision, relate to a kind of face key point localization method.
Background technology
Face datection and key point location technology, as the gordian technique of computer vision research, have been widely used in the aspects such as intelligent monitoring, identification, Expression analysis at present.Face key point location refers to specific face organ such as the eyes, face, nose etc. that accurately navigate to people in facial image, and obtains its geometric parameter, thus provides information accurately for the research such as Expression analysis or recognition of face.
At present, the location of face key point is existing on computers to be realized preferably, real-time and accuracy are all very high, and representational work has active shape model (ActiveShapeModel, ASM), Bayes tangent space shape (BayesianTangentShapeModel, BTSM), active appearance models (ActiveAppearanceModel, AAM), affined partial model (ConstrainedLocalModel, CLM) etc.But existing algorithm computation complexity is higher, rarely has application calculating, on the limited mobile platform of storage capacity.
Nowadays, in state-owned several hundred million intelligent terminal (mainly mobile phone, panel computer) user, intelligent terminal becomes the information carrier of people gradually.Mobile platform performance have also been obtained and significantly promotes, and this is just for human face characteristic point real time location tracking on mobile platform provides possibility.But the general computation complexity of current existing facial modeling track algorithm is higher, memory consumption is comparatively large, processing speed is comparatively slow, is difficult to be grafted directly to mobile platform.
Summary of the invention
In order to solve the problem, the invention discloses high-level efficiency, high-precision " explicit shape recurrence " face key point localization method, predicting the key point of whole face by directly learning a vector regression function.Intrinsic shape constraining is encoded in our cascade learning framework naturally, and is applied in the process of test by from coarse to fine.The present invention tries hard to the performance at many aspects General Promotion algorithm, be embodied in Feature Selection and larger improvement has been done in model training aspect, algorithm operational efficiency is significantly promoted, and ensure that the accuracy of location simultaneously, the detection and location carrying out human face characteristic point on a mobile platform in real time can be realized.
The present invention is divided into two stages of training and testing.Training stage mainly learns regression model.Be different from most of homing method, this method is not the shape with preset parameter.The present invention predicts the key point of whole face by directly learning a vector regression function, then on training set, explicitly minimizes positioning error.
In order to realize efficient recurrence, we use simple pixel difference feature, i.e. the intensity difference of two pixels in the picture.Such feature calculation complexity is very little, is exactly the position according to key point and a side-play amount, obtains the pixel value of this position, then the difference of the pixel that calculating two is such, thus obtains shape indexing feature.What adopt in the method is local coordinate but not global coordinate system, enhances the robustness of feature greatly.
In order to make pixel difference feature keep geometric invariance, by estimating that the shape obtained carrys out index pixel.Here index pixel Δ is carried out with local coordinate l=(Δ x l, Δ y l), l is a standardization key point on the face.Such indexed mode can to yardstick, and the changes such as rotation maintain the invariance and make algorithm robust more.This index also can help us around the point of some static state, list more how useful feature (point of such as eye center is darker than nose, but the center of two eyes is again similar) simultaneously.In the process that reality realizes, we convert back local coordinate system in the global coordinate system of original image and obtain shape indexing pixel, then in calculating pixel difference feature.Such method makes the travelling speed of test phase faster.If the estimation shape of a sample is S.So π is passed through in the position of m key point lο S obtains, π here lthe x of m key point is obtained, y coordinate from shape vector.Δ lfor local coordinate system, corresponding global coordinate system can be expressed as so on the original image for different sample Δs lbe identical, but the global coordinate system for extracting underlying pixel data point correspondingly can regulate and ensure geometric invariance.
The enhancing that present invention uses a two-stage returns device, the i.e. first order 10 grades, the second layer 500 grades.In this secondary structure, in the first order, each node is the cascade of 500 Weak Classifiers, namely the recurrence device of a second layer.Return in device at the second layer, feature remains unchanged, and in ground floor, feature is change.At ground floor, the output of each node is the input of a upper node.It is all the feature of getting in the key point of upper level estimation.
The present invention adopts a two-layer reinforcement Regression Model, i.e. basic face key point positioning framework (exterior layer) and stage regression device R t(interior layer) cascade mode.Interior layer stage regression device is also known as making original recurrence device.The key difference that exterior layer and interior layer return device is that shape indexing feature is fixing in interior layer returns.Namely feature only corresponds to the front shape S once estimated t-1and remained unchanged before current original recurrence device succeeds in school.The training stage can be made more to save time for shape indexing feature invariant in keeping interior layer to return and to make the interior layer that learns return device more stable.The structure returning device in the present invention can be described as T=10, K=500, i.e. ground floor (exterior layer) 10 grades, the second layer (interior layer) 500 grades.Each node in ground floor is the cascade of 500 Weak Classifiers, and namely the second layer returns device.The input of each node of ground floor is the output of a upper node.The second layer returns device and adopts so-called fern structure.Each fern is the combination of 5 characteristic sum threshold values, and feature space is divided into 2 findividual bins.
Each binb corresponds to one and returns output y b: Ω bthe sample of representative in b bin.Y boptimum solution be a contraction factor is introduced in order to prevent over-fitting when training when the sample size in bin is enough large, beta impact is less, otherwise then can play lower estimate and be worth amplitude.
Original recurrence device: the present invention fern represents original recurrence device, a fern is the combination of F characteristic sum threshold value, and feature space and all training samples are divided into 2 finside individual bin.Each bin and one returns and exports y bclosely related. represent the regressive object of training sample, so the output of a bin be exactly minimize fall into this bin all training samples to the mean square distance of regressive object: Ω bit is the sample in b bin.Its optimum solution gets average to all regressive objects in bin: in order to avoid over-fitting introduces a contraction factor β: can find out that contraction factor β does not almost affect when number of training is enough large time, otherwise β can reduce the magnitude of estimation.
The method operating process that this patent is complete is as follows:
Step 1, collection training pictures, and key point is demarcated to training set face picture;
Step 2, calculate each training sample standardization target: m ο S is standardized target sample;
Step 3, combined training sample set each sample packages contains a training image, a true shape and an original shape; N sample in random selecting training set is as the original shape of each training sample;
Step 4, extract shape indexing feature according to the position of each key point;
Step 5, employing correlation analysis select appropriate feature as final feature;
Step 6, regressive structure is set: adopt 2-level to strengthen regressive structure as basic face registration framework (exterior layer), inner stage regression device R is set at every one deck exterior layer t(interior layer);
Step 7, calculate the recurrence device R in each stage t:
The method of step 8, use Face datection estimates face window; Forecast of regression model face key point position according to training:
Accompanying drawing explanation
Fig. 1 is the basic procedure schematic diagram of the real-time face key point localization method that the present invention is based on Android platform.
Embodiment
Below in conjunction with accompanying drawing, technical scheme of the present invention is described in detail.The invention discloses a kind of real-time face key point localization method based on Android platform.
As shown in Figure 1, the present invention can be divided into learning position and two stages of test position fix, specifically, comprises the steps:
The mark of step 1, training set unique point and standardization: the face calibration point related in the present invention is chosen according to MPEG-4 standard, 84 unique points defined due to face defined parameters FDP are too numerous and diverse, therefore we have chosen wherein 68 unique points as required, and the unique point on eye contour also only remains eight (i.e. eyes each four points in upper and lower, left and right near the eyes) to improve operation efficiency.The facial image of MUCT and self-calibration is taken from training storehouse, and total about 4,000 width facial images, carry out specular process to these facial images, final acquisition has the training set of 8,000 width facial images.
Step 2, standardization is carried out to sample image shape: because the attitude of face shape in image is by large small scale, the anglec of rotation, location parameter describes, therefore needing the attitude parameter by suitably changing shape, comprising size, angle, positions etc. make all images reach consistent as far as possible, such as change of scale can allow the distance between calibration point fix, and it is fixing that rotational transform can allow calibration point line point to, and shift transformation makes centroid position fix.First we try to achieve the average face of all training samples then minimize input amendment S with between L2 distance m ο S is standardized shape
Step 3, combined training sample set each sample packages contains a training image, a true shape and an original shape.10 samples in random selecting training set are as the original shape of each training sample.Intangibly add the quantity of training sample with this method producing multiple original shape to a width picture, also improve the generalization of training simultaneously.
Step 4, a certain specified point on standardized face set up local coordinate system Δ l=(Δ x l, Δ y l).For different training samples, local coordinate system Δ lidentical.According to the location index pixel of local coordinate system key point.Subscript l represents the concrete mark point corresponding to pixel.Then local coordinate system is converted back original image global coordinate system and extract pixel difference feature π lrepresent the x extracting m mark point from shape vector, y coordinate.Each weak learner R tfeature based on image I and last estimation shape S t-1.At each stage regression device R tin, a stochastic generation P local coordinate system namely P shape indexing pixel is defined.First by random selecting one mark point (the l αindividual mark point) generate each local coordinate system, then produce X-and Y-skew at random in a uniformly distributed manner.P pixel one is met together
Produce P 2individual pixel difference feature.
Step 5, employing correlation analysis select appropriate feature as final feature.
Creating P 2after individual feature, bring new challenge---the effective feature of fast selecting in huge feature space." in right amount " means and selects much smaller than P 2characteristic Number, be beneficial to less computation complexity like this, improve the efficiency of algorithm.The present invention chooses F/P 2feature is set up good fern and is returned device.We utilize the correlativity between characteristic sum regressive object to carry out selected characteristic.Selected characteristic has following two principles: keep high correlation between each the characteristic sum regressive object in fern; Maintenance between feature to feature is low relevant, in complementary relationship.Representing regressive object with Y, is N (sample number) row N fpthe matrix that (mark is counted) arranges.X represents pixel difference feature, is a capable P of N 2the matrix of row.Our target is the P at X matrix 2select in row and list with the F of Y matrix height correlation.Since Y is matrix, Y, in order to the individual column vector projection υ obtained from unit Gauss, is projected to a column vector Y by us probupper: Y prob=Y υ.(PearsonCorrelation) feature the most relevant to projection target can be selected out: repeat to obtain F the feature be applicable to for F time with multiple projections.This F feature is the appropriate feature that will choose out.Correlation calculations corr (the Y of regressive object and pixel difference feature proj, ρ mn) design as follows: c o r r ( Y p r o j , ρ m - ρ n ) = cov ( Y p r o j , ρ m ) - cov ( Y p r o j , ρ n ) σ ( Y p r o j ) σ ( ρ m - ρ n ) , Wherein
σ(ρ mn)=cov(ρ mm)+cov(ρ nn)-2cov(ρ mn)。Here the calculating of correlativity is made up of two parts: target-pixel covariance and pixel-pixel covariance.Because shape indexing feature is changeless in interior layer cascade returns, then reused in each interior layer so pixel-pixel covariance can calculate in advance.And for original recurrence device, we only need to calculate all target-pixel covariance to form correlativity, correlativity here and the number of pixel characteristic linear.Therefore, correlation calculations complexity is from O (NP 2) drop to O (NP).
Step 6, regressive structure is set: adopt 2-level to strengthen regressive structure as basic face registration framework (exterior layer), inner stage regression device R is set at every one deck exterior layer t(interior layer).The weak learner of interior layer is called as original recurrence device.Here inside and outside recurrence utensil has some similaritys, but their difference be interior layer all stage regression wherein shape indexing feature remain unchanged.
Step 7, learn inner cascade and return, inner cascade returns and comprises K original recurrence device { r 1..., r k, i.e. ferns.These original recurrence devices wolfishly matching regressive object successively.In each original recurrence device process recurrence device carry over remnants.In iterative process each time, these remnants are by as the target being used for learning new recurrence device.Calculate the recurrence device R in each stage t:
The method of step 8, use Face datection estimates face window; Forecast of regression model face key point position according to training:

Claims (1)

1., based on a real-time face key point localization method for Android platform, it is characterized in that structure one is based on the non-ginseng shape returned, and predicting the key point of whole face, comprising the following steps by directly learning a vector regression function:
Step 1, collection training pictures, and key point is demarcated to training set face picture;
Step 2, calculate each training sample standardization target: be standardized target sample;
Step 3, combined training sample set each sample packages contains a training image, a true shape and an original shape; N sample in random selecting training set is as the original shape of each training sample;
Step 4, extract shape indexing feature according to the position of each key point;
Step 5, employing correlation analysis select appropriate feature as final feature;
Step 6, regressive structure is set: adopt 2-level to strengthen regressive structure as basic face registration framework (exterior layer), inner stage regression device R is set at every one deck exterior layer t(interior layer);
Step 7, calculate the recurrence device R in each stage t:
The method of step 8, use Face datection estimates face window; Forecast of regression model face key point position according to training:
CN201510713055.1A 2015-10-28 2015-10-28 A kind of real-time face key independent positioning method based on Android platform Active CN105224935B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510713055.1A CN105224935B (en) 2015-10-28 2015-10-28 A kind of real-time face key independent positioning method based on Android platform

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510713055.1A CN105224935B (en) 2015-10-28 2015-10-28 A kind of real-time face key independent positioning method based on Android platform

Publications (2)

Publication Number Publication Date
CN105224935A true CN105224935A (en) 2016-01-06
CN105224935B CN105224935B (en) 2018-08-24

Family

ID=54993895

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510713055.1A Active CN105224935B (en) 2015-10-28 2015-10-28 A kind of real-time face key independent positioning method based on Android platform

Country Status (1)

Country Link
CN (1) CN105224935B (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105469081A (en) * 2016-01-15 2016-04-06 成都品果科技有限公司 Face key point positioning method and system used for beautifying
CN106096560A (en) * 2016-06-15 2016-11-09 广州尚云在线科技有限公司 A kind of face alignment method
CN106960203A (en) * 2017-04-28 2017-07-18 北京搜狐新媒体信息技术有限公司 A kind of facial feature tracking method and system
CN107463865A (en) * 2016-06-02 2017-12-12 北京陌上花科技有限公司 Face datection model training method, method for detecting human face and device
CN109635752A (en) * 2018-12-12 2019-04-16 腾讯科技(深圳)有限公司 Localization method, face image processing process and the relevant apparatus of face key point
CN109800635A (en) * 2018-12-11 2019-05-24 天津大学 A kind of limited local facial critical point detection and tracking based on optical flow method
CN110781765A (en) * 2019-09-30 2020-02-11 腾讯科技(深圳)有限公司 Human body posture recognition method, device, equipment and storage medium
CN111667403A (en) * 2020-07-02 2020-09-15 北京爱笔科技有限公司 Method and device for generating face image with shielding
WO2020248789A1 (en) * 2019-06-11 2020-12-17 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Method and system for facial landmark detection using facial component-specific local refinement
CN112115845A (en) * 2020-09-15 2020-12-22 中山大学 Active shape model parameterization method for face key point detection

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8837839B1 (en) * 2010-11-03 2014-09-16 Hrl Laboratories, Llc Method for recognition and pose estimation of multiple occurrences of multiple objects in visual images
CN103824049A (en) * 2014-02-17 2014-05-28 北京旷视科技有限公司 Cascaded neural network-based face key point detection method
CN104063700B (en) * 2014-07-04 2017-08-18 武汉工程大学 The method of eye center point location in natural lighting front face image
CN104899575A (en) * 2015-06-19 2015-09-09 南京大学 Human body assembly dividing method based on face detection and key point positioning

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105469081B (en) * 2016-01-15 2019-03-22 成都品果科技有限公司 A kind of face key independent positioning method and system for U.S. face
CN105469081A (en) * 2016-01-15 2016-04-06 成都品果科技有限公司 Face key point positioning method and system used for beautifying
CN107463865B (en) * 2016-06-02 2020-11-13 北京陌上花科技有限公司 Face detection model training method, face detection method and device
CN107463865A (en) * 2016-06-02 2017-12-12 北京陌上花科技有限公司 Face datection model training method, method for detecting human face and device
CN106096560A (en) * 2016-06-15 2016-11-09 广州尚云在线科技有限公司 A kind of face alignment method
CN106960203A (en) * 2017-04-28 2017-07-18 北京搜狐新媒体信息技术有限公司 A kind of facial feature tracking method and system
CN109800635A (en) * 2018-12-11 2019-05-24 天津大学 A kind of limited local facial critical point detection and tracking based on optical flow method
CN109635752A (en) * 2018-12-12 2019-04-16 腾讯科技(深圳)有限公司 Localization method, face image processing process and the relevant apparatus of face key point
WO2020248789A1 (en) * 2019-06-11 2020-12-17 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Method and system for facial landmark detection using facial component-specific local refinement
US20220092294A1 (en) * 2019-06-11 2022-03-24 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Method and system for facial landmark detection using facial component-specific local refinement
CN110781765A (en) * 2019-09-30 2020-02-11 腾讯科技(深圳)有限公司 Human body posture recognition method, device, equipment and storage medium
CN110781765B (en) * 2019-09-30 2024-02-09 腾讯科技(深圳)有限公司 Human body posture recognition method, device, equipment and storage medium
CN111667403A (en) * 2020-07-02 2020-09-15 北京爱笔科技有限公司 Method and device for generating face image with shielding
CN112115845A (en) * 2020-09-15 2020-12-22 中山大学 Active shape model parameterization method for face key point detection
CN112115845B (en) * 2020-09-15 2023-12-29 中山大学 Active shape model parameterization method for face key point detection

Also Published As

Publication number Publication date
CN105224935B (en) 2018-08-24

Similar Documents

Publication Publication Date Title
CN105224935A (en) A kind of real-time face key point localization method based on Android platform
CN111795704B (en) Method and device for constructing visual point cloud map
CN111780763B (en) Visual positioning method and device based on visual map
Yousefhussien et al. A multi-scale fully convolutional network for semantic labeling of 3D point clouds
CN112862874B (en) Point cloud data matching method and device, electronic equipment and computer storage medium
US20150006117A1 (en) Learning Synthetic Models for Roof Style Classification Using Point Clouds
Cui et al. 3D semantic map construction using improved ORB-SLAM2 for mobile robot in edge computing environment
CN112949647B (en) Three-dimensional scene description method and device, electronic equipment and storage medium
Ding et al. Vehicle pose and shape estimation through multiple monocular vision
Alidoost et al. Knowledge based 3D building model recognition using convolutional neural networks from LiDAR and aerial imageries
CN109063549A (en) High-resolution based on deep neural network is taken photo by plane video moving object detection method
CN104463962B (en) Three-dimensional scene reconstruction method based on GPS information video
JP2019185787A (en) Remote determination of containers in geographical region
CN115457492A (en) Target detection method and device, computer equipment and storage medium
Li et al. An aerial image segmentation approach based on enhanced multi-scale convolutional neural network
Zhuang et al. Instance segmentation based 6D pose estimation of industrial objects using point clouds for robotic bin-picking
Gao et al. Road extraction using a dual attention dilated-linknet based on satellite images and floating vehicle trajectory data
Liu et al. A new multi-channel deep convolutional neural network for semantic segmentation of remote sensing image
Li et al. GeoImageNet: a multi-source natural feature benchmark dataset for GeoAI and supervised machine learning
CN111738074A (en) Pedestrian attribute identification method, system and device based on weak supervised learning
Firouznia et al. Chaotic particle filter for visual object tracking
CN117132737B (en) Three-dimensional building model construction method, system and equipment
CN117132649A (en) Ship video positioning method and device for artificial intelligent Beidou satellite navigation fusion
Yudin et al. Hpointloc: point-based indoor place recognition using synthetic RGB-D images
Sayed et al. Point clouds reduction model based on 3D feature extraction

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CP02 Change in the address of a patent holder
CP02 Change in the address of a patent holder

Address after: 210000 No. 219 Ning six road, Jiangbei new district, Nanjing, Jiangsu

Patentee after: NANJING University OF INFORMATION SCIENCE & TECHNOLOGY

Address before: 210000 No. 69 Olympic Sports street, Jianye District, Jiangsu, Nanjing

Patentee before: NANJING University OF INFORMATION SCIENCE & TECHNOLOGY