CN106778628A - A kind of facial expression method for catching based on TOF depth cameras - Google Patents

A kind of facial expression method for catching based on TOF depth cameras Download PDF

Info

Publication number
CN106778628A
CN106778628A CN201611190469.1A CN201611190469A CN106778628A CN 106778628 A CN106778628 A CN 106778628A CN 201611190469 A CN201611190469 A CN 201611190469A CN 106778628 A CN106778628 A CN 106778628A
Authority
CN
China
Prior art keywords
facial expression
image
model
face
depth cameras
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201611190469.1A
Other languages
Chinese (zh)
Inventor
张维忠
袁翠梅
黄松
周绍致
王青林
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to CN201611190469.1A priority Critical patent/CN106778628A/en
Publication of CN106778628A publication Critical patent/CN106778628A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/174Facial expression recognition
    • G06V40/176Dynamic expression
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T13/00Animation
    • G06T13/203D [Three Dimensional] animation
    • G06T13/403D [Three Dimensional] animation of characters, e.g. humans, animals or virtual beings
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2200/00Indexing scheme for image data processing or generation, in general
    • G06T2200/08Indexing scheme for image data processing or generation, in general involving all processing steps from image acquisition to 3D model generation

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Processing Or Creating Images (AREA)
  • Image Processing (AREA)

Abstract

The invention provides a kind of facial expression method for catching based on TOF depth cameras, it includes:The first step, facial expression seizure is carried out using TOF depth cameras to real person;Second step, processes the facial expression data for catching, and generate facial expression seizure file;3rd step, builds three-dimensional personage's geometrical model, imports facial expression and catches file, driving model.The method is directly recognized and is recorded the exercise data of facial key point without hand labeled Marker using the method for image recognition.The data of its capture are the real motion data of face, and action is coherent true to nature, and expression is fine and smooth abundant, and the facial expression that gets of the method catches file and can reuse, and has saved cost.

Description

A kind of facial expression method for catching based on TOF depth cameras
Technical field
Single TOF depths are utilized the present invention relates to computer vision and computer graphics techniques field, more particularly to one kind Spend the facial expressions and acts of cameras capture face face and be stored as specific data form and caught for the facial expression of cartoon model etc. Catch method.
Background technology
In recent years, with the development of computer graphics, the facial expression animation of Computer Animated Graph synthesis people is to work as One of emphasis for preceding animation disciplinary study, it enables animator with the expression direct drive picture image mould of performer Type, receives animators and more pays close attention to.Facial expression catches the expression that can in real time detect, record performing artist, Digitized " abstract expression " is converted into, so that animation software is by its " imparting " model, model is had as performing artist Expression, and generate final expression animation sequence.
At present, facial expression animation can generally be divided into based on geometry, based on image, the research side based on motion tracking Method.Wherein the research method based on geometry is mainly including keyframe interpolation method, parametric method, muscle model method etc.;Based on figure The method of picture mainly includes:Anamorphose, dynamic texture mapping, expression of complexion change etc..
In the research method based on geometry, keyframe interpolation method sets up geometrical model at two different moment first, Then enter row interpolation between the two models and obtain the model of intermediate time, so as to obtain facial expression animation, this is a kind of Most traditional, most original face cartoon method, while be also a kind of face cartoon method of standard, but its weak point is Need very big workload.Parametric method is in human face animation (Parke F I.A by Parke first Applications Parameteric Model for Human Faces [D].Salt Lake City:University of Utah, 1974) action when, lip is spoken is suitable in this way, and the deficiency of this method is the human face expression for producing inadequate certainly So, and in arrange parameter value substantial amounts of manual setting is needed.Muscle model method is by one group of bullet of composition muscle model Property muscle contraction control face surface mesh to change and then simulate human face expression.
In the method based on image, anamorphose one kind is that character pair line segment hand-manipulated comes between two images Realize that 2D deforms;Another kind is to be mapped to the parameter space of 2D by the 3D models for deforming to realize 3D model deformations and pass through The 3D conversion and 2D deformations of geometrical model are combined, using 3D geometry interpolations, and image are carried out between correspondence texture image Deformation operation obtains real facial expression animation.In addition, also based on Bezier indicatrixes and based on radial direction base The anamorphose scheduling algorithm of neutral net.Dynamic texture mapping relies on the texture mapping of viewpoint, and it is allowed using different every time Texture maps, are dynamically adjusted by drawing repeatedly for model come the mediation weights to current view point, and deficiency is the meter for needing Calculate and amount of ram is big, and multi-texturing is merged if independently of viewpoint, because there is error and usually make texture in record and sampling Thicken.
At present, most widely used is that, based on motion tracking method, motion tracking method mainly uses motion capture system To carry out expression seizure, it mainly uses existing capture data to move on object module the human face expression of source model, So as to the facial expression for realizing face is captured, facial expression true to nature is obtained.
Existing movement capturing technology is broadly divided into four major classes according to seizure equipment is different with principle:Mechanically, acoustics formula, Electricity formula and optical motion catch, wherein, data are convenient, sampling precision is high, frequency with obtaining for optical motion capture mode Rate is high, using scope it is wide the advantages of, the data that it is gathered are sequence identifier point (Marker) the point set data in units of frame, Face's key point of performing artist sticks Marker, and vision system will recognize and process these Maker, it is possible to realize that expression is caught Catch.
The present invention uses the method based on motion tracking, different from the method for catching of traditional optical profile type.The present invention Without hand labeled Marker in capture-process of expressing one's feelings, directly recognize and record facial key point using the method for image recognition Exercise data.It mainly uses existing capture data to move on object module the human face expression of source model, so that The facial expression capture of face is realized, facial expression true to nature is obtained.
The content of the invention
For the demand that the industries such as current animation, film are captured to facial expression, the application provides a kind of based on TOF depth The facial expression method for catching of camera, without hand labeled Marker, is directly recognized and recording surface using the method for image recognition The exercise data of portion's key point.The data of its capture are the real motion data of face, and action is coherent true to nature, and expression is fine and smooth rich Richness, and the facial expression that gets of the method catches file and can reuse, and has saved cost.
In order to solve the above technical problems, the invention provides a kind of facial expression method for catching based on TOF depth cameras, It includes:
The first step, using TOF depth camera sampling depth images and coloured image, using AAM (Active Appearances Models) algorithm positioning face feature point, and calculate acquisition facial expression data;
Second step, treatment catches facial expression data, and generation facial expression catches file;
3rd step, builds three-dimensional personage's geometrical model, imports facial expression and catches file, driving model.
Wherein, the face feature point of real person is demarcated using AAM algorithms in the first step, in the feelings that head pose is different Under condition, we also can accurately be positioned to face feature point, so as to carry out facial expression seizure to real person.
Wherein, the first step is further specifically included:
A, using TOF depth cameras obtain face coloured image and depth image;
B, set up for head pose judge random regression forest model;
C, the visual angle model trained and set up AAM algorithms;
D, using AAM algorithms position face feature point;
E, the three-dimensional coordinate for obtaining face feature point.
Wherein, a steps are specially connection TOF depth cameras, switch on power, the data control process list of TOF depth cameras Unit sends open command, and its body-sensing camera is opened, and it is single with treatment that color data stream and depth data are streamed into data control Unit, data control changes color data circulation into coloured image with processing unit, changes depth data circulation into depth image.
Wherein, b steps are specially and obtain the corresponding depth image of different head poses using TOF depth cameras, so The depth image that will be got according to the position of head pose afterwards is divided into multiple different classification based training collection, is approximately put down using face The normal direction in face represents different head poses, so as to set up the random regression forest model of head pose judgement.
The present invention be extended on the basis of original AAM algorithms, for when portion's characteristic point is positioned over there not Same head pose, sets up corresponding AAM visual angles model respectively, and each AAM visual angles model is divided into shape and texture again Model two parts.
Wherein, the d steps are further specially and are input into new image, what the head pose set up using step b was judged Random regression forest model, the new depth image to being given judges head position therein and face's direction, according to what is obtained Immediate AAM visual angles model therewith is selected in the AAM visual angles model that head pose information is set up from c steps.Using the AAM for choosing Model carries out the positioning of face feature point to current input picture.
Wherein, after the e steps are further specially the position for determining face feature point, calculate and obtain face face special Levy coordinate information a little, the coordinate system of TOF depth cameras is that, with its own as the origin of coordinates, front is Z-direction, to the left for The positive direction of X-axis, is upwards the positive direction of Y-axis.
Beneficial effects of the present invention:
The application provides a kind of facial expression method for catching based on TOF depth cameras, and the data of the method capture are existing The real exercise data of real world's face, action is coherent true, and expression is fine and smooth abundant, and the facial table that the method gets Feelings catch file and can reuse, and have saved cost.
Brief description of the drawings
Fig. 1 is the coloured image that TOF depth cameras get;
Fig. 2 is the depth image that TOF depth cameras get;
Fig. 3 is the training sample of hand labeled;
Fig. 4 is the characteristic point distribution of face and title;
Fig. 5 is that expression catches the design sketch that file imports faceform;
Fig. 6 is to set up the depth image block used by head pose judgment models.
Specific embodiment
The invention provides a kind of facial expression method for catching based on TOF depth cameras, it includes:
The first step, using TOF depth camera sampling depth images and coloured image, using AAM (Active Appearances Models) algorithm positioning face feature point, and calculate acquisition facial expression data;
Second step, treatment catches facial expression data, and generation facial expression catches file;
3rd step, builds three-dimensional personage's geometrical model, imports facial expression and catches file, driving model.
The data-driven three-dimensional personage's geometrical model captured using TOF depth cameras makes consistent with performing artist Expression, it is thus necessary to determine that the positional information of face feature point, this is also to realize the key that face facial expression catches.How to determine and The validity for extracting face facial expression of the characteristic point information to capturing serves conclusive effect.Face feature information is very Easily influenceed by illumination and attitude, TOF depth cameras can simultaneously provide depth image, RGB color image, thus greatly Illumination and the influence of attitude are reduced greatly, face features point can be more accurately extracted.The present invention is directed to different heads Portion's attitude sets up corresponding AAM models respectively, demarcates the face spy of real person using 3D AAM algorithms in the first step Levy a little, we also can accurately be positioned to face feature point in the case of making head pose different, so as to true people Thing carries out facial expression seizure.
The first step is further specifically included:
A, using TOF depth cameras obtain face coloured image and depth image;
B, set up head pose judgment models;
C, train and set up AAM algorithm models;
D, using AAM algorithms position face feature point;
E, the three-dimensional coordinate for obtaining face feature point.
A steps are specially connection TOF depth cameras, switch on power, and the data control process unit of TOF depth cameras sends Open command, its body-sensing camera is opened, and color data stream and depth data are streamed into data control and processing unit, data Control changes color data circulation into coloured image with processing unit, changes depth data circulation into depth image..
B steps are specially and obtain the corresponding depth image of different head poses using TOF depth cameras, then according to The depth image that the position of head pose will get is divided into multiple different classification based training collection, uses the method for face's myopia plane Vector represents different head poses, so as to set up the random regression forest model of head pose judgement.
What the method that the b steps set up head pose judgment models was set up is that a kind of random regression based on statistical method is gloomy Woods model, it is further specially firstly the need of the head zone and non-head region for identifying depth image, then at this The image-region of fixed size is chosen respectively as training set in two regions.Whether depth image block includes picture number, belongs to Head zone, head center position and head pose information.Each tree uses one group in training set respectively in random regression forest The depth image block for randomly selecting is set up.The foundation of each node classification is two rectangular blocks of selection from image block, meter on tree The average of the depth value of rectangular block is calculated, the threshold value that this average sets with certain is compared into entrance right subtrees of the more than this threshold value, it is no Then enter left subtree, the decision tree for carrying out selecting to ultimately generate symmetery to threshold value and rectangular block using Shannon entropy, such as formula Shown in 1.Another group of depth image block being taken at random and setting up another decision tree, all decision trees constitute random regression forest.
In formula, H (P) is Shannon entropy,For left and right subtree Shannon entropy and.
Wherein, the used AAM algorithms of the c steps and traditional AAM algorithms are otherwise varied.Traditional AAM algorithms enter The object of row training and identification is all almost close in the case of the visual angle of front, the present invention is on the basis of traditional AAM algorithms On be improved, for three-dimensional head attitude different when portion's characteristic point is positioned over there, set up respectively corresponding AAM visual angles model, each AAM visual angles model is divided into shape and texture model two parts again.
The c steps are further specially according to the difference of face's direction, and the direction of wherein face is approximately put down using face The normal direction in face represented, training set is divided into the sub- training set under multiple different visual angles.One individually is set up to every sub- training set AAM models, such a single AAM models are called an AAM visual angles model.For each AAM visual angles model, point shape and Texture two parts are modeled.
It is that training set under visual angle carries out manual characteristic point demarcation, characteristic point to set up shape and be specially first to each Location sets be referred to as shape vector Si, obtain training set L={ (Ii, Si) | i=1,2 ..., m;Wherein, IiRepresent training sample, SiRepresent the coordinate set of manual markings.i It is the number of point;X, y are the coordinate of point, (to image removal translation, rotation, scaling) are then normalized to shape and to returning One shape changed is alignd, and average shape is drawn by PCA transformation calculationsTraining image is deformed to average shape, because This any face shape S can use linear equationExpression, PsIt is shape principal component that PCA transformation calculations are obtained The transformation matrix that characteristic component is formed, BsIt is the Statistical Shape parameter for controlling shape conversion, obtains shape Statistics model.
Set up being specially for texture model and obtain corresponding AAM shape eigenvectors SiAnd average shapeSelection byThrough Shaped grid that Delaunay Triangulation obtains is crossed as the benchmark grid for carrying out texture mapping, by the image root in training set Delaunay Triangulation is carried out according to correspondence nominal shape characteristic point position, the pixel value in shaped grid envelope after subdivision Exactly need sampling texture information, secondly, by the affine method of piecewise linearity map that to setting benchmark grid in, Realize the normalization sampling to texture;Again, PCA dimensionality reductions are carried out to the texture information after normalization and obtains average textureThis The arbitrary face texture of sample can be usedRepresent, the texture statistics model for finally drawing is:
It is average texture, PgIt is the transformation matrix that the texture main component characteristic component that PCA transformation calculations are obtained is formed, BgIt is the statistic texture parameter for controlling texture transformation.
Model Fusion
Active apparent model is:Wherein,It is average apparent vector, Q is The matrix that apparent principal component characteristic vector is formed, c is the apparent parameter of statistics for controlling apparent change.
The d steps are further specially and are input into new image, and the head pose set up using step b judges random regression Forest model, the new depth image to being given judges head position therein and face's direction, according to the head pose for obtaining Immediate AAM visual angles model therewith is selected in the AAM visual angles model that information is set up from c steps.Using the AAM models chosen to working as Preceding input picture carries out the positioning of face feature point.Wherein, the process for carrying out facial characteristics point location is exactly that AAM is matched The process of iterative calculation, comprises the following steps that:
1. initialization model parameter c;
2. the poor Δ g=g of computation model texture and current textures-gm
3. using the changes delta c=R of equation of linear regression Prediction Parameters ccΔg;
4. new model parameter c '=c-k Δs c, k=1 is attempted;
5. error function Δ g ' is recalculated;
If 6., | | Δ g ' | | < | | Δs g | |, receives c ' as new parameter;4 are otherwise jumped to, k=1.5 is tasted, 2, 2.5 etc.
7. a two field picture is removed, turns 2.
After the e steps are further specially the position for determining face feature point, face features point is calculated and obtained Coordinate information, the coordinate system of TOF depth cameras is that, with its own as the origin of coordinates, front is Z-direction, is to the left X-axis Positive direction, upwards for Y-axis positive direction.The horizontal view angle angle value α and vertical angle of view angle value β of TOF depth cameras are obtained, By conversion relation RealWorldXtoZ=2tan (α/2), RealWorldYtoZ=2tan (β/2) has obtained depth map The width value w/ height values h of picture and the ratio apart from d.
NormalizedX=x/512-0.5
NormalizedY=0.5-y/424
X=NormalizedX*Z*RealWorldXtoZ
Y=NormalizedY*Z*RealWorldYtoZ
To sum up, index can be tried to achieve for x, the three-dimensional coordinate of the characteristic point at y is (X, Y, Z).
The second step further drives dummy model to realize face face table specifically by motion capture file The seizure of feelings.The three-dimensional coordinate information that the first step is obtained, rotation information is calculated by transfer algorithm, and by its according to The form write activity of BVH is caught in file.
The transfer algorithm is specially:
A, structural information and initial state information to input feature vector point in BVH files;
B, the exercise data for reading face feature point, data form are the three-dimensional coordinate of each point;
C, three-dimensional coordinate is converted into quaternary number;
D, calculating spin matrix;
E, node coordinate is transformed under local coordinate system;
F, Eulerian angles are tried to achieve by formula 3;
θ x=arcsin2 (yz-wx)
Y, z, w are the coordinates of spatial point;
G, Eulerian angles are added in BVH files.
Specific implementation process of the present invention is as follows:
Step 1. obtains face coloured image and depth image
Color data stream and depth data stream are obtained using TOF depth cameras, and changes color data circulation into cromogram Picture, changes depth data circulation into depth image and changes color data circulation into coloured image, changes depth data circulation into depth Degree image, as depicted in figs. 1 and 2.
Step 2. sets up head pose judgment models
The data-driven threedimensional model captured using TOF depth cameras makes the expression consistent with performing artist, it is necessary to really Determine the positional information of face feature point, this is also to realize the key that face facial expression catches.The present invention is directed to different heads Attitude sets up corresponding AAM models respectively, and we can also enter to face feature point in the case of making head pose different The accurate positioning of row.
The present invention is a kind of random regression forest mould based on statistical method using head pose judgment models method is set up Type.Then the method is selected respectively firstly the need of the head zone and non-head region that identify depth image in the two regions The image-region of fixed size is taken as training set.Whether depth image block includes picture number, belong to head zone, head Heart position and head pose information.Each tree uses one group of depth map for randomly selecting in training set respectively in random regression forest As block is set up.The foundation of each node classification is two rectangular blocks of selection from image block on tree, calculates the depth value of rectangular block Average, by the threshold value that this average sets with certain compare more than this threshold value entrance right subtree, otherwise into left subtree, profit The decision tree for carrying out selecting to ultimately generate symmetery to threshold value and rectangular block with Shannon entropy, takes another group of depth image at random Block sets up another decision tree, and all decision trees constitute random regression forest.
In formula, H (P) is Shannon entropy,For left and right subtree Shannon entropy and.
Step 3. is trained and sets up the model of AAM algorithms
Model is divided into two parts of shape and texture model by AAM algorithms, and AAM algorithms are on the basis of ASM algorithms Upper (will facial image be deformed to average shape) carries out texture analysis to image to position its characteristic point.
Using AAM algorithm locating human face face feature points firstly the need of setting up shape and texture model.In view of head appearance Influence of the state to the facial positioning feature point degree of accuracy, we set up an AAM model for each head pose, choose every time The AAM models being best suitable for carry out the matching of face feature point.
Set up shape
Manual characteristic point demarcation, the location sets quilt of characteristic point are carried out to the training sample under each head pose first Referred to as shape vector Si, obtain training set
L={ (Ii, Si) | i=1,2 ..., m;Then shape is carried out Normalization (to image removal translation, rotation, scaling) is simultaneously alignd to normalized shape, is drawn by PCA transformation calculations Average shapeTraining image is deformed to average shape, therefore any face shape S can use linear equation,Expression, PsIt is the transformation matrix of the shape principal component characteristic component formation that PCA transformation calculations are obtained, BsIt is control The Statistical Shape parameter of shape conversion, obtains shape Statistics model.
Set up texture model
The process for setting up texture model is consistent with the process for setting up shape, and the texture statistics model for finally drawing is:
It is average texture, PgIt is the transformation matrix that the texture main component characteristic component that PCA transformation calculations are obtained is formed, bgIt is the statistic texture parameter for controlling texture transformation.
Model Fusion
Active apparent model is:Wherein,It is average apparent vector, Q is The matrix that apparent principal component characteristic vector is formed, c is the apparent parameter of statistics for controlling apparent change.
Step 4. is using AAM algorithms positioning face feature point
The new image of input, the random regression forest model set up using step 2, the new depth image to being given is judged Head position therein, and one group of estimate of head pose is given, calculate the head center of the depth image for providing Position and head pose information.Then the 3DAAM models for most matching therewith are selected.Head pose estimation information acquisition head center Position and head pose, thus calculate the rotation and translation of head model, using end rotation angle calculation spin matrix R, Using head center position as translation matrix T.The 3D face feature points being calculated in upper section are revolved using R chis and T Turn and translate, camera internal reference is reused afterwards by the spot projection after conversion to RGB image plane, obtain the feature on RGB image Point set, in this, as the shape initial value of AAM model instances.
1. initialization model parameter c;
2. the poor Δ g=g of computation model texture and current textures-gm
3. using the changes delta c=R of equation of linear regression Prediction Parameters ccΔg;
4. new model parameter c '=c-k Δs c, k=1 is attempted;
5. error function Δ g ' is recalculated;
6. c ' is received as new parameter if | | Δ g ' | | < | | Δs g | |;4 are otherwise jumped to, k=1.5 is tasted, 2, 2.5 etc.
7. a two field picture is removed, turns 2.
Step 5. obtains the three-dimensional coordinate of face feature point
After determining the position of face feature point, followed by calculating and obtain the coordinate information of face features point, The coordinate system of TOF depth cameras is that, with its own as the origin of coordinates, front is Z-direction, is to the left the positive direction of X-axis, to Upper is the positive direction of Y-axis.TOF depth cameras can get depth image, and in 3D computer graphics, depth image refers to from sight Examine visual angle to look, image includes a kind of information image related to object scene surface distance or an image channel.So, Assuming that the visual field direction Z side of the change direction (i.e. video camera shooting direction) of image depth values and the three-dimensional scenic of required description If identical, then just can easily describe whole three-dimensional scenic.Therefore, depth image is also referred to as range image.
The depth image pixel value that TOF depth cameras get is and TOF depth cameras camera lens is to water actual between object What flat distance was associated, can be in the hope of the Z values of real space by this incidence relation, the X and Y of real space can be by spies Levy index value x (row) and y (OK) a little in depth image to try to achieve, computational methods are as follows:
The horizontal view angle angle value α and vertical angle of view angle value β of TOF depth cameras are obtained, by conversion relation RealWorldXtoZ=2tan (α/2), RealWorldYtoZ=2tan (β/2), that is, obtained the width value w/ of depth image Height value h and the ratio apart from d.
NormalizedX=x/512-0.5
NormalizedY=0.5-y/424
X=NormalizedX*Z*RealWorldXtoZ
Y=NormalizedY*Z*RealWorldYtoZ
To sum up, index can be tried to achieve for x, the three-dimensional coordinate of the characteristic point at y is (X, Y, Z).
Step 6. generation facial expression catches file
The present invention drives dummy model by motion capture file to realize the seizure of face facial expression.Treatment TOF Each two field picture of depth camera acquisition simultaneously therefrom determines to extract the three-dimensional coordinate information of characteristic point, is calculated by transfer algorithm Go out rotation information, and by its form write activity seizure file according to bvh.
The step of realizing of transfer algorithm is:
To the structural information and initial state information of input feature vector point in BVH files;
Exercise data is read to be stored in structure;
Three-dimensional coordinate is converted into quaternary number;
Calculate spin matrix;
Node coordinate is transformed under local coordinate system;
Eulerian angles are input in BVH files.
All above-mentioned this intellectual properties of primarily implementation, the not this new product of implementation of setting limitation other forms And/or new method.Those skilled in the art will be using this important information, the above modification, to realize similar execution feelings Condition.But, all modifications or transformation are based on the right that new product of the present invention belongs to reservation.
The above, is only presently preferred embodiments of the present invention, is not the limitation for making other forms to the present invention, is appointed What those skilled in the art changed possibly also with the technology contents of the disclosure above or be modified as equivalent variations etc. Effect embodiment.But it is every without departing from technical solution of the present invention content, according to technical spirit of the invention to above example institute Any simple modification, equivalent variations and the remodeling made, still fall within the protection domain of technical solution of the present invention.

Claims (8)

1. a kind of facial expression method for catching based on TOF depth cameras, it is characterised in that including:
The first step, facial expression seizure is carried out using TOF depth cameras to real person;
Second step, processes the facial expression data for catching, and generate facial expression seizure file;
3rd step, builds three-dimensional personage's geometrical model, imports facial expression and catches file, driving model.
2. the facial expression method for catching of TOF depth cameras is based on as claimed in claim 1, it is characterised in that:In the first step Middle use AAM algorithms demarcate the face feature point of real person, and we also can be to face in the case of making head pose different Characteristic point is accurately positioned, so as to carry out facial expression seizure to real person.
3. the facial expression method for catching of TOF depth cameras is based on as claimed in claim 1 or 2, it is characterised in that:Described One step is further specifically included,
A, using TOF depth cameras obtain face coloured image and depth image;
B, set up head pose judgment models;
C, the model trained and set up AAM algorithms;
D, using AAM algorithms position face feature point;
E, the three-dimensional coordinate for obtaining face feature point.
4. the facial expression method for catching based on TOF depth cameras as described in claims 1 to 3, it is characterised in that:A is walked The color data stream and depth data stream of three-dimensional face are specially obtained using TOF depth cameras, and color data circulation is changed Into coloured image, depth data circulation is changed into depth image.
5. the facial expression method for catching based on TOF depth cameras as described in Claims 1-4, it is characterised in that:B is walked The corresponding depth image of different head poses specially is obtained using TOF depth cameras, then according to the position of head pose Put the depth image that will be got and be divided into multiple different classification based training collection, represented not using the normal vector of face's myopia plane Same head pose, so as to set up head pose judgment models.
6. the facial expression method for catching based on TOF depth cameras as described in claim 1 to 5, it is characterised in that:Described C walks used AAM algorithms and model is divided into two parts of shape and texture model, and AAM algorithms are the bases in ASM algorithms (will facial image be deformed to average shape) carries out texture analysis to image to position its characteristic point on plinth.
7. the facial expression method for catching based on TOF depth cameras as described in claim 1 to 6, it is characterised in that:Described D steps are further specially and are input into new image, the random regression forest model set up using step b, to the new depth for being given Image judges head position therein, and provides one group of estimate of head pose, calculates the head of the depth image for providing The position at portion center and head pose information, then select the AAM models for most matching therewith, head pose estimation information acquisition head Portion center and head pose, thus calculate the rotation and translation of head model, are rotated using end rotation angle calculation Matrix R, using head center position as translation matrix, rotation peace is carried out to the 3D face feature points for obtaining using R and T Move, recycle camera internal reference by the spot projection after conversion to RGB image plane afterwards, obtain the feature point set on RGB image, with This as AAM model instances shape initial value.
8. the facial expression method for catching based on TOF depth cameras as described in claim 1 to 7, it is characterised in that:Described After e steps are further specially the position for determining face feature point, the coordinate information of face features point, TOF are calculated and obtained The coordinate system of depth camera is that, with its own as the origin of coordinates, front is Z-direction, is to the left the positive direction of X-axis, is upwards The positive direction of Y-axis, TOF depth cameras can get depth image, finally obtain whole three-dimensional face.
CN201611190469.1A 2016-12-21 2016-12-21 A kind of facial expression method for catching based on TOF depth cameras Pending CN106778628A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201611190469.1A CN106778628A (en) 2016-12-21 2016-12-21 A kind of facial expression method for catching based on TOF depth cameras

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201611190469.1A CN106778628A (en) 2016-12-21 2016-12-21 A kind of facial expression method for catching based on TOF depth cameras

Publications (1)

Publication Number Publication Date
CN106778628A true CN106778628A (en) 2017-05-31

Family

ID=58893609

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201611190469.1A Pending CN106778628A (en) 2016-12-21 2016-12-21 A kind of facial expression method for catching based on TOF depth cameras

Country Status (1)

Country Link
CN (1) CN106778628A (en)

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107368778A (en) * 2017-06-02 2017-11-21 深圳奥比中光科技有限公司 Method for catching, device and the storage device of human face expression
CN107403170A (en) * 2017-08-08 2017-11-28 灵然创智(天津)动画科技发展有限公司 A kind of seizure point positioning system based on facial seizure system
CN107657652A (en) * 2017-09-11 2018-02-02 广东欧珀移动通信有限公司 Image processing method and device
CN107680034A (en) * 2017-09-11 2018-02-09 广东欧珀移动通信有限公司 Image processing method and device, electronic installation and computer-readable recording medium
CN107705276A (en) * 2017-09-11 2018-02-16 广东欧珀移动通信有限公司 Image processing method and device, electronic installation and computer-readable recording medium
CN107705356A (en) * 2017-09-11 2018-02-16 广东欧珀移动通信有限公司 Image processing method and device
CN107886558A (en) * 2017-11-13 2018-04-06 电子科技大学 A kind of human face expression cartoon driving method based on RealSense
CN108573526A (en) * 2018-03-30 2018-09-25 盎锐(上海)信息科技有限公司 Face snap device and image generating method
CN108648280A (en) * 2018-04-25 2018-10-12 深圳市商汤科技有限公司 virtual role driving method and device, electronic equipment and storage medium
CN109299639A (en) * 2017-07-25 2019-02-01 虹软(杭州)多媒体信息技术有限公司 A kind of method and apparatus for Expression Recognition
CN109325393A (en) * 2017-08-01 2019-02-12 苹果公司 Using the face detection of single network, Attitude estimation and away from the estimation of camera distance
WO2019154231A1 (en) * 2018-02-12 2019-08-15 腾讯科技(深圳)有限公司 Image processing method, electronic device and storage medium
CN110378855A (en) * 2019-07-18 2019-10-25 博瓦鲸视(深圳)科技有限公司 A method of personage's statistical model that building is acted comprising personage's figure, facial expression, gesture, posture
CN110427917A (en) * 2019-08-14 2019-11-08 北京百度网讯科技有限公司 Method and apparatus for detecting key point
CN110738717A (en) * 2019-10-16 2020-01-31 网易(杭州)网络有限公司 Method and device for correcting motion data and electronic equipment
CN111290568A (en) * 2018-12-06 2020-06-16 阿里巴巴集团控股有限公司 Interaction method and device and computer equipment
CN111416934A (en) * 2020-03-16 2020-07-14 南昌航空大学 Facial expression acquisition method and system based on functional music intervention
CN111741285A (en) * 2020-06-08 2020-10-02 上海龙旗科技股份有限公司 Real-time 3D scene implementation method and device
CN112241658A (en) * 2019-07-17 2021-01-19 青岛大学 Fatigue driving early warning system and method based on depth camera
CN113221600A (en) * 2020-01-21 2021-08-06 初速度(苏州)科技有限公司 Method and device for calibrating image feature points
CN114463817A (en) * 2022-02-10 2022-05-10 北京中科深智科技有限公司 Lightweight 2D video-based facial expression driving method and system
CN116453384A (en) * 2023-06-19 2023-07-18 江西德瑞光电技术有限责任公司 Immersion type intelligent learning system based on TOF technology and control method

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102479388A (en) * 2010-11-22 2012-05-30 北京盛开互动科技有限公司 Expression interaction method based on face tracking and analysis
US20130100140A1 (en) * 2011-10-25 2013-04-25 Cywee Group Limited Human body and facial animation systems with 3d camera and method thereof
CN103093490A (en) * 2013-02-02 2013-05-08 浙江大学 Real-time facial animation method based on single video camera
CN103324905A (en) * 2012-03-21 2013-09-25 天津生态城动漫园投资开发有限公司 Next-generation virtual photostudio facial capture system
CN103608841A (en) * 2011-05-17 2014-02-26 高通股份有限公司 Head pose estimation using RGBD camera
CN103942822A (en) * 2014-04-11 2014-07-23 浙江大学 Facial feature point tracking and facial animation method based on single video vidicon
CN104345801A (en) * 2013-08-09 2015-02-11 三星电子株式会社 Hybrid visual communication
CN105654537A (en) * 2015-12-30 2016-06-08 中国科学院自动化研究所 Expression cloning method and device capable of realizing real-time interaction with virtual character
CN105900144A (en) * 2013-06-07 2016-08-24 费斯史福特股份公司 Online modeling for real-time facial animation
CN106228119A (en) * 2016-07-13 2016-12-14 天远三维(天津)科技有限公司 A kind of expression catches and Automatic Generation of Computer Animation system and method

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102479388A (en) * 2010-11-22 2012-05-30 北京盛开互动科技有限公司 Expression interaction method based on face tracking and analysis
CN103608841A (en) * 2011-05-17 2014-02-26 高通股份有限公司 Head pose estimation using RGBD camera
US20130100140A1 (en) * 2011-10-25 2013-04-25 Cywee Group Limited Human body and facial animation systems with 3d camera and method thereof
CN103324905A (en) * 2012-03-21 2013-09-25 天津生态城动漫园投资开发有限公司 Next-generation virtual photostudio facial capture system
CN103093490A (en) * 2013-02-02 2013-05-08 浙江大学 Real-time facial animation method based on single video camera
CN105900144A (en) * 2013-06-07 2016-08-24 费斯史福特股份公司 Online modeling for real-time facial animation
CN104345801A (en) * 2013-08-09 2015-02-11 三星电子株式会社 Hybrid visual communication
CN103942822A (en) * 2014-04-11 2014-07-23 浙江大学 Facial feature point tracking and facial animation method based on single video vidicon
CN105654537A (en) * 2015-12-30 2016-06-08 中国科学院自动化研究所 Expression cloning method and device capable of realizing real-time interaction with virtual character
CN106228119A (en) * 2016-07-13 2016-12-14 天远三维(天津)科技有限公司 A kind of expression catches and Automatic Generation of Computer Animation system and method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
CHENLEE_1: "AAM(Active Appearance Model)算法介绍", 《CSDN》 *

Cited By (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107368778A (en) * 2017-06-02 2017-11-21 深圳奥比中光科技有限公司 Method for catching, device and the storage device of human face expression
CN109299639B (en) * 2017-07-25 2021-03-16 虹软科技股份有限公司 Method and device for facial expression recognition
CN109299639A (en) * 2017-07-25 2019-02-01 虹软(杭州)多媒体信息技术有限公司 A kind of method and apparatus for Expression Recognition
CN109325393A (en) * 2017-08-01 2019-02-12 苹果公司 Using the face detection of single network, Attitude estimation and away from the estimation of camera distance
CN109325393B (en) * 2017-08-01 2022-12-09 苹果公司 Face detection, pose estimation, and estimation of distance to camera using a single network
CN107403170A (en) * 2017-08-08 2017-11-28 灵然创智(天津)动画科技发展有限公司 A kind of seizure point positioning system based on facial seizure system
CN107705276B (en) * 2017-09-11 2021-03-02 Oppo广东移动通信有限公司 Image processing method and apparatus, electronic apparatus, and computer-readable storage medium
CN107657652A (en) * 2017-09-11 2018-02-02 广东欧珀移动通信有限公司 Image processing method and device
CN107680034A (en) * 2017-09-11 2018-02-09 广东欧珀移动通信有限公司 Image processing method and device, electronic installation and computer-readable recording medium
CN107705276A (en) * 2017-09-11 2018-02-16 广东欧珀移动通信有限公司 Image processing method and device, electronic installation and computer-readable recording medium
CN107705356A (en) * 2017-09-11 2018-02-16 广东欧珀移动通信有限公司 Image processing method and device
CN107886558A (en) * 2017-11-13 2018-04-06 电子科技大学 A kind of human face expression cartoon driving method based on RealSense
US11436779B2 (en) 2018-02-12 2022-09-06 Tencent Technology (Shenzhen) Company Ltd Image processing method, electronic device, and storage medium
WO2019154231A1 (en) * 2018-02-12 2019-08-15 腾讯科技(深圳)有限公司 Image processing method, electronic device and storage medium
CN108573526A (en) * 2018-03-30 2018-09-25 盎锐(上海)信息科技有限公司 Face snap device and image generating method
CN108648280B (en) * 2018-04-25 2023-03-31 深圳市商汤科技有限公司 Virtual character driving method and device, electronic device and storage medium
CN108648280A (en) * 2018-04-25 2018-10-12 深圳市商汤科技有限公司 virtual role driving method and device, electronic equipment and storage medium
CN111290568A (en) * 2018-12-06 2020-06-16 阿里巴巴集团控股有限公司 Interaction method and device and computer equipment
CN112241658A (en) * 2019-07-17 2021-01-19 青岛大学 Fatigue driving early warning system and method based on depth camera
CN112241658B (en) * 2019-07-17 2023-09-01 青岛大学 Fatigue driving early warning method based on depth camera
CN110378855B (en) * 2019-07-18 2023-04-25 博瓦鲸视(深圳)科技有限公司 Method for constructing character statistical model containing character shape, facial expression, gesture and gesture actions
CN110378855A (en) * 2019-07-18 2019-10-25 博瓦鲸视(深圳)科技有限公司 A method of personage's statistical model that building is acted comprising personage's figure, facial expression, gesture, posture
CN110427917A (en) * 2019-08-14 2019-11-08 北京百度网讯科技有限公司 Method and apparatus for detecting key point
CN110427917B (en) * 2019-08-14 2022-03-22 北京百度网讯科技有限公司 Method and device for detecting key points
CN110738717A (en) * 2019-10-16 2020-01-31 网易(杭州)网络有限公司 Method and device for correcting motion data and electronic equipment
CN110738717B (en) * 2019-10-16 2021-05-11 网易(杭州)网络有限公司 Method and device for correcting motion data and electronic equipment
CN113221600B (en) * 2020-01-21 2022-06-21 魔门塔(苏州)科技有限公司 Method and device for calibrating image feature points
CN113221600A (en) * 2020-01-21 2021-08-06 初速度(苏州)科技有限公司 Method and device for calibrating image feature points
CN111416934A (en) * 2020-03-16 2020-07-14 南昌航空大学 Facial expression acquisition method and system based on functional music intervention
CN111741285A (en) * 2020-06-08 2020-10-02 上海龙旗科技股份有限公司 Real-time 3D scene implementation method and device
CN114463817A (en) * 2022-02-10 2022-05-10 北京中科深智科技有限公司 Lightweight 2D video-based facial expression driving method and system
CN116453384A (en) * 2023-06-19 2023-07-18 江西德瑞光电技术有限责任公司 Immersion type intelligent learning system based on TOF technology and control method

Similar Documents

Publication Publication Date Title
CN106778628A (en) A kind of facial expression method for catching based on TOF depth cameras
CN109325437B (en) Image processing method, device and system
CN109472198B (en) Gesture robust video smiling face recognition method
Liao et al. Automatic caricature generation by analyzing facial features
WO2021093453A1 (en) Method for generating 3d expression base, voice interactive method, apparatus and medium
Min et al. Kinectfacedb: A kinect database for face recognition
Li et al. Realtime facial animation with on-the-fly correctives.
US8624901B2 (en) Apparatus and method for generating facial animation
CN103443826B (en) mesh animation
CN103473801A (en) Facial expression editing method based on single camera and motion capturing data
CN102157007A (en) Performance-driven method and device for producing face animation
CN113449570A (en) Image processing method and device
CN109034131A (en) A kind of semi-automatic face key point mask method and storage medium
Wang et al. Digital twin: Acquiring high-fidelity 3D avatar from a single image
Darujati et al. Facial motion capture with 3D active appearance models
CN110223368A (en) A kind of unmarked motion capture method of face based on deep learning
Luo et al. Facial metamorphosis using geometrical methods for biometric applications
Luo et al. Alignment and tracking of facial features with component-based active appearance models and optical flow
Pei et al. Transferring of speech movements from video to 3D face space
Somepalli et al. Implementation of single camera markerless facial motion capture using blendshapes
Agianpuye et al. Synthesizing neutral facial expression on 3D faces using Active Shape Models
Tian et al. Robust facial marker tracking based on a synthetic analysis of optical flows and the YOLO network
Abeysundera et al. Nearest neighbor weighted average customization for modeling faces
Li et al. Example-based 3D face reconstruction from uncalibrated frontal and profile images
Liu et al. Face cartoon synthesis based on the active appearance model

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20170531