CN109508654A - Merge the human face analysis method and system of multitask and multiple dimensioned convolutional neural networks - Google Patents
Merge the human face analysis method and system of multitask and multiple dimensioned convolutional neural networks Download PDFInfo
- Publication number
- CN109508654A CN109508654A CN201811260674.XA CN201811260674A CN109508654A CN 109508654 A CN109508654 A CN 109508654A CN 201811260674 A CN201811260674 A CN 201811260674A CN 109508654 A CN109508654 A CN 109508654A
- Authority
- CN
- China
- Prior art keywords
- face
- multitask
- loss
- convolutional neural
- neural networks
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/174—Facial expression recognition
- G06V40/175—Static expression
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- General Health & Medical Sciences (AREA)
- Data Mining & Analysis (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a kind of human face analysis method and system for merging multitask combining multi-scale convolutional neural networks, the picture to be learned for being first N × N to a size, the face area-of-interest for extracting K different scale from the picture using key area searching algorithm, as the input in multiple dimensioned tri- channels CNN;Then feature extraction is carried out to the K face area-of-interests respectively using CNN, obtains the face characteristic of different scale, the face characteristic of the different scale extracted is merged using cascade mode, obtains fused feature representation;The loss function of multiple tasks is merged again to obtain associated losses function, is inputted using the feature representation as study, the optimal solution about the associated losses function is obtained, to obtain the processing result of the multiple task.The present invention is promoted mutually using the correlation between task to improve the predictablity rate of individual task.
Description
Technical field
The present invention relates to face character fields, more specifically to a kind of fusion multitask and multiple dimensioned convolutional Neural
The human face analysis method and system of network.
Background technique
In recent years, with the development of internet and artificial intelligence the relevant technologies achieve in practical applications centainly at
Fruit, artificial intelligence field have attracted the concern of more and more researchers, and the application range of artificial intelligence technology is also more next
It is wider.In computer vision field, Face datection and analysis are always the popular direction studied.Although being based on depth convolution recently
The research of neural network (Convolutional Neural network, CNN) achieves significant achievement, and is widely answered
For fields such as recognition of face, face tracking, Face datections, but for Face datection task, still it is difficult to from comprising extreme appearance
Facial key point, head pose, gender and expression information are obtained in the facial image of state, illumination and change resolution.Face inspection
It surveys, the task of crucial point location, Attitude estimation, Gender Classification and Expression Recognition is normally used as individual Resolving probiems.Recently,
Research shows that learning multiple relevant tasks simultaneously can be improved the performance of individual task.
Summary of the invention
The technical problem to be solved in the present invention is that carrying out multitask (Multi-Task to the relevant attribute of face
Learning, MTL) study, the correlation between attribute, and the multitask convolutional Neural for passing through multiple channels are excavated deeper into ground
The case where same picture of e-learning a variety of resolution ratio, enriches the extraction of face character feature with this, thus by part attribute
Accuracy of identification promoted.
The technical solution adopted by the present invention to solve the technical problems is: constructing a kind of fusion multitask combining multi-scale volume
The human face analysis method of product neural network, comprises the following steps:
(1) multiple dimensioned face attention region extraction step:
The picture to be learned for being N × N to a size, is mentioned from the picture using key area searching algorithm
The face area-of-interest for taking out K different scale, as the input in multiple dimensioned tri- channels CNN;Wherein, N indicates picture
Plain size, K >=2 and be integer;
(2) multiple dimensioned learning procedure, including feature extraction sub-step and Fusion Features sub-step are merged;
Feature extraction sub-step: feature extraction is carried out to the K face area-of-interests respectively using CNN, is obtained not
With the face characteristic of scale;
Fusion Features sub-step: using the face spy for the different scale that cascade mode extracts feature extraction sub-step
Sign is merged, and fused feature representation is obtained;
(3) multitask human face analysis step:
The loss function of multiple tasks is merged to obtain associated losses function, the feature obtained with step (3)
Expression obtains the optimal solution about the associated losses function, to obtain the processing of the multiple task as study input
As a result.
Further, in the human face analysis method of fusion multitask combining multi-scale convolutional neural networks of the invention,
In step (1), when extracting face area-of-interest, the face sense for positioning 3 scales is helped using " attention " mechanism
Interest region.
Further, in the human face analysis method of fusion multitask combining multi-scale convolutional neural networks of the invention,
Specifically include in the feature extraction sub-step of step (2): the feature channel in K channel of building, each feature channel are right respectively
The face area-of-interest of one scale independently carries out feature extraction, extracts the face characteristic of K different scale altogether.
Further, in the human face analysis method of fusion multitask combining multi-scale convolutional neural networks of the invention,
Feature representation in step (3), between the multiple task parallel training and shared inter-related task.
Further, in the human face analysis method of fusion multitask combining multi-scale convolutional neural networks of the invention,
It is characterized in that, K=3.
Further, in the human face analysis method of fusion multitask combining multi-scale convolutional neural networks of the invention,
In step (3), it is described merged the loss function of multiple tasks to obtain associated losses function specifically refer to: will be the multiple
The loss function of task is weighted.
Further, in the human face analysis method of fusion multitask combining multi-scale convolutional neural networks of the invention,
The multiple task refers to: Face datection, Attitude estimation, crucial point location, gender identification and five tasks of Expression Recognition.
Further, in the human face analysis method of fusion multitask combining multi-scale convolutional neural networks of the invention,
(a) Face datection task returns to face frame coordinate, loss function for detecting and positioning the face in picture are as follows:
lossD=-(1-l) log (l-pD)-l·log(pD),
It is face wherein when l=1;L=0 is non-face;pDIt is the probability of face for face key area;
(b) gender of gender identification mission face picture for identification, loss function are as follows:
lossG=-(1-g) log (1-pg0)-g·log(pg1),
Wherein, male g=0, women g=1, (pg0, pg1) face picture be men and women probability;
(c) face key point location tasks are used for the face feature point location of face alignment, are on the basis of Face datection
Upper progress, to the positioning feature point on face;Face key area is described with (x, y, w, h), wherein (x, y) is face
The coordinate of regional center point, (w, h) are the width and height in the region respectively, and each visible key point is relative in region
Heart point (x, y) displacement, and as follows according to (w, h) normalized formula:
Wherein, NLIt is face key point number, (ai,bi) be training when label, (xi,yi) it is true value, face is crucial
The loss function of point location are as follows:
It is the prediction result of i-th of face key point, viFor the visible factor of i-th of face key point, key point
As it can be seen that vi=1, otherwise vi=0;
It can be seen that whether the factor is for characterizing face key point as it can be seen that being additional key point location, loss function are as follows:
(d) head pose estimation task is used for face picture head in roll (p1), pitch (p2) and yaw (p3) three
Prediction on direction, loss function are as follows:
Wherein,It is the estimated result of the angle of three directions head portion posture, (p1,p2,p3) it is institute
State the training label of the angle of three direction head portion postures;
(e) Expression Recognition task from given still image or dynamic video sequence for isolating specific expression shape
State, so that it is determined that the mental emotion of identified object, loss function are as follows:
Wherein, M is the type number of expression, and M is more than or equal to 2,Indicate the classification of expression
Training label, (pe1, pe2..., peM) indicate expression classification estimated result.
Associated losses function are as follows: lossfull=λDlossD+λLlossL+λVlossV+λPlossP+λGlossG+λElossE;
Wherein λ indicates weight in formula, and λ is greater than 0.
Further, in the human face analysis method of fusion multitask combining multi-scale convolutional neural networks of the invention,
(e) in,
Wherein the type of expression shares 6 kinds, respectively represents 6 kinds of basic facial expressions: anger is detested, fears, is happy, is sad, is frightened
It is surprised.
The present invention is to solve its technical problem, is additionally provided a kind of in fusion multitask combining multi-scale convolution of the invention
The human face analysis system of neural network, the system use the fusion multitask combining multi-scale convolutional neural networks of any of the above-described
Human face analysis method carry out human face analysis.
The human face analysis method and system for implementing fusion multitask combining multi-scale convolutional neural networks of the invention, have
Below the utility model has the advantages that the present invention uses end-to-end training method, while multiple tasks to be used for, multiple tasks parallel training is simultaneously shared
Characterization between inter-related task promotes multiple tasks to learn jointly, is promoted mutually using the correlation between task to improve list
The predictablity rate of a task.
Detailed description of the invention
Present invention will be further explained below with reference to the attached drawings and examples, in attached drawing:
Fig. 1 is the human face analysis method task definition figure for merging multitask combining multi-scale convolutional neural networks;
Fig. 2 is result of study exemplary diagram;
Fig. 3 is the human face analysis method network structure for merging multitask combining multi-scale convolutional neural networks;
Fig. 4 is the exemplary diagram that element is searched in iteration region.
Specific embodiment
For a clearer understanding of the technical characteristics, objects and effects of the present invention, now control attached drawing is described in detail
A specific embodiment of the invention.
With reference to Fig. 1-Fig. 3, the human face analysis method of the fusion multitask combining multi-scale convolutional neural networks of the present embodiment
It specifically includes the following steps.
(1) multiple dimensioned face attention region extraction step:
The picture to be detected for being N × N to a size extracts the people of three scales using attention detection network
Face area-of-interest, size are respectively 227 × 227,147 × 147,59 × 59, as tri- channels multiple dimensioned CNN
Input;Wherein, N indicates pixel size
(2) multiple dimensioned learning procedure is merged, the fusion multitask Triple network learning model in three channels is constructed, including
Feature extraction sub-step and Fusion Features sub-step;
Feature extraction sub-step: the face key area for three scales for respectively extracting above-mentioned steps (1) is as defeated
Enter, they respectively correspond three sub-networks of table 1:
1 network detail parameters table of table
The input scale of sub-network one is 227 × 227, has used 8 convolutional layers, 3 pond layers, 1 full articulamentum.By
It is layer distributed in the characteristic information that CNN is extracted, low convolutional layer includes more corner information, is suitble to study face key point
Positioning and head pose assess task, and high convolutional layer is then suitble to study Face datection, Gender Classification and Expression Recognition etc., thus
This patent is in the performance for promoting crucial point location and head state to assess by multi-scale feature fusion.The present invention has merged pond
Change layer pool1 layers of (27 × 27 × 96), conv3 layers of convolutional layer (13 × 13 × 384), layer pool5 layers of pond (6 × 6 × 256).
Further, it since their dimensions are inconsistent, cannot be directly connected to, so being convolution conv1a to pool1 layers and conv3 layers
With convolution conv3a, the vector of 6 × 6 × 192 dimensions then is obtained to result dimensionality reduction, finally obtains 2048 by full articulamentum Fc6
The output vector of dimension.
The input scale 147 × 147 of sub-network two has used 4 convolutional layers, 3 pond layers and 2 full articulamentums, warp
It crosses full articulamentum Fc61 and obtains 512 dimension output vectors.
The input scale of sub-network three is 59 × 59, has used 2 convolutional layers, 2 pond layers and 2 full articulamentums, is passed through
It crosses full articulamentum Fc42 and obtains 512 dimension output vectors.
Fusion Features sub-step: using the face spy for the different scale that cascade mode extracts feature extraction sub-step
Sign is merged, and fused feature representation is obtained.Three sub- network layers are merged to obtain 3072 dimension output vectors.Finally incite somebody to action
6 independent branches are divided into according to task to 3072 dimensional vectors: being Face datection, crucial point location, visibility detection, table respectively
Feelings identification, Attitude estimation, gender identification.Wherein visibility detection is assisted face key point location.They respectively correspond Fc_
Detection, Fc_landmarks, Fc_visibility, Fc_pose, Fc_gender, Fc_expression six connect entirely
Layer is connect, each branch carries out primary full connection respectively again and obtains the recognition result of each task, and see Table 1 for details for network parameter table.
(3) multitask human face analysis step:
Multi-task learning provides multiple supervision messages (label), is promoted mutually using the correlation between task, and more
One feature of business is single tensor input (X), multiple outputs (task _ 1, task _ 2...), in the present invention as face is examined
Survey, Expression Recognition etc., it is therefore desirable to have multiple loss functions for calculating separately the loss of each task.The present invention identifies simultaneously
Five tasks, and for additional key point location task, crucial visibility of a point is proposed, therefore there are six loss functions:
(1) Face datection detects and positions the face in picture, high-precision face frame coordinate is returned to, as the people of prediction
Face frame and the intersection of real human face range are when being greater than 0.5 with the union (IOU) of real human face range than the face frame of upper prediction
Positive example (l=1), less than the negative example of 0.35 conduct (l=0), loss function is as follows:
lossD=-(1-l) log (l-pD)-l·log(pD)
Wherein pDIt is the probability of face for the face key area, is acquired from above-mentioned Fc_detection.
(2) gender identification identifies the gender of face on picture, and IOU is greater than 0.5 conduct positive example, and loss function is such as
Under:
lossG=-(1-g) log (1-pg0)-g·log(pg1)
Wherein male g=0, women then g=1, (pg0, pg1) it is the probability that face picture is men and women;
(3) face key point location, i.e. the face feature point location for face alignment, are on the basis of Face datection
It carries out, characteristic point on the face such as corners of the mouth, canthus etc. is positioned, when IOU is used greater than 0.35 conduct positive example,
All key point coordinates are normalized, and using relative value, mapping function is as follows:
(xi, yi) it is key point coordinate, (x, y) is the coordinate at face center, and w, h are that face is wide and high, for sightless
Key point is set as (0,0).Loss function is calculated using Euclidean distance, as follows:
For the relative coordinate values after the normalization of prediction, N is keypoint quantity, and AFLW data set includes 21 passes
Key point information, i.e. N=21, viIndicate the key point in this region as it can be seen that being otherwise 0 when being 1, that is to say, that not for those
Visible point is not involved in operation.
(4) head pose estimation is the prediction to face picture head on tri- directions pitch, yaw and roll, is used
Euclidean distance loss function training roll (p1), pitch (p2), it is positive example that yaw (p3), IOU, which are greater than 0.5, and loss function is public
Formula is as follows:
WhereinFor the predicted value of the angle of three directions head portion posture, (p1,p2,p3) it is described
The training label of the angle of three direction head portion postures.
(5) Expression Recognition, which refers to, isolates specific emotional state from given still image or dynamic video sequence,
So that it is determined that the mental emotion of identified object, similarly, when IOU is greater than 0.5 conduct positive example, loss function formula is as follows:
Wherein pejFor the predicted value of the classification of expression,For the training label of the classification of expression, j 0,1,2,3,5 divides
Do not represent 6 kinds of basic facial expressions: angry, detest, fear, is happy, is sad, is surprised etc..
Further, associated losses function is the weighted sum of five task loss functions, formula are as follows:
lossfull=λDlossD+λLlossL+λVlossV+λPlossP+λGlossG+λElossE
Wherein λ is that the significance level according to each task in general assignment is determined, λ is greater than 0.The weight of this patent is arranged
Are as follows:
λD=1, λL=5, λE=2, λP=5, λG=2
(4) last handling process
Face datection algorithm of the invention faces two challenges: first, network is different surely to capture lesser face;The
Two, human face region cannot be accurately positioned in the face frame predicted.Because confidence score is greater than 0.5 candidate regions by the present invention
It is considered as comprising face, being considered as less than 0.35 is non-face, so the two challenges will lead to the reduction of Face datection rate, while shadow
Ring the recognition result of other tasks.Therefore, the present invention promotes the accuracy of Face detection by last handling process.
The present invention has abandoned traditional bounding box regression algorithm, but proposes the iteration range searching suitable for this task
With the non-maxima suppression algorithm based on key point.Iteration range searching is generated more by using the face feature point information of prediction
Recall rate is improved in more face candidate area.Non-maxima suppression based on key point then by using it is predicted that facial markers
Point information readjusts the bounding box of detection to promote the accuracy of positioning, reuses non-maxima suppression algorithm and goes out de-redundancy side
Frame.Both methods does not need any training process.
(1) iteration range searching: when the correct Face datection score of key point is low, according to it is predicted that face feature point believe
Breath generates new candidate regions using FaceRectCalculator, and it is carried out propagated forward in a model.This is new
The candidate regions of generation can obtain higher score in this task of Face datection, to improve recall rate, realize process such as
Fig. 4.
(2) based on the non-maxima suppression of key point: with the difference of other non-maxima suppression algorithms.Based on key point
Non-maxima suppression algorithm not using face frame coordinate as parameter, but using four angular coordinates of minimum in key point as ginseng
Number can avoid adjacent face from being merged or situations such as missing inspection with compact face in this way.
The embodiment of the present invention is described with above attached drawing, but the invention is not limited to above-mentioned specific
Embodiment, the above mentioned embodiment is only schematical, rather than restrictive, those skilled in the art
Under the inspiration of the present invention, without breaking away from the scope protected by the purposes and claims of the present invention, it can also make very much
Form, all of these belong to the protection of the present invention.
Claims (10)
1. a kind of human face analysis method for merging multitask combining multi-scale convolutional neural networks, which is characterized in that comprising as follows
Step:
(1) multiple dimensioned face attention region extraction step:
The picture to be learned for being N × N to a size, is extracted from the picture using key area searching algorithm
The face area-of-interest of K different scale, as the input in multiple dimensioned tri- channels CNN;Wherein, N indicates that pixel is big
It is small, K >=2 and be integer;
(2) multiple dimensioned learning procedure, including feature extraction sub-step and Fusion Features sub-step are merged;
Feature extraction sub-step: feature extraction is carried out to the K face area-of-interests respectively using CNN, obtains different rulers
The face characteristic of degree;
Fusion Features sub-step: the face characteristic of the different scale extracted feature extraction sub-step using cascade mode into
Row fusion, obtains fused feature representation;
(3) multitask human face analysis step:
The loss function of multiple tasks is merged to obtain associated losses function, the feature representation obtained with step (3)
It is inputted as study, the optimal solution about the associated losses function is obtained, to obtain the processing result of the multiple task.
2. the human face analysis method of fusion multitask combining multi-scale convolutional neural networks according to claim 1, special
Sign is, in step (1), when extracting face area-of-interest, helps to position 3 scales using " attention " mechanism
Face area-of-interest.
3. the human face analysis method of fusion multitask combining multi-scale convolutional neural networks according to claim 1, special
Sign is, specifically includes in the feature extraction sub-step of step (2): the feature channel in K channel of building, each feature channel
Feature extraction is independently carried out to the face area-of-interest of a scale respectively, extracts the face characteristic of K different scale altogether.
4. the human face analysis method of fusion multitask combining multi-scale convolutional neural networks according to claim 1, special
Sign is, the feature representation in step (3), between the multiple task parallel training and shared inter-related task.
5. the human face analysis method of fusion multitask combining multi-scale convolutional neural networks according to claim 1, special
Sign is, K=3.
6. the human face analysis method of fusion multitask combining multi-scale convolutional neural networks according to claim 1, step
(3) in, it is described merged the loss function of multiple tasks to obtain associated losses function specifically refer to: by the multiple task
Loss function be weighted.
7. the human face analysis method of fusion multitask combining multi-scale convolutional neural networks according to claim 6, special
Sign is that the multiple task refers to: Face datection, Attitude estimation, crucial point location, gender identification and Expression Recognition five appoint
Business.
8. the human face analysis method of fusion multitask combining multi-scale convolutional neural networks according to claim 7, special
Sign is,
(a) Face datection task returns to face frame coordinate, loss function for detecting and positioning the face in picture are as follows:
lossD=-(1-l) log (l-pD)-l·log(pD),
It is face wherein when l=1;L=0 is non-face;pDIt is the probability of face for face key area;
(b) gender of gender identification mission face picture for identification, loss function are as follows:
lossG=-(1-g) log (1-pg0)-g·log(pg1),
Wherein, male g=0, women g=1, (pg0, pg1) face picture be men and women probability;
(c) face key point location tasks be used for face alignment face feature point location, be on the basis of Face datection into
Capable, to the positioning feature point on face;Face key area is described with (x, y, w, h), wherein (x, y) is human face region
The coordinate of central point, (w, h) are the width and height in the region respectively, and each visible key point is relative to regional center point
(x, y) displacement, and as follows according to (w, h) normalized formula:
Wherein, NLIt is face key point number, (ai,bi) be training when label, (xi,yi) it is true value, face key point is fixed
The loss function of position are as follows:
It is the prediction result of i-th of face key point, viFor the visible factor of i-th of face key point, key point can
See, vi=1, otherwise vi=0;
It can be seen that whether the factor is for characterizing face key point as it can be seen that being additional key point location, loss function are as follows:
(d) head pose estimation task is used for face picture head in roll (p1), pitch (p2) and yaw (p3) three directions
On prediction, loss function are as follows:
Wherein,It is the estimated result of the angle of three directions head portion posture, (p1,p2,p3) it is described three
The training label of the angle of a direction head portion posture;
(e) Expression Recognition task is used to isolate specific emotional state from given still image or dynamic video sequence,
So that it is determined that the mental emotion of identified object, loss function are as follows:
Wherein, M is the type number of expression, and M is more than or equal to 2,Indicate the training of the classification of expression
Label, (pe1, pe2..., peM) indicate expression classification estimated result.
Associated losses function are as follows: lossfull=λDlossD+λLlossL+λVlossV+λPlossP+λGlossG+λElossE;Wherein formula
Middle λ indicates weight, and λ is greater than 0.
9. the human face analysis method of fusion multitask combining multi-scale convolutional neural networks according to claim 8, special
Sign is, (e) in,
Wherein the type of expression shares 6 kinds, respectively represents 6 kinds of basic facial expressions: anger is detested, fears, is happy, is sad, is surprised.
10. a kind of human face analysis system for merging multitask combining multi-scale convolutional neural networks, which is characterized in that using as weighed
Benefit requires the human face analysis method of the described in any item fusion multitask combining multi-scale convolutional neural networks of 1-9 to carry out face point
Analysis.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811260674.XA CN109508654B (en) | 2018-10-26 | 2018-10-26 | Face analysis method and system fusing multitask and multi-scale convolutional neural network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811260674.XA CN109508654B (en) | 2018-10-26 | 2018-10-26 | Face analysis method and system fusing multitask and multi-scale convolutional neural network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN109508654A true CN109508654A (en) | 2019-03-22 |
CN109508654B CN109508654B (en) | 2021-01-05 |
Family
ID=65746931
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201811260674.XA Expired - Fee Related CN109508654B (en) | 2018-10-26 | 2018-10-26 | Face analysis method and system fusing multitask and multi-scale convolutional neural network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109508654B (en) |
Cited By (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110082822A (en) * | 2019-04-09 | 2019-08-02 | 中国科学技术大学 | The method for carrying out earthquake detection using convolutional neural networks |
CN110135251A (en) * | 2019-04-09 | 2019-08-16 | 上海电力学院 | A kind of group's image Emotion identification method based on attention mechanism and hybrid network |
CN110135373A (en) * | 2019-05-20 | 2019-08-16 | 北京探境科技有限公司 | A kind of multiple dimensioned face recognition method, system and electronic equipment |
CN110188615A (en) * | 2019-04-30 | 2019-08-30 | 中国科学院计算技术研究所 | A kind of facial expression recognizing method, device, medium and system |
CN110232652A (en) * | 2019-05-27 | 2019-09-13 | 珠海格力电器股份有限公司 | Image processing engine processing method, the image processing method for terminal, terminal |
CN110263774A (en) * | 2019-08-19 | 2019-09-20 | 珠海亿智电子科技有限公司 | A kind of method for detecting human face |
CN110287846A (en) * | 2019-06-19 | 2019-09-27 | 南京云智控产业技术研究院有限公司 | A kind of face critical point detection method based on attention mechanism |
CN110298325A (en) * | 2019-07-02 | 2019-10-01 | 四川长虹电器股份有限公司 | Expression impaired patients assisted care system based on video Expression Recognition |
CN110348416A (en) * | 2019-07-17 | 2019-10-18 | 北方工业大学 | Multi-task face recognition method based on multi-scale feature fusion convolutional neural network |
CN110457572A (en) * | 2019-05-23 | 2019-11-15 | 北京邮电大学 | Commodity information recommendation method and electronic equipment based on figure network |
CN110633624A (en) * | 2019-07-26 | 2019-12-31 | 北京工业大学 | Machine vision human body abnormal behavior identification method based on multi-feature fusion |
CN110705563A (en) * | 2019-09-07 | 2020-01-17 | 创新奇智(重庆)科技有限公司 | Industrial part key point detection method based on deep learning |
CN110866466A (en) * | 2019-10-30 | 2020-03-06 | 平安科技(深圳)有限公司 | Face recognition method, face recognition device, storage medium and server |
CN111062269A (en) * | 2019-11-25 | 2020-04-24 | 珠海格力电器股份有限公司 | User state identification method and device, storage medium and air conditioner |
CN111178183A (en) * | 2019-12-16 | 2020-05-19 | 深圳市华尊科技股份有限公司 | Face detection method and related device |
CN111291670A (en) * | 2020-01-23 | 2020-06-16 | 天津大学 | Small target facial expression recognition method based on attention mechanism and network integration |
CN111325108A (en) * | 2020-01-22 | 2020-06-23 | 中能国际建筑投资集团有限公司 | Multitask network model, using method, device and storage medium |
CN111507248A (en) * | 2020-04-16 | 2020-08-07 | 成都东方天呈智能科技有限公司 | Face forehead area detection and positioning method and system of low-resolution thermodynamic diagram |
CN111814706A (en) * | 2020-07-14 | 2020-10-23 | 电子科技大学 | Face recognition and attribute classification method based on multitask convolutional neural network |
CN111914758A (en) * | 2020-08-04 | 2020-11-10 | 成都奥快科技有限公司 | Face in-vivo detection method and device based on convolutional neural network |
CN112036260A (en) * | 2020-08-10 | 2020-12-04 | 武汉星未来教育科技有限公司 | Expression recognition method and system for multi-scale sub-block aggregation in natural environment |
CN112036339A (en) * | 2020-09-03 | 2020-12-04 | 福建库克智能科技有限公司 | Face detection method and device and electronic equipment |
CN112085540A (en) * | 2020-09-27 | 2020-12-15 | 湖北科技学院 | Intelligent advertisement pushing system and method based on artificial intelligence technology |
GB2588747A (en) * | 2019-06-28 | 2021-05-12 | Huawei Tech Co Ltd | Facial behaviour analysis |
CN113537115A (en) * | 2021-07-26 | 2021-10-22 | 东软睿驰汽车技术(沈阳)有限公司 | Method and device for acquiring driving state of driver and electronic equipment |
CN117275069A (en) * | 2023-09-26 | 2023-12-22 | 华中科技大学 | End-to-end head gesture estimation method based on learnable vector and attention mechanism |
WO2024027146A1 (en) * | 2022-08-01 | 2024-02-08 | 五邑大学 | Array-type facial beauty prediction method, and device and storage medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105654049A (en) * | 2015-12-29 | 2016-06-08 | 中国科学院深圳先进技术研究院 | Facial expression recognition method and device |
CN106529402A (en) * | 2016-09-27 | 2017-03-22 | 中国科学院自动化研究所 | Multi-task learning convolutional neural network-based face attribute analysis method |
CN106599883A (en) * | 2017-03-08 | 2017-04-26 | 王华锋 | Face recognition method capable of extracting multi-level image semantics based on CNN (convolutional neural network) |
US20170344808A1 (en) * | 2016-05-28 | 2017-11-30 | Samsung Electronics Co., Ltd. | System and method for a unified architecture multi-task deep learning machine for object recognition |
CN107423690A (en) * | 2017-06-26 | 2017-12-01 | 广东工业大学 | A kind of face identification method and device |
-
2018
- 2018-10-26 CN CN201811260674.XA patent/CN109508654B/en not_active Expired - Fee Related
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105654049A (en) * | 2015-12-29 | 2016-06-08 | 中国科学院深圳先进技术研究院 | Facial expression recognition method and device |
US20170344808A1 (en) * | 2016-05-28 | 2017-11-30 | Samsung Electronics Co., Ltd. | System and method for a unified architecture multi-task deep learning machine for object recognition |
US20180307897A1 (en) * | 2016-05-28 | 2018-10-25 | Samsung Electronics Co., Ltd. | System and method for a unified architecture multi-task deep learning machine for object recognition |
CN106529402A (en) * | 2016-09-27 | 2017-03-22 | 中国科学院自动化研究所 | Multi-task learning convolutional neural network-based face attribute analysis method |
CN106599883A (en) * | 2017-03-08 | 2017-04-26 | 王华锋 | Face recognition method capable of extracting multi-level image semantics based on CNN (convolutional neural network) |
CN107423690A (en) * | 2017-06-26 | 2017-12-01 | 广东工业大学 | A kind of face identification method and device |
Non-Patent Citations (8)
Title |
---|
CHEN, Q.,ET.AL: "A multi-scale fusion convolutional neural network for face detection", 《2017 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS》 * |
RAJEEV RANJAN,ET.AL: "Hyperface: A deep multi-task learning framework for face detection, landmark localization, pose estimation, and gender recognition", 《IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE》 * |
RANJAN, R.,ET.AL: "Hyperface: A deep multi-task learning framework for face detection, landmark localization, pose estimation, and gender recognition.", 《IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE》 * |
ZHU, C.,ET.AL: "Cms-rcnn: contextual multi-scale region-based cnn for unconstrained face detection", 《DEEP LEARNING FOR BIOMETRICS》 * |
刘璐: "基于深度神经网络的多任务视觉感知研究与应用", 《中国优秀硕士学位论文全文数据库 信息科技辑》 * |
姜涛: "室外动态场景中的人脸识别", 《中国优秀硕士学位论文全文数据库 信息科技辑》 * |
翁昊: "随机多尺度核学习及应用", 《中国优秀博士学位论文全文数据库 信息科技辑》 * |
鄢宇烈: "一种在线学习实时长期目标跟踪算法研究与实现", 《中国优秀硕士学位论文全文数据库 信息科技辑》 * |
Cited By (43)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110135251A (en) * | 2019-04-09 | 2019-08-16 | 上海电力学院 | A kind of group's image Emotion identification method based on attention mechanism and hybrid network |
CN110135251B (en) * | 2019-04-09 | 2023-08-08 | 上海电力学院 | Group image emotion recognition method based on attention mechanism and hybrid network |
CN110082822B (en) * | 2019-04-09 | 2020-07-28 | 中国科学技术大学 | Method for detecting earthquake by using convolution neural network |
CN110082822A (en) * | 2019-04-09 | 2019-08-02 | 中国科学技术大学 | The method for carrying out earthquake detection using convolutional neural networks |
CN110188615A (en) * | 2019-04-30 | 2019-08-30 | 中国科学院计算技术研究所 | A kind of facial expression recognizing method, device, medium and system |
CN110188615B (en) * | 2019-04-30 | 2021-08-06 | 中国科学院计算技术研究所 | Facial expression recognition method, device, medium and system |
CN110135373A (en) * | 2019-05-20 | 2019-08-16 | 北京探境科技有限公司 | A kind of multiple dimensioned face recognition method, system and electronic equipment |
CN110457572A (en) * | 2019-05-23 | 2019-11-15 | 北京邮电大学 | Commodity information recommendation method and electronic equipment based on figure network |
CN110457572B (en) * | 2019-05-23 | 2022-05-24 | 北京邮电大学 | Commodity information recommendation method based on graph network and electronic equipment |
CN110232652A (en) * | 2019-05-27 | 2019-09-13 | 珠海格力电器股份有限公司 | Image processing engine processing method, the image processing method for terminal, terminal |
CN110287846A (en) * | 2019-06-19 | 2019-09-27 | 南京云智控产业技术研究院有限公司 | A kind of face critical point detection method based on attention mechanism |
CN110287846B (en) * | 2019-06-19 | 2023-08-04 | 南京云智控产业技术研究院有限公司 | Attention mechanism-based face key point detection method |
GB2588747B (en) * | 2019-06-28 | 2021-12-08 | Huawei Tech Co Ltd | Facial behaviour analysis |
GB2588747A (en) * | 2019-06-28 | 2021-05-12 | Huawei Tech Co Ltd | Facial behaviour analysis |
CN110298325A (en) * | 2019-07-02 | 2019-10-01 | 四川长虹电器股份有限公司 | Expression impaired patients assisted care system based on video Expression Recognition |
CN110348416A (en) * | 2019-07-17 | 2019-10-18 | 北方工业大学 | Multi-task face recognition method based on multi-scale feature fusion convolutional neural network |
CN110633624A (en) * | 2019-07-26 | 2019-12-31 | 北京工业大学 | Machine vision human body abnormal behavior identification method based on multi-feature fusion |
CN110633624B (en) * | 2019-07-26 | 2022-11-22 | 北京工业大学 | Machine vision human body abnormal behavior identification method based on multi-feature fusion |
CN110263774A (en) * | 2019-08-19 | 2019-09-20 | 珠海亿智电子科技有限公司 | A kind of method for detecting human face |
CN110705563B (en) * | 2019-09-07 | 2020-12-29 | 创新奇智(重庆)科技有限公司 | Industrial part key point detection method based on deep learning |
CN110705563A (en) * | 2019-09-07 | 2020-01-17 | 创新奇智(重庆)科技有限公司 | Industrial part key point detection method based on deep learning |
CN110866466B (en) * | 2019-10-30 | 2023-12-26 | 平安科技(深圳)有限公司 | Face recognition method, device, storage medium and server |
CN110866466A (en) * | 2019-10-30 | 2020-03-06 | 平安科技(深圳)有限公司 | Face recognition method, face recognition device, storage medium and server |
CN111062269A (en) * | 2019-11-25 | 2020-04-24 | 珠海格力电器股份有限公司 | User state identification method and device, storage medium and air conditioner |
CN111178183B (en) * | 2019-12-16 | 2023-05-23 | 深圳市华尊科技股份有限公司 | Face detection method and related device |
CN111178183A (en) * | 2019-12-16 | 2020-05-19 | 深圳市华尊科技股份有限公司 | Face detection method and related device |
CN111325108B (en) * | 2020-01-22 | 2023-05-26 | 中能国际高新科技研究院有限公司 | Multitasking network model, using method, device and storage medium |
CN111325108A (en) * | 2020-01-22 | 2020-06-23 | 中能国际建筑投资集团有限公司 | Multitask network model, using method, device and storage medium |
CN111291670B (en) * | 2020-01-23 | 2023-04-07 | 天津大学 | Small target facial expression recognition method based on attention mechanism and network integration |
CN111291670A (en) * | 2020-01-23 | 2020-06-16 | 天津大学 | Small target facial expression recognition method based on attention mechanism and network integration |
CN111507248A (en) * | 2020-04-16 | 2020-08-07 | 成都东方天呈智能科技有限公司 | Face forehead area detection and positioning method and system of low-resolution thermodynamic diagram |
CN111507248B (en) * | 2020-04-16 | 2023-05-26 | 成都东方天呈智能科技有限公司 | Face forehead region detection and positioning method and system based on low-resolution thermodynamic diagram |
CN111814706B (en) * | 2020-07-14 | 2022-06-24 | 电子科技大学 | Face recognition and attribute classification method based on multitask convolutional neural network |
CN111814706A (en) * | 2020-07-14 | 2020-10-23 | 电子科技大学 | Face recognition and attribute classification method based on multitask convolutional neural network |
CN111914758A (en) * | 2020-08-04 | 2020-11-10 | 成都奥快科技有限公司 | Face in-vivo detection method and device based on convolutional neural network |
CN112036260A (en) * | 2020-08-10 | 2020-12-04 | 武汉星未来教育科技有限公司 | Expression recognition method and system for multi-scale sub-block aggregation in natural environment |
CN112036339A (en) * | 2020-09-03 | 2020-12-04 | 福建库克智能科技有限公司 | Face detection method and device and electronic equipment |
CN112036339B (en) * | 2020-09-03 | 2024-04-09 | 福建库克智能科技有限公司 | Face detection method and device and electronic equipment |
CN112085540A (en) * | 2020-09-27 | 2020-12-15 | 湖北科技学院 | Intelligent advertisement pushing system and method based on artificial intelligence technology |
CN113537115A (en) * | 2021-07-26 | 2021-10-22 | 东软睿驰汽车技术(沈阳)有限公司 | Method and device for acquiring driving state of driver and electronic equipment |
WO2024027146A1 (en) * | 2022-08-01 | 2024-02-08 | 五邑大学 | Array-type facial beauty prediction method, and device and storage medium |
CN117275069A (en) * | 2023-09-26 | 2023-12-22 | 华中科技大学 | End-to-end head gesture estimation method based on learnable vector and attention mechanism |
CN117275069B (en) * | 2023-09-26 | 2024-06-04 | 华中科技大学 | End-to-end head gesture estimation method based on learnable vector and attention mechanism |
Also Published As
Publication number | Publication date |
---|---|
CN109508654B (en) | 2021-01-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109508654A (en) | Merge the human face analysis method and system of multitask and multiple dimensioned convolutional neural networks | |
Sun et al. | Deep spatial-temporal feature fusion for facial expression recognition in static images | |
CN111339903B (en) | Multi-person human body posture estimation method | |
Schmidt et al. | Self-supervised visual descriptor learning for dense correspondence | |
CN112800903B (en) | Dynamic expression recognition method and system based on space-time diagram convolutional neural network | |
CN109559320A (en) | Realize that vision SLAM semanteme builds the method and system of figure function based on empty convolution deep neural network | |
CN110097553A (en) | The semanteme for building figure and three-dimensional semantic segmentation based on instant positioning builds drawing system | |
Deng et al. | MVF-Net: A multi-view fusion network for event-based object classification | |
CN111625667A (en) | Three-dimensional model cross-domain retrieval method and system based on complex background image | |
WO2023024658A1 (en) | Deep video linkage feature-based behavior recognition method | |
Núñez et al. | Multiview 3D human pose estimation using improved least-squares and LSTM networks | |
Kumar et al. | 3D sign language recognition using spatio temporal graph kernels | |
US20230222841A1 (en) | Ensemble Deep Learning Method for Identifying Unsafe Behaviors of Operators in Maritime Working Environment | |
Sun et al. | F3-Net: Multiview scene matching for drone-based geo-localization | |
CN112906520A (en) | Gesture coding-based action recognition method and device | |
Amrutha et al. | Human Body Pose Estimation and Applications | |
CN115018999A (en) | Multi-robot-cooperation dense point cloud map construction method and device | |
Yin et al. | Msa-gcn: Multiscale adaptive graph convolution network for gait emotion recognition | |
CN113076905B (en) | Emotion recognition method based on context interaction relation | |
Jiao et al. | Facial attention based convolutional neural network for 2D+ 3D facial expression recognition | |
Kumar | Motion trajectory based human face and hands tracking for sign language recognition | |
CN116311518A (en) | Hierarchical character interaction detection method based on human interaction intention information | |
CN114973305B (en) | Accurate human body analysis method for crowded people | |
de Paiva et al. | Estimating human body orientation using skeletons and extreme gradient boosting | |
CN113688864B (en) | Human-object interaction relation classification method based on split attention |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20210105 Termination date: 20211026 |