CN108985181A - A kind of end-to-end face mask method based on detection segmentation - Google Patents

A kind of end-to-end face mask method based on detection segmentation Download PDF

Info

Publication number
CN108985181A
CN108985181A CN201810654160.6A CN201810654160A CN108985181A CN 108985181 A CN108985181 A CN 108985181A CN 201810654160 A CN201810654160 A CN 201810654160A CN 108985181 A CN108985181 A CN 108985181A
Authority
CN
China
Prior art keywords
face
segmentation
sub
network
module
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201810654160.6A
Other languages
Chinese (zh)
Other versions
CN108985181B (en
Inventor
温世平
董明辉
黄廷文
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huazhong University of Science and Technology
Original Assignee
Huazhong University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huazhong University of Science and Technology filed Critical Huazhong University of Science and Technology
Priority to CN201810654160.6A priority Critical patent/CN108985181B/en
Publication of CN108985181A publication Critical patent/CN108985181A/en
Application granted granted Critical
Publication of CN108985181B publication Critical patent/CN108985181B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • G06V10/267Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a kind of end-to-end face mask methods based on detection segmentation, comprising: the region that every facial image needs to detect in mark facial image training set;End-to-end face segmentation mark neural network model is constructed, the neural network model includes being made of sharing feature module, face composition detection module, ROI feature extraction module, three sub- network modules, three sorter network modules;Based on facial image training set, the end-to-end face segmentation mark neural network model of training, modules synchronize training;Divide mark neural network model using trained face and mark is split to test facial image.The present invention obtains the sharing feature of different scale using deconvolution up-sampling step by step and height Fusion Features, captures more details information;By designing sub-network for the less semantic classes of distribution area, segmentation precision is improved;The ability to express of entire model is promoted by unified training.

Description

A kind of end-to-end face mask method based on detection segmentation
Technical field
The invention belongs to artificial intelligence Image Information Processing fields, more particularly, to a kind of end based on detection segmentation Opposite end face mask method.
Background technique
Face mark, which refers to, carries out region segmentation to the image comprising face, and difference according to demand divides the image into 3 Class (hair, background, face) or more multiclass (face is continued to divide).Face mark is used as front-end processing, has been applied to In the applications such as recognition of face, virtual makeups, face exchange.In such applications, effect of the precision of face mark to inter-related task Fruit plays very crucial effect.
With the promotion of depth learning technology development and big data computing capability, deep learning is applied to people by researcher Face mark, makes precision be greatly improved.Face mask method based on deep learning is broadly divided into based on sliding window Method and method based on full convolutional neural networks, since the method process based on sliding window is complicated and compared to full convolutional Neural The method precision of network does not have advantage, is all based on the face mark of the deep learning of full convolutional neural networks greatly in the prior art Method.
Patent CN105354565A is disclosed one kind and is positioned based on full convolutional network human face five-sense-organ and sentence method for distinguishing, is passed through It collects facial image and mark face forms training set by hand, using the full convolutional neural networks of training set training, by face to be measured Image is input to trained full convolutional neural networks, the face segmentation result and facial feature localization and mark of output pixel rank. Although it discloses the human face five-sense-organ mask method based on full convolutional neural networks, consider sample in distribution area compared with Few semantic classes has that the area markings such as eyes, mouth, nose are inaccurate.
Patent CN107729819A discloses a kind of face mask method based on sparse full convolutional neural networks, passes through receipts Collect facial image and by hand mark hair, skin, background formed training set, using full convolutional neural networks semantic segmentation method with Group lasso Sparse methods constitute model and facial image to be measured are then input to training using training set to model training Good model is completed to mark the Pixel-level of hair, skin and background in facial image.Although it discloses based on full convolution mind Face dimensioning algorithm through network, but do not consider the segmentation of facial face, there is a problem of that application scenarios are not extensive enough.
Summary of the invention
In view of the drawbacks of the prior art, it is an object of the invention to solve the face dividing method based on artificial design features The problem of existing precision is low and poor robustness, while solving traditional face dividing method based on full convolutional neural networks and depositing Output homogeneity problem, and designed method can be very good the processing classification present in face segmentation task point The unbalanced problem of cloth.
To achieve the above object, the embodiment of the invention provides a kind of end-to-end face mark sides based on detection segmentation Method, comprising the following steps:
S1. the region that every facial image needs to detect in facial image training set is marked;
S2. an end-to-end face segmentation mark neural network model is constructed, the neural network model is by shared spy Module, face composition detection module, ROI feature extraction module, three sub- network modules, three sorter network modules composition are levied, In, the input of the sharing feature module is facial image, is exported as sharing feature figure;The face composition detection module it is defeated Enter the position frame information exported for sharing feature figure as each region for needing to detect;The input of the ROI feature extraction module is Sharing feature figure and each position frame information export as each ROI subcharacter figure;The input of the sub-network module is ROI subcharacter Figure, exports the segmentation result for each subregion;The input of the three sorter networks module is sharing feature figure, is exported as head The segmentation result of hair, skin and background;
S3. it is based on facial image training set, the end-to-end face segmentation of training marks neural network model, and modules are same Step training;
S4. divide mark neural network model using trained face and mark is split to test facial image, three The segmentation result of sorter network and the fusion of the segmentation result of each sub-network are segmentation annotation results.
Specifically, the region for needing to detect described in step S1 includes: left eyebrow, right eyebrow, left eye eyeball, right eye eyeball, nose Son, upper lip, in mouth, lower lip, skin, hair, background.
Specifically, sharing feature module described in step S2 uses coding-decoding structure, and coding structure passes through full convolution mind Characteristic pattern C1-CN is converted by facial image through network;Decode structure by characteristic pattern C1-CN step by step deconvolution up-sampling and height Fusion Features obtain sharing feature figure P1-PN.
Specifically, face composition detection module described in step S2 exports the position frame information of different ingredients.
Specifically, ROI feature extraction module described in step S2 is after obtaining the position ROI frame information, from sharing feature figure ROI feature is intercepted in P1-PN respectively, identical size is uniformly transformed to, the fusion of feature concatenation, is uniformly transformed to identical channel After dimension, each ROI subcharacter figure is obtained.
Specifically, three sub- network modules include that eyes add eyebrow sub-network module, nose sub-network module, mouth Sub-network module three classes are responsible for that individual features are further processed, and the segmentation result of decoded output subregion, In, subregion is divided into 2 semantic classes of nose and background by nasal area sub-network, and mouth region sub-network is by mouth region Be divided into upper lip, in mouth, 4 semantic classes of lower lip and background, eyes add eyebrow sub-network that eyes are added brow region point It is segmented into 3 eyes, eyebrow and background semantic classes.
Specifically, characteristic dimension is risen to input picture by warp lamination and convolutional layer by the three sorter networks module Dimension, the segmentation result of output skin area, hair zones and background area.
Specifically, the optimization loss function of training process is defined as follows:
Lall=Lseg+Ldet+Lreg
Wherein, LallFor global optimization loss function, LregFor regularization loss function, asked for alleviating model over-fitting Topic, LdetFor face composition detection module loss function, loss, L are returned comprising Classification Loss and positionsegLetter is lost for segmentation Number, is defined as follows:
Wherein,Divide loss functions for three classification,For three classes community finding loss function, 4 segmentations Loss function is all made of cross entropy loss function.
In general, through the invention it is contemplated above technical scheme is compared with the prior art, have below beneficial to effect Fruit:
(1) present invention obtains sharing feature by the way of deconvolution up-sampling step by step and height Fusion Features, compared to biography It unites full convolutional neural networks model, can capture more details profile informations, alleviate homogeneous problem.Simultaneously as using The profile information of different scale, network model can learn the information to more different scales from data.
(2) present invention employs sub-network structure designs, individually set for the less semantic classes of distribution area in the sample A sub-network structure is counted, the information for losing the category in trained and test process is avoided.It is designed by this, is alleviated significantly Since sample class is distributed uneven the problem of reducing segmentation precision.
(3) all modules of the invention can carry out end-to-end unified training, be not necessarily to stage by stage or sub-module is trained, In unified training process, the ability to express of entire model can be promoted with sharing feature information between modules.
Detailed description of the invention
Fig. 1 is a kind of end-to-end face mask method flow chart based on detection segmentation provided by the invention.
Fig. 2 is ROI feature extraction module operation principle schematic diagram provided by the invention.
Fig. 3 is mouth region sub-network flow diagram provided by the invention.
Fig. 4 provides a kind of end-to-end face mask method segmentation result schematic diagram based on detection segmentation for the present invention.
Specific embodiment
In order to make the objectives, technical solutions, and advantages of the present invention clearer, with reference to the accompanying drawings and embodiments, right The present invention is further elaborated.It should be appreciated that the specific embodiments described herein are merely illustrative of the present invention, and It is not used in the restriction present invention.
Fig. 1 is a kind of end-to-end face mask method flow chart based on detection segmentation provided by the invention.Such as Fig. 1 institute Show, method includes the following steps:
S1. the region that every facial image needs to detect in facial image training set is marked;
S2. an end-to-end face segmentation mark neural network model is constructed, the neural network model is by shared spy Levy module, face composition detection module, ROI feature extraction module, three sub- network modules, three sorter network modules composition;
S3. it is based on facial image training set, the end-to-end face segmentation of training marks neural network model, and modules are same Step training;
S4. divide mark neural network model using trained face and mark is split to test facial image.
The region for needing to detect described in step S1 includes: left eyebrow, right eyebrow, left eye eyeball, right eye eyeball, nose, upper mouth In lip, mouth, lower lip, skin, hair, background.
The input of sharing feature module described in step S2 is facial image, is exported as sharing feature figure;The face at The input for dividing detection module is sharing feature figure, exports and adds left brow region, right eye eyeball to add right brow region, nose for left eye eyeball The position frame information in region, mouth region and facial area;The input of the ROI feature extraction module is for sharing feature figure and respectively Position frame information exports as each ROI subcharacter figure;The input of the sub-network module is ROI subcharacter figure, is exported as every height The segmentation result in region;The input of the three sorter networks module is sharing feature figure, exports point for hair, skin and background Cut result.
Sharing feature module described in step S2 uses coding-decoding structure, and coding structure passes through full convolutional neural networks Characteristic pattern C1-CN is converted by facial image;Decoding structure, deconvolution up-sampling and height feature are melted step by step by characteristic pattern C1-CN Conjunction obtains sharing feature figure P1-PN.
The process of deconvolution up-sampling and height Fusion Features is as follows step by step by characteristic pattern C1-CN: first by characteristic pattern C1-CN It is unified for identical dimension of the channel K, the characteristic pattern CN after unified channel dimension is sharing feature figure PN, then characteristic pattern Ci reversed Dimension of the channel K, which is unified for, after being added after convolution with adjacent characteristic pattern C (i-1) obtains sharing feature figure P (i-1), i=N ..., 2, Finally obtain sharing feature figure P1-PN.
Specifically, coding structure uses Res50 network module, and Res50 module is one kind that Kaiming He et al. is proposed Convolutional neural networks model with residual error connection structure has stronger feature representation ability and information transfer capacity.It will The last layer convolutional layer of 5 convolution blocks is denoted as characteristic pattern C1-C5 respectively in Res50;Decoding structure first will be to this 5 convolution Layer uses the convolution kernel of 1 × 1 size respectively, and dimension of the channel is uniformly dropped to 256, the use of step-length is step by step 2 then since C5 Deconvolution operation by characteristic pattern scale size promoted 2 times, then with corresponding next stage characteristic pattern be added, for example, C5 promoted 2 It is added after times dimension of the channel with C4, characteristic pattern after being added imposes 3 × 3 convolution kernel again, and dimension of the channel 256 obtains most Whole sharing feature figure P1-P5.
For face composition detection module described in step S2 using Faster RCNN structure as detector, the detector is defeated The position frame information of different ingredients out, the rectangular area that position frame is indicated where target using 5, format are rectangular area The height and width of upper left corner x, y and rectangle.Facial area without feature extraction, the purpose for detecting facial area is to make detector Have when detecting face ingredient it is certain stress region, improve learning model in the precision of face composition detection.Other The position frame information in four class regions is used for the extraction of individual features.
ROI feature extraction module described in step S2 is after obtaining the position ROI frame information, from sharing feature figure P1-PN ROI feature is intercepted respectively, after being uniformly transformed to identical size, the fusion of feature concatenation, being uniformly transformed to identical dimension of the channel, Obtain each ROI subcharacter figure.
Specifically, high-level characteristic generally comprises more semantic informations, and low-level feature generally comprises more detailed information, In face segmentation task, these two types of information are all particularly significant.Fig. 2 is that ROI feature extraction module provided by the invention work is former Manage schematic diagram.As shown in Fig. 2, in order to integrate the feature of different levels, the present invention after the position frame information for obtaining some ROI, ROI feature is intercepted from the sharing feature figure of tetra- levels of P1-P4 respectively.For the ease of training, different size of ROI feature quilt Uniformly transform to identical size 14 × 14.After obtaining the subcharacter figure that four groups of channels are 256, by four groups of subcharacter figures according to The sequential concatenation of P1-P4 together, is then obtained dimension of the channel dimensionality reduction 256 final extracted using 1 × 1 convolution kernel Subcharacter, the part are ROI feature extraction module.Extracted subcharacter will be entered in different sub-networks, individually It is handled.
The sub-network module includes that eyes add eyebrow sub-network module, nose sub-network module, mouth sub-network module Three classes, wherein eyes add the input of eyebrow sub-network module to be that left eye eyeball adds left eyebrow characteristic pattern and right eye eyeball to add right eyebrow special Sign figure, exports and adds left brow region and right eye eyeball to add right brow region for left eye eyeball;The input of nose sub-network module is nose Characteristic pattern exports as nasal area;The input of mouth sub-network module is mouth subcharacter figure, is exported as mouth region.
Sub-network is responsible for that individual features are further processed, and the segmentation result of decoded output subregion.Wherein Subregion is divided into 2 semantic classes of nose and background by nasal area sub-network, and mouth region sub-network divides mouth region Be segmented into upper lip, in mouth, 4 semantic classes of lower lip and background, eyes add eyebrow sub-network to add brow region to divide eyes For 3 eyes, eyebrow and background semantic classes.ROI feature figure is re-converted into original ROI size by sub-network, and will be defeated Result is placed on corresponding position out, as shown in Figure 3.
Fig. 3 is mentioned mouth region sub-network flow diagram in method by the present invention, and nasal area and eyes add eyebrow area Domain process flow having the same, three sub-networks are run parallel, and three sub-networks are full convolutional neural networks network model, Specifically, sub-network rises dimension, the result of output 56 × 56 for Fig. 4 times of ROI subcharacter.
Since the class imbalance phenomenon of skin, hair and background is not serious, It is not necessary to the individual subnet of design Network structure.Characteristic dimension is risen to input picture dimension, is then exported by three sorter network modules by warp lamination and convolutional layer Skin area, hair zones and the segmentation result of background area.Decoder rear end part forms three sorter network modules.
Specifically, after C1 characteristic pattern is added with C2 liter dimensional feature figure, convolutional layer 1 (256 channels, 3 × 3 convolution kernels), Warp lamination 1 (128 channels, step-length 2), convolutional layer 2 (128 channels, 3 × 3 convolution kernels) and convolutional layer 3 (3 channels, volume 3 × 3 Product core) it is sequentially overlapped, output hair, skin and the other segmentation result of three type of background, the part are three sorter network pattern dies Block.
In training process, face composition detection module mentions for the position the ROI frame that each sub-network provides 20 respective classes View, 20 proposal extracted features of position frame are exported by corresponding sub-network as a result, then true with corresponding position Value, which calculates, intersects entropy loss, and combination is used to optimize network parameter by the intersection entropy loss of all proposals, and the true value is from artificial Divide face and divides database.
The optimization loss function of training process is defined as follows:
Lall=Lseg+Ldet+Lreg
Wherein, LallFor global optimization loss function, LregFor regularization loss function, asked for alleviating model over-fitting Topic, LdetFor face composition detection module loss function, it is original fixed that partial loss function definition is derived from Faster RCNN model Justice returns loss, L comprising Classification Loss and positionsegFor segmentation loss function, it is defined as follows:
Wherein,Divide loss functions for three classification,For three classes community finding loss function, 4 segmentations Loss function is all made of cross entropy loss function.
Step 4: mark being split to facial image to be processed using algorithm model trained in step 3.
The segmentation result of three sorter networks and the fusion of the segmentation result of each sub-network just obtain the classification segmentation mark of face 11 Infuse result.(left eye, right eye, Zuo Mei after acquisition mouth region respectively, nasal area, eyes add the segmentation result of brow region In hair, right eyebrow, nose, upper lip, mouth, lower lip), with three sorter networks export result (skin, hair, background) phase group Close, obtain final 11 classifications (left eye, right eye, left eyebrow, right eyebrow, nose, upper lip, in mouth, lower lip, skin, head Hair, background) segmentation result.
Fig. 4 provides a kind of end-to-end face mask method segmentation result schematic diagram based on detection segmentation for the present invention.Such as Shown in Fig. 4, the first behavior input picture, the second behavior segmentation result, the corresponding true value of third behavior.Shown sample is on head Posture, hair style, the colour of skin, block, expression etc. all has biggish variation, mentioned method model still can obtain preferably Segmentation result.
More than, the only preferable specific embodiment of the application, but the protection scope of the application is not limited thereto, and it is any Within the technical scope of the present application, any changes or substitutions that can be easily thought of by those familiar with the art, all answers Cover within the scope of protection of this application.Therefore, the protection scope of the application should be subject to the protection scope in claims.

Claims (8)

1. a kind of end-to-end face mask method based on detection segmentation, which comprises the following steps:
S1. the region that every facial image needs to detect in facial image training set is marked;
S2. an end-to-end face segmentation mark neural network model is constructed, the neural network model is by sharing feature mould Block, face composition detection module, ROI feature extraction module, three sub- network modules, three sorter network modules composition, wherein institute The input for stating sharing feature module is facial image, is exported as sharing feature figure;The input of the face composition detection module is Sharing feature figure exports as the position frame information in each region for needing to detect;The input of the ROI feature extraction module is shared Characteristic pattern and each position frame information export as each ROI subcharacter figure;The input of the sub-network module is ROI subcharacter figure, defeated It is out the segmentation result of each subregion;The input of the three sorter networks module is sharing feature figure, is exported as hair, skin With the segmentation result of background;
S3. facial image training set, the end-to-end face segmentation mark neural network model of training, the synchronous instruction of modules are based on Practice;
S4. divide mark neural network model using trained face and mark, three classification are split to test facial image The segmentation result of network and the fusion of the segmentation result of each sub-network are segmentation annotation results.
2. face mask method as described in claim 1, which is characterized in that the region packet for needing to detect described in step S1 Include: left eyebrow, right eyebrow, left eye eyeball, right eye eyeball, nose, upper lip, in mouth, lower lip, skin, hair, background.
3. face mask method as claimed in claim 1 or 2, which is characterized in that sharing feature module described in step S2 is adopted With coding-decoding structure, coding structure converts characteristic pattern C1-CN for facial image by full convolutional neural networks;Decoding knot By characteristic pattern C1-CN, deconvolution up-sampling and height Fusion Features obtain sharing feature figure P1-PN to structure step by step.
4. face mask method as claimed in claim 1 or 2, which is characterized in that face composition detection mould described in step S2 Block exports the position frame information of different ingredients.
5. face mask method as claimed in claim 4, which is characterized in that ROI feature extraction module described in step S2 exists After obtaining the position ROI frame information, intercept ROI feature respectively from sharing feature figure P1-PN, be uniformly transformed to identical size, The fusion of feature concatenation after being uniformly transformed to identical dimension of the channel, obtains each ROI subcharacter figure.
6. face mask method as claimed in claim 1 or 2, which is characterized in that three sub- network modules include eyes Add eyebrow sub-network module, nose sub-network module, mouth sub-network module three classes, is responsible for carrying out individual features further Processing, and the segmentation result of decoded output subregion, wherein subregion is divided into nose and background 2 by nasal area sub-network A semantic classes, mouth region sub-network by mouth region segmentation be upper lip, in mouth, 4 semantic classes of lower lip and background, Eyes add eyebrow sub-network to add brow region to be divided into 3 eyes, eyebrow and background semantic classes eyes.
7. face mask method as claimed in claim 1 or 2, which is characterized in that the three sorter networks module passes through warp Characteristic dimension is risen to input picture dimension, exports the segmentation of skin area, hair zones and background area by lamination and convolutional layer As a result.
8. face mask method as claimed in claim 1 or 2, which is characterized in that the optimization loss function of training process defines It is as follows:
Lall=Lseg+Ldet+Lreg
Wherein, LallFor global optimization loss function, LregFor regularization loss function, for alleviating model overfitting problem, Ldet For face composition detection module loss function, loss, L are returned comprising Classification Loss and positionsegTo divide loss function, determine Justice is as follows:
Wherein,Divide loss functions for three classification,For three classes community finding loss function, 4 segmentation losses Function is all made of cross entropy loss function.
CN201810654160.6A 2018-06-22 2018-06-22 End-to-end face labeling method based on detection segmentation Expired - Fee Related CN108985181B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810654160.6A CN108985181B (en) 2018-06-22 2018-06-22 End-to-end face labeling method based on detection segmentation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810654160.6A CN108985181B (en) 2018-06-22 2018-06-22 End-to-end face labeling method based on detection segmentation

Publications (2)

Publication Number Publication Date
CN108985181A true CN108985181A (en) 2018-12-11
CN108985181B CN108985181B (en) 2020-07-24

Family

ID=64538442

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810654160.6A Expired - Fee Related CN108985181B (en) 2018-06-22 2018-06-22 End-to-end face labeling method based on detection segmentation

Country Status (1)

Country Link
CN (1) CN108985181B (en)

Cited By (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109657596A (en) * 2018-12-12 2019-04-19 天津卡达克数据有限公司 A kind of vehicle appearance component identification method based on deep learning
CN109741338A (en) * 2018-12-13 2019-05-10 北京爱奇艺科技有限公司 A kind of face dividing method, device and equipment
CN109801249A (en) * 2018-12-27 2019-05-24 深圳豪客互联网有限公司 Image interfusion method, device, computer equipment and storage medium
CN109829374A (en) * 2018-12-27 2019-05-31 深圳豪客互联网有限公司 Image processing method, device, computer equipment and storage medium
CN109903257A (en) * 2019-03-08 2019-06-18 上海大学 A kind of virtual hair-dyeing method based on image, semantic segmentation
CN110059768A (en) * 2019-04-30 2019-07-26 福州大学 The semantic segmentation method and system of the merging point and provincial characteristics that understand for streetscape
CN110148085A (en) * 2019-04-22 2019-08-20 智慧眼科技股份有限公司 Face image super-resolution reconstruction method and computer-readable storage medium
CN110287930A (en) * 2019-07-01 2019-09-27 厦门美图之家科技有限公司 Wrinkle disaggregated model training method and device
CN110378278A (en) * 2019-07-16 2019-10-25 北京地平线机器人技术研发有限公司 Training method, object search method, apparatus and the electronic equipment of neural network
CN110472605A (en) * 2019-08-21 2019-11-19 广州纳丽生物科技有限公司 A kind of skin problem diagnostic method based on deep learning face subregion
CN111144310A (en) * 2019-12-27 2020-05-12 创新奇智(青岛)科技有限公司 Face detection method and system based on multi-layer information fusion
CN111179287A (en) * 2020-01-03 2020-05-19 厦门美图之家科技有限公司 Portrait instance segmentation method, device, equipment and storage medium
CN111310718A (en) * 2020-03-09 2020-06-19 成都川大科鸿新技术研究所 High-accuracy detection and comparison method for face-shielding image
CN111382746A (en) * 2018-12-29 2020-07-07 杭州光启人工智能研究院 Data marking method, computer device and computer readable storage medium
CN111666905A (en) * 2020-06-10 2020-09-15 重庆紫光华山智安科技有限公司 Model training method, pedestrian attribute identification method and related device
CN111723596A (en) * 2019-03-18 2020-09-29 北京市商汤科技开发有限公司 Method, device and equipment for detecting gazing area and training neural network
CN111739025A (en) * 2020-05-08 2020-10-02 北京迈格威科技有限公司 Image processing method, device, terminal and storage medium
CN112330696A (en) * 2020-12-02 2021-02-05 青岛大学 Face segmentation method, face segmentation device and computer-readable storage medium
CN112560701A (en) * 2020-12-17 2021-03-26 成都新潮传媒集团有限公司 Face image extraction method and device and computer storage medium
CN112669197A (en) * 2019-10-16 2021-04-16 顺丰科技有限公司 Image processing method, image processing device, mobile terminal and storage medium
CN112733632A (en) * 2020-12-28 2021-04-30 华南理工大学 Robot control method based on face recognition and gesture recognition
CN113052247A (en) * 2021-03-31 2021-06-29 清华苏州环境创新研究院 Garbage classification method and garbage classifier based on multi-label image recognition
CN113469040A (en) * 2021-06-30 2021-10-01 北京市商汤科技开发有限公司 Image processing method and device, computer equipment and storage medium
CN116269285A (en) * 2022-11-28 2023-06-23 电子科技大学 Non-contact normalized heart rate variability estimation system

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100508202B1 (en) * 2003-04-14 2005-08-17 주식회사 윈포넷 A face feature extraction method for a face detection
CN105654420A (en) * 2015-12-21 2016-06-08 小米科技有限责任公司 Face image processing method and device
CN105869159A (en) * 2016-03-28 2016-08-17 联想(北京)有限公司 Image segmentation method and apparatus
CN107507211A (en) * 2017-07-24 2017-12-22 中国科学院合肥物质科学研究院 Remote sensing image segmentation method based on multi-Agent and MRF

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100508202B1 (en) * 2003-04-14 2005-08-17 주식회사 윈포넷 A face feature extraction method for a face detection
CN105654420A (en) * 2015-12-21 2016-06-08 小米科技有限责任公司 Face image processing method and device
CN105869159A (en) * 2016-03-28 2016-08-17 联想(北京)有限公司 Image segmentation method and apparatus
CN107507211A (en) * 2017-07-24 2017-12-22 中国科学院合肥物质科学研究院 Remote sensing image segmentation method based on multi-Agent and MRF

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
SIFEI,LIU ET AL.: "Face Parsing via Recurrent Propagation", 《ARXIV》 *
牛新亚: "基于深度学习的人脸表情识别研究", 《中国优秀硕士学位论文全文数据库》 *

Cited By (35)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109657596A (en) * 2018-12-12 2019-04-19 天津卡达克数据有限公司 A kind of vehicle appearance component identification method based on deep learning
CN109741338A (en) * 2018-12-13 2019-05-10 北京爱奇艺科技有限公司 A kind of face dividing method, device and equipment
CN109801249A (en) * 2018-12-27 2019-05-24 深圳豪客互联网有限公司 Image interfusion method, device, computer equipment and storage medium
CN109829374A (en) * 2018-12-27 2019-05-31 深圳豪客互联网有限公司 Image processing method, device, computer equipment and storage medium
CN111382746A (en) * 2018-12-29 2020-07-07 杭州光启人工智能研究院 Data marking method, computer device and computer readable storage medium
CN109903257A (en) * 2019-03-08 2019-06-18 上海大学 A kind of virtual hair-dyeing method based on image, semantic segmentation
CN111723596B (en) * 2019-03-18 2024-03-22 北京市商汤科技开发有限公司 Gaze area detection and neural network training method, device and equipment
CN111723596A (en) * 2019-03-18 2020-09-29 北京市商汤科技开发有限公司 Method, device and equipment for detecting gazing area and training neural network
CN110148085A (en) * 2019-04-22 2019-08-20 智慧眼科技股份有限公司 Face image super-resolution reconstruction method and computer-readable storage medium
CN110059768B (en) * 2019-04-30 2022-11-15 福州大学 Semantic segmentation method and system for fusion point and region feature for street view understanding
CN110059768A (en) * 2019-04-30 2019-07-26 福州大学 The semantic segmentation method and system of the merging point and provincial characteristics that understand for streetscape
CN110287930A (en) * 2019-07-01 2019-09-27 厦门美图之家科技有限公司 Wrinkle disaggregated model training method and device
CN110287930B (en) * 2019-07-01 2021-08-20 厦门美图之家科技有限公司 Wrinkle classification model training method and device
CN110378278B (en) * 2019-07-16 2021-11-02 北京地平线机器人技术研发有限公司 Neural network training method, object searching method, device and electronic equipment
CN110378278A (en) * 2019-07-16 2019-10-25 北京地平线机器人技术研发有限公司 Training method, object search method, apparatus and the electronic equipment of neural network
CN110472605A (en) * 2019-08-21 2019-11-19 广州纳丽生物科技有限公司 A kind of skin problem diagnostic method based on deep learning face subregion
CN110472605B (en) * 2019-08-21 2022-10-14 广州纳丽生物科技有限公司 Skin problem classification method based on deep learning face partition
CN112669197A (en) * 2019-10-16 2021-04-16 顺丰科技有限公司 Image processing method, image processing device, mobile terminal and storage medium
CN111144310A (en) * 2019-12-27 2020-05-12 创新奇智(青岛)科技有限公司 Face detection method and system based on multi-layer information fusion
CN111179287A (en) * 2020-01-03 2020-05-19 厦门美图之家科技有限公司 Portrait instance segmentation method, device, equipment and storage medium
CN111310718A (en) * 2020-03-09 2020-06-19 成都川大科鸿新技术研究所 High-accuracy detection and comparison method for face-shielding image
CN111739025B (en) * 2020-05-08 2024-03-19 北京迈格威科技有限公司 Image processing method, device, terminal and storage medium
CN111739025A (en) * 2020-05-08 2020-10-02 北京迈格威科技有限公司 Image processing method, device, terminal and storage medium
CN111666905A (en) * 2020-06-10 2020-09-15 重庆紫光华山智安科技有限公司 Model training method, pedestrian attribute identification method and related device
CN112330696B (en) * 2020-12-02 2022-08-09 青岛大学 Face segmentation method, face segmentation device and computer-readable storage medium
CN112330696A (en) * 2020-12-02 2021-02-05 青岛大学 Face segmentation method, face segmentation device and computer-readable storage medium
CN112560701A (en) * 2020-12-17 2021-03-26 成都新潮传媒集团有限公司 Face image extraction method and device and computer storage medium
CN112733632A (en) * 2020-12-28 2021-04-30 华南理工大学 Robot control method based on face recognition and gesture recognition
CN112733632B (en) * 2020-12-28 2023-02-14 华南理工大学 Robot control method based on face recognition and gesture recognition
CN113052247A (en) * 2021-03-31 2021-06-29 清华苏州环境创新研究院 Garbage classification method and garbage classifier based on multi-label image recognition
CN113469040A (en) * 2021-06-30 2021-10-01 北京市商汤科技开发有限公司 Image processing method and device, computer equipment and storage medium
CN113469040B (en) * 2021-06-30 2023-10-24 北京市商汤科技开发有限公司 Image processing method, device, computer equipment and storage medium
WO2023273022A1 (en) * 2021-06-30 2023-01-05 北京市商汤科技开发有限公司 Image processing method and apparatus, computer device, storage medium, and computer program product
CN116269285A (en) * 2022-11-28 2023-06-23 电子科技大学 Non-contact normalized heart rate variability estimation system
CN116269285B (en) * 2022-11-28 2024-05-28 电子科技大学 Non-contact normalized heart rate variability estimation system

Also Published As

Publication number Publication date
CN108985181B (en) 2020-07-24

Similar Documents

Publication Publication Date Title
CN108985181A (en) A kind of end-to-end face mask method based on detection segmentation
CN107273876B (en) A kind of micro- expression automatic identifying method of ' the macro micro- transformation model of to ' based on deep learning
CN110163239B (en) Weak supervision image semantic segmentation method based on super-pixel and conditional random field
CN104143079B (en) The method and system of face character identification
CN102496023B (en) Region of interest extraction method of pixel level
CN110428432A (en) The deep neural network algorithm of colon body of gland Image Automatic Segmentation
CN110008832A (en) Based on deep learning character image automatic division method, information data processing terminal
CN108288035A (en) The human motion recognition method of multichannel image Fusion Features based on deep learning
CN109712145A (en) A kind of image matting method and system
CN106295584A (en) Depth migration study is in the recognition methods of crowd's attribute
CN111462183A (en) Behavior identification method and system based on attention mechanism double-current network
CN109598234A (en) Critical point detection method and apparatus
CN110472495B (en) Deep learning face recognition method based on graphic reasoning global features
CN110827312B (en) Learning method based on cooperative visual attention neural network
CN106339719A (en) Image identification method and image identification device
CN110223304A (en) A kind of image partition method, device and computer readable storage medium based on multipath polymerization
CN109214298A (en) A kind of Asia women face value Rating Model method based on depth convolutional network
Mo et al. Background noise filtering and distribution dividing for crowd counting
CN109886391A (en) A kind of neural network compression method based on the positive and negative diagonal convolution in space
CN106067016A (en) A kind of facial image eyeglass detection method and device
CN109978074A (en) Image aesthetic feeling and emotion joint classification method and system based on depth multi-task learning
Yuan et al. Watershed-based superpixels with global and local boundary marching
CN108416304A (en) A kind of three classification method for detecting human face using contextual information
AU2021240205B1 (en) Object sequence recognition method, network training method, apparatuses, device, and medium
Liang et al. Image segmentation and recognition for multi-class chinese food

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20200724

Termination date: 20210622