CN108921037A - A kind of Emotion identification method based on BN-inception binary-flow network - Google Patents

A kind of Emotion identification method based on BN-inception binary-flow network Download PDF

Info

Publication number
CN108921037A
CN108921037A CN201810579049.5A CN201810579049A CN108921037A CN 108921037 A CN108921037 A CN 108921037A CN 201810579049 A CN201810579049 A CN 201810579049A CN 108921037 A CN108921037 A CN 108921037A
Authority
CN
China
Prior art keywords
network
inception
binary
flow
spp
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201810579049.5A
Other languages
Chinese (zh)
Other versions
CN108921037B (en
Inventor
卿粼波
王露
滕奇志
何小海
熊文诗
吴晓红
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sichuan University
Original Assignee
Sichuan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sichuan University filed Critical Sichuan University
Priority to CN201810579049.5A priority Critical patent/CN108921037B/en
Publication of CN108921037A publication Critical patent/CN108921037A/en
Application granted granted Critical
Publication of CN108921037B publication Critical patent/CN108921037B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The individual Emotion identification method based on posture information that the present invention provides a kind of relates generally to the mood that individual is judged using deep learning technique study individual posture.This method includes:It is firstly introduced into and the static state and behavioral characteristics of list entries is extracted by the study to original image and light stream image based on BN-inception binary-flow network model;Then spatial pyramid pond (Space Pyramid Pooling, SPP) is added on the basis of binary-flow network, image is made to input network with original size, to reduce because being influenced caused by deforming model performance.The invention firstly uses binary-flow networks to carry out the study of space-time characteristic to list entries, and introduces pyramid pond, to retain the raw information of video frame, enables network effectively to learn the feature of individual posture mood, obtains higher discrimination.

Description

A kind of Emotion identification method based on BN-inception binary-flow network
Technical field
The present invention relates to the Emotion identification problems in deep learning field, are based on BN-inception more particularly, to one kind The individual mood analysis method of+SPP binary-flow network.
Background technique
Mood is the feeling for combining people, a kind of state of thought and act, is played in the exchange of person to person important Effect.Its emotional state can be judged generally according to people's facial expression, but in certain specific environment, such as monitoring view In the presence of when blocking, we can not necessarily obtain clearly face facial expression for angle, face.In fact, people True emotional not only relies on human face expression to be expressed, and individual limb action can also express certain emotional information.Therefore, Research of the invention is concentrated mainly on the Emotion identification of the individual posture based on video.
Emotion identification is important research content and direction in computer vision field, the International Periodicals of current many authoritys It is equipped with relevant theme and content with top-level meeting, and many external elite schools have also opened up relevant course.It is traditional based on view The Emotion identification method of frequency relies primarily on the feature manually chosen, this method take time and effort and the model parameter that obtains it is extensive Performance is poor, serves the limitation of Emotion identification.Deep learning is an important component of artificial intelligence field development, The very popular research direction of artificial intelligence field is had become in recent years.It is in many fields (such as image recognition, speech recognition Deng) very big breakthrough is all achieved, higher discrimination and generalization ability have especially been achieved in video analysis.Therefore originally Patent utilizes advantage of the deep learning in video analysis, studies Emotion identification individual in video.
Just grown up based on the Emotion identification of posture information in recent years, relevant research is less, focuses primarily upon biography The research of system algorithm.Li et al.[1]It proposes and a kind of raw skeleton coordinate and skeleton motion is utilized to carry out Activity recognition and classification; Piana etc.[2]It is proposed a mood automatic identification model and system from all-around exercises, it is used to help autism children Association identifies and is showed emotion by all-around exercises.Equally also someone by the motion feature of human body attitude and advanced kinematics Geometrical characteristic is combined, and is carried out cluster and is classified.Crenn etc.[3]It is low that operation data etc. is obtained using the 3D frame sequence of people Then feature decomposition is geometrical characteristic, motion feature and three kinds of Fourier's feature, calculates the member of these low-level features by layer feature Feature (such as mean value, standard deviation) finally uses classifier by first tagsort.Deep learning regardless of recognition time still It in accuracy, is compared with the traditional method and all improves a lot, but due to the shortage of mood data collection relevant to posture, adopt It is also seldom that the individual Emotion identification correlative study based on posture information is carried out with deep learning.
Summary of the invention
The individual Emotion identification method based on posture that the object of the present invention is to provide a kind of, will be in deep learning and video Human body attitude combines, and makes full use of the superiority of BN-inception+SPP network structure, while introducing binary-flow network structure The individual Emotion identification based on video is carried out, effectively learns the emotional characteristics of individual posture, obtains higher discrimination.
For convenience of explanation, it is firstly introduced into following concept:
Optical flow method:It is a kind of expression way of simple and practical image motion, is normally defined in an image sequence The apparent motion of brightness of image mode, i.e., the movement velocity of the point on space object surface is on the imaging plane of visual sensor Expression.
Convolutional neural networks:A kind of multilayer feedforward neural network, every layer is made of multiple two-dimensional surfaces, the mind of each plane It works independently through member, convolutional neural networks include convolutional layer and pond layer.
Double-current convolutional neural networks:For video behavioural characteristic extraction and design, network with single frames RGB original image and The light stream image obtained based on video data is inputted respectively as two, is indicated with realizing the apparent information in object of action space And the extraction of action process temporal aspect.
Spatial pyramid pond (Space Pyramid Pooling, SPP):It is to be composed of multiple down-sampled layers, it The fixed feature vector of a length can be converted to dividing to input feature vector figure from thick to thin, and characteristic pattern, So SPP layers can extract various local messages.
The present invention specifically adopts the following technical scheme that:
Propose the individual Emotion identification method based on BN-inception+SPP binary-flow network, the main feature of this method It is:
1. individual attitude data collection is divided into four mood classifications:Boring (bored), exciting (excited) is angry (frantic), loosen (relaxed);
2. spatial pyramid pond (Space is added before the full articulamentum of BN-inception binary-flow network Pyramid Pooling, SPP), carry out the training of time-space network respectively to data set;
This method mainly includes the following steps that:
(1) by individual posture sequence data collection, four mood classifications are divided into:Boring, excitement is angry, loosens;
(2) the corresponding light stream image sequence of data set is generated using the optical flow algorithm of document [4], indicates the fortune of individual posture Dynamic feature;
(3) raw data set and optical flow data collection are proportionally divided into training set, verifying collection and test set respectively;
(4) the double-current convolutional neural networks model based on BN-inception is introduced, and is added before its full articulamentum SPP layers of optimization BN-inception network, using training set and verifying collection carry out time-space network training, using test set into Row verifying;
(5) by based on BN-inception+SPP spatial flow and two channel network of time flow carry out average fusion, obtain Accuracy ACC (Accuracy) and macro accuracy of the mean MAP (Macro Average Precision) on test set.
Detailed description of the invention
Fig. 1 is that the present invention is based on the individual Emotion identification general frame schematic diagrames of BN-inception+SPP binary-flow network.
Fig. 2-a~Fig. 2-b is the accuracy confusion matrix that the present invention obtains on test set when SPP layers are not added, In, 2-a is the test matrix of spatial flow BN-inception network, and 2-b is the test square of time flow BN-inception network Battle array.
Fig. 3-a~Fig. 3-b is the accuracy confusion matrix that the present invention obtains on test set when being added SPP layers, wherein 3-a is the test matrix of spatial flow BN-inception+SPP network, and 3-b is time flow BN-inception+SPP network Test matrix.
Fig. 4 is that the spatial flow by based on BN-inception+SPP and two channel network of time flow of the invention averagely merge Afterwards, the ACC and MAP on test set are obtained.
Specific embodiment
The present invention is described in further detail with reference to the accompanying drawings and embodiments, it is necessary to, it is noted that below Embodiment is served only for that the present invention is described further, should not be understood as limiting the scope of the invention, fields Personnel be skillful at according to foregoing invention content, some nonessential modifications and adaptations are made to the present invention and are embodied, Protection scope of the present invention should be still fallen within.
In Fig. 1, a kind of individual Emotion identification method based on BN-inception+SPP binary-flow network, including following step Suddenly:
(1) it first after obtaining the individual data items collection in public space, is generated using the optical flow algorithm of document [4] former The light stream image sequence of beginning data set indicates the motion feature of individual posture;
(2) raw data set and obtained optical flow data collection are respectively divided into test set in proportion, verifying collects and training Collect three parts, and gives corresponding mood classification;
(3) remove SPP layers shown in FIG. 1, the data of training set and verifying collection inputted time-space network respectively and learn, Training pattern is obtained, test is carried out using the data of test set and verifies its effect;
(4) SPP layers are added, training set is inputted into time-space network with original size respectively and is learnt, training pattern is obtained, Test, which is carried out, using the data of test set collection verifies its effect;
(5) it by the spatial flow based on BN-inception+SPP and after two channel network of time flow averagely merges, is surveyed ACC and MAP on examination collection;
The present invention is separately trained the convolutional neural networks in two channels of spatial flow and time flow using Caffe, leads to It crosses experiment the parameter of time flow and space flow network is arranged, as shown in table 1.Due to the individual posture mood number of foundation It is less according to the sample size of collection, to prevent over-fitting, Dropout layers of side is added using data extending and in a network Method.
The setting of 1 training parameter of table
Bibliography:
[1]Li C,Zhong Q,Xie D,et al.Skeleton-based Action Recognition with Convolutional Neural Networks[J]. 2017:597-600.
[2]Piana S,StaglianòA,Odone F,et al.Adaptive Body Gesture Representation for Automatic Emotion Recognition[J].ACM Transactions on Interactive Intelligent Systems(TiiS),2016,6(1):6.
[3]Crenn A,Khan R A,Meyer A,et al.Body Expression Recognition from Animated 3D Skeleton[C]// International Conference on 3D Imaging.IEEE,2017:1- 7.
[4]Brox T,Bruhn A,Papenberg N,et al.High Accuracy Optical Flow Estimation Based on A Theory for Warping[C]//European Conference on Computer Vision(ECCV),2004:25-36.

Claims (3)

1. a kind of individual Emotion identification method based on BN-inception+SPP binary-flow network, it is characterised in that:
A. individual attitude data collection is divided into four mood classifications:Boring (bored), exciting (excited) is angry (frantic), loosen (relaxed);
B. spatial pyramid pond (Space Pyramid is added before the full articulamentum of BN-inception binary-flow network Pooling, SPP), carry out the training of time-space network respectively to data set;
This method mainly includes the following steps that:
(1) using the corresponding light stream image sequence of optical flow algorithm processing data set generation of document [1], the fortune of individual posture is indicated Dynamic feature;
(2) data set is divided into training set, verifying collection and test set three parts, and gives the mood classification of each sequence;
(3) the double-current convolutional neural networks model based on BN-inception is introduced, and SPP layers are added before its full articulamentum Optimize BN-inception network, the training of time-space network is carried out using training set and verifying collection, is verified using test set;
(4) by based on BN-inception+SPP spatial flow and two channel network of time flow carry out average fusion, tested Accuracy ACC (Accuracy) and macro accuracy of the mean MAP (Macro Average Precision) on collection.
2. the individual Emotion identification method based on BN-inception+SPP binary-flow network as described in claim 1, feature It is respectively to learn the space-time characteristic of data set using binary-flow network in step (3).
3. as described in claim 1 based on the Emotion identification method of BN-inception binary-flow network, it is characterised in that in step Suddenly SPP layers are added in (3) first before the full articulamentum of BN-inception binary-flow network, so that training set is with original size Network is inputted, avoids fixed input size bring motion information from losing, time-space network is then carried out respectively to data set again Training.
CN201810579049.5A 2018-06-07 2018-06-07 Emotion recognition method based on BN-acceptance double-flow network Active CN108921037B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810579049.5A CN108921037B (en) 2018-06-07 2018-06-07 Emotion recognition method based on BN-acceptance double-flow network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810579049.5A CN108921037B (en) 2018-06-07 2018-06-07 Emotion recognition method based on BN-acceptance double-flow network

Publications (2)

Publication Number Publication Date
CN108921037A true CN108921037A (en) 2018-11-30
CN108921037B CN108921037B (en) 2022-06-03

Family

ID=64418934

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810579049.5A Active CN108921037B (en) 2018-06-07 2018-06-07 Emotion recognition method based on BN-acceptance double-flow network

Country Status (1)

Country Link
CN (1) CN108921037B (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109766856A (en) * 2019-01-16 2019-05-17 华南农业大学 A kind of method of double fluid RGB-D Faster R-CNN identification milking sow posture
CN109814565A (en) * 2019-01-30 2019-05-28 上海海事大学 The unmanned boat intelligence navigation control method of space-time double fluid data-driven depth Q study
CN109886160A (en) * 2019-01-30 2019-06-14 浙江工商大学 It is a kind of it is non-limiting under the conditions of face identification method
CN110147729A (en) * 2019-04-16 2019-08-20 深圳壹账通智能科技有限公司 User emotion recognition methods, device, computer equipment and storage medium
CN110175596A (en) * 2019-06-04 2019-08-27 重庆邮电大学 The micro- Expression Recognition of collaborative virtual learning environment and exchange method based on double-current convolutional neural networks
CN110414561A (en) * 2019-06-26 2019-11-05 武汉大学 A kind of construction method of the natural scene data set suitable for machine vision
CN112131908A (en) * 2019-06-24 2020-12-25 北京眼神智能科技有限公司 Action identification method and device based on double-flow network, storage medium and equipment
WO2022037642A1 (en) * 2020-08-19 2022-02-24 南京图格医疗科技有限公司 Method for detecting and classifying lesion area in clinical image

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050265580A1 (en) * 2004-05-27 2005-12-01 Paul Antonucci System and method for a motion visualizer
CN102663429A (en) * 2012-04-11 2012-09-12 上海交通大学 Method for motion pattern classification and action recognition of moving target
CN103544963A (en) * 2013-11-07 2014-01-29 东南大学 Voice emotion recognition method based on core semi-supervised discrimination and analysis
CN104732203A (en) * 2015-03-05 2015-06-24 中国科学院软件研究所 Emotion recognizing and tracking method based on video information
CN106295568A (en) * 2016-08-11 2017-01-04 上海电力学院 The mankind's naturalness emotion identification method combined based on expression and behavior bimodal
CN106897671A (en) * 2017-01-19 2017-06-27 山东中磁视讯股份有限公司 A kind of micro- expression recognition method encoded based on light stream and FisherVector
CN107368798A (en) * 2017-07-07 2017-11-21 四川大学 A kind of crowd's Emotion identification method based on deep learning
CN107491731A (en) * 2017-07-17 2017-12-19 南京航空航天大学 A kind of Ground moving target detection and recognition methods towards precision strike
CN107784114A (en) * 2017-11-09 2018-03-09 广东欧珀移动通信有限公司 Recommendation method, apparatus, terminal and the storage medium of facial expression image
CN107944442A (en) * 2017-11-09 2018-04-20 北京智芯原动科技有限公司 Based on the object test equipment and method for improving convolutional neural networks

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050265580A1 (en) * 2004-05-27 2005-12-01 Paul Antonucci System and method for a motion visualizer
CN102663429A (en) * 2012-04-11 2012-09-12 上海交通大学 Method for motion pattern classification and action recognition of moving target
CN103544963A (en) * 2013-11-07 2014-01-29 东南大学 Voice emotion recognition method based on core semi-supervised discrimination and analysis
CN104732203A (en) * 2015-03-05 2015-06-24 中国科学院软件研究所 Emotion recognizing and tracking method based on video information
CN106295568A (en) * 2016-08-11 2017-01-04 上海电力学院 The mankind's naturalness emotion identification method combined based on expression and behavior bimodal
CN106897671A (en) * 2017-01-19 2017-06-27 山东中磁视讯股份有限公司 A kind of micro- expression recognition method encoded based on light stream and FisherVector
CN107368798A (en) * 2017-07-07 2017-11-21 四川大学 A kind of crowd's Emotion identification method based on deep learning
CN107491731A (en) * 2017-07-17 2017-12-19 南京航空航天大学 A kind of Ground moving target detection and recognition methods towards precision strike
CN107784114A (en) * 2017-11-09 2018-03-09 广东欧珀移动通信有限公司 Recommendation method, apparatus, terminal and the storage medium of facial expression image
CN107944442A (en) * 2017-11-09 2018-04-20 北京智芯原动科技有限公司 Based on the object test equipment and method for improving convolutional neural networks

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
KAIMING HE ET AL: "Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition", 《IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE》 *
陈胜娣 等: "基于改进的深度卷积神经网络的人体动作识别方法", 《计算机应用研究》 *

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109766856A (en) * 2019-01-16 2019-05-17 华南农业大学 A kind of method of double fluid RGB-D Faster R-CNN identification milking sow posture
CN109766856B (en) * 2019-01-16 2022-11-15 华南农业大学 Method for recognizing postures of lactating sows through double-current RGB-D Faster R-CNN
CN109814565A (en) * 2019-01-30 2019-05-28 上海海事大学 The unmanned boat intelligence navigation control method of space-time double fluid data-driven depth Q study
CN109886160A (en) * 2019-01-30 2019-06-14 浙江工商大学 It is a kind of it is non-limiting under the conditions of face identification method
CN110147729A (en) * 2019-04-16 2019-08-20 深圳壹账通智能科技有限公司 User emotion recognition methods, device, computer equipment and storage medium
CN110175596A (en) * 2019-06-04 2019-08-27 重庆邮电大学 The micro- Expression Recognition of collaborative virtual learning environment and exchange method based on double-current convolutional neural networks
CN112131908A (en) * 2019-06-24 2020-12-25 北京眼神智能科技有限公司 Action identification method and device based on double-flow network, storage medium and equipment
CN112131908B (en) * 2019-06-24 2024-06-11 北京眼神智能科技有限公司 Action recognition method, device, storage medium and equipment based on double-flow network
CN110414561A (en) * 2019-06-26 2019-11-05 武汉大学 A kind of construction method of the natural scene data set suitable for machine vision
WO2022037642A1 (en) * 2020-08-19 2022-02-24 南京图格医疗科技有限公司 Method for detecting and classifying lesion area in clinical image

Also Published As

Publication number Publication date
CN108921037B (en) 2022-06-03

Similar Documents

Publication Publication Date Title
CN108921037A (en) A kind of Emotion identification method based on BN-inception binary-flow network
CN107292813B (en) A kind of multi-pose Face generation method based on generation confrontation network
CN105005774B (en) A kind of recognition methods of face kinship and device based on convolutional neural networks
CN107423678A (en) A kind of training method and face identification method of the convolutional neural networks for extracting feature
CN108830252A (en) A kind of convolutional neural networks human motion recognition method of amalgamation of global space-time characteristic
CN109919031A (en) A kind of Human bodys' response method based on deep neural network
CN111274921B (en) Method for recognizing human body behaviors by using gesture mask
CN109508669A (en) A kind of facial expression recognizing method based on production confrontation network
CN109101865A (en) A kind of recognition methods again of the pedestrian based on deep learning
CN106570474A (en) Micro expression recognition method based on 3D convolution neural network
CN108052884A (en) A kind of gesture identification method based on improvement residual error neutral net
CN105160310A (en) 3D (three-dimensional) convolutional neural network based human body behavior recognition method
CN107092894A (en) A kind of motor behavior recognition methods based on LSTM models
CN109376720A (en) Classification of motion method based on artis space-time simple cycle network and attention mechanism
CN107392131A (en) A kind of action identification method based on skeleton nodal distance
CN110232361B (en) Human behavior intention identification method and system based on three-dimensional residual dense network
CN104063721B (en) A kind of human behavior recognition methods learnt automatically based on semantic feature with screening
CN108537181A (en) A kind of gait recognition method based on the study of big spacing depth measure
CN109858407A (en) A kind of video behavior recognition methods based on much information stream feature and asynchronous fusion
CN110119707A (en) A kind of human motion recognition method
CN114821640A (en) Skeleton action identification method based on multi-stream multi-scale expansion space-time diagram convolution network
CN113111857A (en) Human body posture estimation method based on multi-mode information fusion
CN110163567A (en) Classroom roll calling system based on multitask concatenated convolutional neural network
CN110046544A (en) Digital gesture identification method based on convolutional neural networks
CN110059593A (en) A kind of human facial expression recognition method based on feedback convolutional neural networks

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant