CN111329498A - Multi-modal driver emotion auxiliary adjusting method - Google Patents

Multi-modal driver emotion auxiliary adjusting method Download PDF

Info

Publication number
CN111329498A
CN111329498A CN202010157896.XA CN202010157896A CN111329498A CN 111329498 A CN111329498 A CN 111329498A CN 202010157896 A CN202010157896 A CN 202010157896A CN 111329498 A CN111329498 A CN 111329498A
Authority
CN
China
Prior art keywords
driver
road rage
information
emotion
signal
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010157896.XA
Other languages
Chinese (zh)
Inventor
李翠霞
李英豪
王亚博
吴卫东
周元元
杨珊珊
闫凯波
许书宁
叶帅
赵若琰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhengzhou University
Original Assignee
Zhengzhou University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhengzhou University filed Critical Zhengzhou University
Priority to CN202010157896.XA priority Critical patent/CN111329498A/en
Publication of CN111329498A publication Critical patent/CN111329498A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/74Details of notification to user or communication with user or patient ; user input means
    • A61B5/7405Details of notification to user or communication with user or patient ; user input means using sound
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/0059Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
    • A61B5/0077Devices for viewing the surface of the body, e.g. camera, magnifying lens
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/16Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
    • A61B5/18Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state for vehicle drivers or machine operators
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/48Other medical applications
    • A61B5/4803Speech analysis specially adapted for diagnostic purposes
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/74Details of notification to user or communication with user or patient ; user input means
    • A61B5/746Alarms related to a physiological condition, e.g. details of setting alarm thresholds or avoiding false alarms

Landscapes

  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Veterinary Medicine (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Medical Informatics (AREA)
  • Molecular Biology (AREA)
  • Surgery (AREA)
  • Animal Behavior & Ethology (AREA)
  • Physics & Mathematics (AREA)
  • Public Health (AREA)
  • Pathology (AREA)
  • Child & Adolescent Psychology (AREA)
  • Developmental Disabilities (AREA)
  • Educational Technology (AREA)
  • Hospice & Palliative Care (AREA)
  • Psychiatry (AREA)
  • Psychology (AREA)
  • Social Psychology (AREA)
  • Physiology (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Traffic Control Systems (AREA)

Abstract

The invention discloses a multi-modal driver emotion auxiliary adjusting method, which relates to the field of artificial intelligence and machine learning, and comprises the steps of collecting multi-modal data information of a driver; according to the multi-modal data information, the emotional states of the drivers in the single modes are respectively identified, and whether the emotional state of the driver in each single mode is a road rage signal or not is judged; fusing judgment results of emotional states of a plurality of drivers in single mode, judging whether the fusion result is a total road rage signal, and grading the total road rage signal; and correspondingly outputting different voice reminding information according to the grade of the total road rage signal, and adjusting the emotion of the driver. The invention can monitor the emotion of the driver in real time, and can adopt related measures to assist in adjusting the emotion of the driver when the emotional state of the driver is abnormal, thereby reducing the occurrence of accidents and protecting the driving safely.

Description

Multi-modal driver emotion auxiliary adjusting method
Technical Field
The invention relates to the field of artificial intelligence and machine learning, in particular to a multi-modal driver emotion auxiliary adjusting method.
Background
The concept of road rage (Roadrage) was first derived from foreign psychology. The medical community classifies the road rage as paroxysmal rage disorder, which means that multiple anger fires explode, and the violent degree is called as a big feeling and an accident. For example, when a driver encounters traffic jam or other things affecting mood, the driver can show behaviors of operation modes such as anger, frequent whistling, frequent braking and the like.
Traffic accidents caused by road rage are rising every year, and the public traffic safety is seriously influenced. Although some driving assistance systems are already available in the market, the influence of the emotion of the driver on the driving safety is not considered, and an appropriate adjusting strategy is not matched according to the emotion recognition result.
Disclosure of Invention
In order to solve the problems in the background technology, the invention provides a multi-modal driver emotion auxiliary adjusting method, which collects data related to emotion judgment of a driver through various channels, performs emotion analysis and judgment through a deep learning method, and further adopts related measures to assist in adjusting the emotion of the driver, so that accidents are reduced.
The technical scheme of the invention is as follows:
a multi-modal driver emotion aid adjustment method comprises the following steps:
s1, collecting multi-modal data information of the driver; the multi-modal data information comprises image information, voice information and driving behavior information;
s2, according to the multi-modal data information, respectively identifying the emotional states of the drivers in a plurality of single modes, and judging whether the emotional state of each driver in the single mode is a road rage signal;
s3, fusing judgment results of emotional states of the drivers in the single modes, judging whether the fusion result is a general road rage signal, and grading the general road rage signal;
and S4, correspondingly outputting different voice reminding information according to the grade of the total road rage signal, and adjusting the emotion of the driver.
Preferably, the S2 includes image information identification, specifically:
the method comprises the steps that driver image information captured by a camera is disassembled into frames, the frames are respectively sent to an OpenCv face selector, and face images of drivers in a vehicle are cut out in real time;
and (3) converting the size of the facial image into 224 × 224 pixels, sending the facial image into a VGG16 deep convolution neural network model to extract features, and dividing the facial image into a road anger signal and a non-road anger signal.
Preferably, the S2 includes speech information recognition, specifically:
capturing the voice information of a driver in the vehicle in real time by adopting a VAD algorithm, and sending the voice information of the driver into an ASR model to convert the voice into character information;
judging whether the voice information of the driver contains a wake-up word or not; if the wake-up word is contained, sending the wake-up word into a voice interaction module to carry out human-computer interaction normally; if the awakening words are not contained, emotion analysis is carried out, and the voice information of the driver is divided into a road rage signal and a non-road rage signal.
Preferably, the S2 includes driving behavior information identification, specifically:
collecting driving behavior data of driver by film sensor mounted on steering wheel and brake
The collected driving behavior data of the driver are sent to a random forest model, and the driving behavior data of the driver are divided into a road rage signal and a non-road rage signal through the random forest model;
the driving behavior data of the driver comprise average starting hot-car time, on-site idling ratio, rapid accelerator pedaling frequency and rapid braking frequency of the driver within three days.
Preferably, in S3, if the emotional state of the driver in the single mode is a road rage signal, the emotional state is marked as "1"; the emotional state of the driver in each single mode is a road rage signal and is marked as '0'; the total road rage signal is graded, and the grading comprises the following steps:
when the emotional states of the drivers in the single modes are added to be 0, the driver is not angry;
when the emotional states of the drivers in the single modes are added to be 1, the driver is in mild road rage;
when the emotional states of the drivers in the single modes are added to be 2, the driver is moderate road rage;
and when the emotional states of the drivers in the single modes are added to be 3, the road is heavily irritated.
Compared with the prior art, the invention has the beneficial effects that: the invention can monitor the emotion of the driver in real time according to the image information, the voice information and the driving behavior information of the driver, and can adopt related voice reminding measures to assist in adjusting the emotion of the driver when the emotion state of the driver is abnormal, so that the occurrence of accidents is reduced, and the driving safety is protected.
Drawings
FIG. 1 is a flow chart of the present invention;
fig. 2 is a working principle diagram of the present invention.
Detailed Description
The technical solutions of the embodiments of the present invention are clearly and completely described below with reference to the drawings in the present invention, and it is obvious that the described embodiments are some embodiments of the present invention, but not all embodiments. All other embodiments, which can be obtained by a person skilled in the art without any inventive step based on the embodiments of the present invention, shall fall within the scope of protection of the present invention.
As shown in fig. 1, the invention provides a multimodal driver emotion auxiliary adjusting method, which includes:
s1, collecting multi-modal data information of the driver; the multi-modal data information comprises image information, voice information and driving behavior information;
s2, according to the multi-modal data information, respectively identifying the emotional states of the drivers in a plurality of single modes, and judging whether the emotional state of each driver in the single mode is a road rage signal;
s3, fusing judgment results of emotional states of the drivers in the single modes, judging whether the fusion result is a general road rage signal, and grading the general road rage signal;
and S4, correspondingly outputting different voice reminding information according to the grade of the total road rage signal, and adjusting the emotion of the driver.
Specifically, as shown in fig. 2, the emotional state data of the driver in each of the plurality of single modalities includes image information, voice information, and driving behavior information, respectively.
Image information identification, specifically: the method comprises the steps that driver image information captured by a camera is disassembled into frames, the frames are respectively sent to an OpenCv face selector, and face images of drivers in a vehicle are cut out in real time; and (3) converting the size of the facial image into 224 × 224 pixels, sending the facial image into a VGG16 deep convolution neural network model to extract features, and dividing the facial image into a road anger signal and a non-road anger signal.
The voice information recognition specifically comprises the following steps: capturing the voice information of a driver in the vehicle in real time by adopting a VAD algorithm, and sending the voice information of the driver into an ASR model to convert the voice into character information; judging whether the voice information of the driver contains a wake-up word or not; if the awakening words are contained, the voice interaction module is sent to carry out human-computer interaction normally. Such as voice short message sending, voice weather and road condition information inquiry, voice playing, song cutting and the like.
If the awakening words are not contained, emotion analysis is carried out, and the voice information of the driver is divided into a road rage signal and a non-road rage signal.
The driving behavior information identification specifically comprises the following steps: collecting driving behavior data of a driver through a film sensor arranged at a steering wheel and a brake; the collected driving behavior data of the driver are sent to a random forest model, and the driving behavior data of the driver are divided into a road rage signal and a non-road rage signal through the random forest model; the driving behavior data of the driver comprise average starting hot-car time, on-site idling ratio, rapid accelerator pedaling frequency and rapid braking frequency of the driver within three days.
Wherein, the emotional state of the driver in each single mode is a road rage signal, and is marked as '1'; the emotional state of the driver in each single mode is a road rage signal and is marked as '0'; the total road rage signal is graded, and the grading comprises the following steps:
(1) when the emotional states of the drivers in the single modes are added to be 0, the driver is not angry, namely the emotional state of the driver is 0, and the driver is not angry, the voice prompt is not carried out;
(2) when the emotional states of the drivers in the single modes are added to be 1, the driver is in mild road rage, namely the emotional state of the driver is 1, the level 1 prompt is activated, and meanwhile voice reminding information can be played, for example, whether you are not happy and a relaxing music bar is played. The soothing music in the song library may be collected by the driver in advance according to personal preferences or selected from selected songs.
(3) When the emotional states of the drivers in the single modes are added to be 2, the driver is moderate road rage, namely the emotional state of the driver is 2, 2-level prompt is activated, and meanwhile voice reminding information can be played, for example, "you are suspected of road rage now, mom wants to speak to you". And calling the voice recorded by the mother in advance from the voice library to prompt the driver.
(4) When the emotional states of the drivers in the single modes are added to be 3, the driver is severe road rage, namely the emotional state of the driver is 3, 3-level prompt is activated, and meanwhile voice reminding information can be played, for example, "dangerous now, your baby comes". And calling out the sound recorded by the child in advance from the voice library to prompt the driver.
Wherein, the level 2 and the level 3 can also be selected and set by the driver according to the importance degree in the mind.
The above disclosure is only for the preferred embodiments of the present invention, but the embodiments of the present invention are not limited thereto, and any variations that can be made by those skilled in the art are intended to fall within the scope of the present invention.

Claims (5)

1. A multi-modal driver emotion aid adjustment method is characterized by comprising the following steps:
s1, collecting multi-modal data information of the driver; the multi-modal data information comprises image information, voice information and driving behavior information;
s2, according to the multi-modal data information, respectively identifying the emotional states of the drivers in a plurality of single modes, and judging whether the emotional state of each driver in the single mode is a road rage signal;
s3, fusing judgment results of emotional states of the drivers in the single modes, judging whether the fusion result is a general road rage signal, and grading the general road rage signal;
and S4, correspondingly outputting different voice reminding information according to the grade of the total road rage signal, and adjusting the emotion of the driver.
2. The multi-modal driver emotion aid adjustment method as claimed in claim 1, wherein said S2 includes image information recognition, specifically:
the method comprises the steps that driver image information captured by a camera is disassembled into frames, the frames are respectively sent to an OpenCv face selector, and face images of drivers in a vehicle are cut out in real time;
and (3) converting the size of the facial image into 224 × 224 pixels, sending the facial image into a VGG16 deep convolution neural network model to extract features, and dividing the facial image into a road anger signal and a non-road anger signal.
3. The multi-modal driver emotion aid adjustment method as claimed in claim 1, wherein said S2 includes speech information recognition, specifically:
capturing the voice information of a driver in the vehicle in real time by adopting a VAD algorithm, and sending the voice information of the driver into an ASR model to convert the voice into character information;
judging whether the voice information of the driver contains a wake-up word or not; if the wake-up word is contained, sending the wake-up word into a voice interaction module to carry out human-computer interaction normally; if the awakening words are not contained, emotion analysis is carried out, and the voice information of the driver is divided into a road rage signal and a non-road rage signal.
4. The multi-modal driver emotion aid adjustment method as claimed in claim 1, wherein the S2 includes driving behavior information identification, specifically:
collecting driving behavior data of driver by film sensor mounted on steering wheel and brake
The collected driving behavior data of the driver are sent to a random forest model, and the driving behavior data of the driver are divided into a road rage signal and a non-road rage signal through the random forest model;
the driving behavior data of the driver comprise average starting hot-car time, on-site idling ratio, rapid accelerator pedaling frequency and rapid braking frequency of the driver within three days.
5. The multi-modal driver emotion aid adjusting method as claimed in claim 1, wherein in S3, each of the driver emotion states in the single mode is a road rage signal, which is recorded as "1"; the emotional state of the driver in each single mode is a road rage signal and is marked as '0'; the total road rage signal is graded, and the grading comprises the following steps:
when the emotional states of the drivers in the single modes are added to be 0, the driver is not angry;
when the emotional states of the drivers in the single modes are added to be 1, the driver is in mild road rage;
when the emotional states of the drivers in the single modes are added to be 2, the driver is moderate road rage;
and when the emotional states of the drivers in the single modes are added to be 3, the road is heavily irritated.
CN202010157896.XA 2020-03-09 2020-03-09 Multi-modal driver emotion auxiliary adjusting method Pending CN111329498A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010157896.XA CN111329498A (en) 2020-03-09 2020-03-09 Multi-modal driver emotion auxiliary adjusting method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010157896.XA CN111329498A (en) 2020-03-09 2020-03-09 Multi-modal driver emotion auxiliary adjusting method

Publications (1)

Publication Number Publication Date
CN111329498A true CN111329498A (en) 2020-06-26

Family

ID=71174294

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010157896.XA Pending CN111329498A (en) 2020-03-09 2020-03-09 Multi-modal driver emotion auxiliary adjusting method

Country Status (1)

Country Link
CN (1) CN111329498A (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112078588A (en) * 2020-08-11 2020-12-15 大众问问(北京)信息科技有限公司 Vehicle control method and device and electronic equipment
CN112784695A (en) * 2020-12-31 2021-05-11 南京视察者智能科技有限公司 Driver abnormal state detection method based on image and voice recognition
CN113658580A (en) * 2021-06-24 2021-11-16 大众问问(北京)信息科技有限公司 Voice prompt method and device, computer equipment and storage medium
US20220355804A1 (en) * 2021-05-05 2022-11-10 Mitac Digital Technology Corporation Method and system for enhancing driving safety
CN115359532A (en) * 2022-08-23 2022-11-18 润芯微科技(江苏)有限公司 Human face emotion capturing and outputting device based on 3D sensing

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100134302A1 (en) * 2008-12-01 2010-06-03 Electronics And Telecommunications Research Institute System and method for controlling emotion of car driver
CN106114516A (en) * 2016-08-31 2016-11-16 合肥工业大学 The angry driver behavior modeling of a kind of drive automatically people's characteristic and tampering devic
CN106803423A (en) * 2016-12-27 2017-06-06 智车优行科技(北京)有限公司 Man-machine interaction sound control method, device and vehicle based on user emotion state
CN107235045A (en) * 2017-06-29 2017-10-10 吉林大学 Consider physiology and the vehicle-mounted identification interactive system of driver road anger state of manipulation information
CN107316436A (en) * 2017-07-31 2017-11-03 努比亚技术有限公司 Dangerous driving state processing method, electronic equipment and storage medium
CN108364373A (en) * 2018-02-07 2018-08-03 安徽星网软件技术有限公司 Vehicle behavior monitoring method and device
CN110751381A (en) * 2019-09-30 2020-02-04 东南大学 Road rage vehicle risk assessment and prevention and control method

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100134302A1 (en) * 2008-12-01 2010-06-03 Electronics And Telecommunications Research Institute System and method for controlling emotion of car driver
CN106114516A (en) * 2016-08-31 2016-11-16 合肥工业大学 The angry driver behavior modeling of a kind of drive automatically people's characteristic and tampering devic
CN106803423A (en) * 2016-12-27 2017-06-06 智车优行科技(北京)有限公司 Man-machine interaction sound control method, device and vehicle based on user emotion state
CN107235045A (en) * 2017-06-29 2017-10-10 吉林大学 Consider physiology and the vehicle-mounted identification interactive system of driver road anger state of manipulation information
CN107316436A (en) * 2017-07-31 2017-11-03 努比亚技术有限公司 Dangerous driving state processing method, electronic equipment and storage medium
CN108364373A (en) * 2018-02-07 2018-08-03 安徽星网软件技术有限公司 Vehicle behavior monitoring method and device
CN110751381A (en) * 2019-09-30 2020-02-04 东南大学 Road rage vehicle risk assessment and prevention and control method

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112078588A (en) * 2020-08-11 2020-12-15 大众问问(北京)信息科技有限公司 Vehicle control method and device and electronic equipment
CN112784695A (en) * 2020-12-31 2021-05-11 南京视察者智能科技有限公司 Driver abnormal state detection method based on image and voice recognition
CN112784695B (en) * 2020-12-31 2024-02-20 南京视察者智能科技有限公司 Method for detecting abnormal state of driver based on image and voice recognition
US20220355804A1 (en) * 2021-05-05 2022-11-10 Mitac Digital Technology Corporation Method and system for enhancing driving safety
CN113658580A (en) * 2021-06-24 2021-11-16 大众问问(北京)信息科技有限公司 Voice prompt method and device, computer equipment and storage medium
CN115359532A (en) * 2022-08-23 2022-11-18 润芯微科技(江苏)有限公司 Human face emotion capturing and outputting device based on 3D sensing

Similar Documents

Publication Publication Date Title
CN111329498A (en) Multi-modal driver emotion auxiliary adjusting method
CN106803423B (en) Man-machine interaction voice control method and device based on user emotion state and vehicle
CN112277955B (en) Driving assistance method, device, equipment and storage medium
CN106448221B (en) Pushing method of online driving assistance information pushing system and application of pushing method
CN112215097A (en) Method for monitoring driving state of vehicle, vehicle and computer readable storage medium
CN112785837A (en) Method and device for recognizing emotion of user when driving vehicle, storage medium and terminal
US20210362594A1 (en) Car driver alcohol level and sleeping status detection and notification system
CN112784695A (en) Driver abnormal state detection method based on image and voice recognition
CN109541600A (en) A kind of heavy type commercial automobile safe and intelligent driving assistance system
CN116729403A (en) Emotion recognition reminding and control system and method in automobile driving process
CN111904806A (en) Blind guiding system
CN206049658U (en) Angry driver behavior modeling and tampering devic based on drive automatically people's characteristic
Kamaruddin et al. Driver behaviour state recognition based on speech
JP2010271794A (en) Driving behavior guiding system
KR102270011B1 (en) Deep learning-based autonomous vehicle visualize system for the visually impaired and method thereof
Malta et al. Multi-modal real-world driving data collection, transcription, and integration using Bayesian network
KR102132058B1 (en) Interactive voice communication system embedded in a car
CN114655241A (en) Automobile intelligent multimedia central control screen fatigue driving reminding method
CN112773349A (en) Method and device for constructing emotion log of user driving vehicle, storage medium and terminal
CN111660927A (en) Whistling method and system
Kashevnik et al. Driver intelligent support system in internet of transportation things: Smartphone-based approach
JP2021114081A (en) Drive recorder, recording method, and program
JP3471042B2 (en) Automotive electronics
CN112706691A (en) Vehicle reminding method and device
US20230063708A1 (en) Systems and methods for testing driver awareness within a vehicle and/or intervening during a possible crisis

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination