WO2023017291A1 - 運転支援装置及び運転支援方法 - Google Patents
運転支援装置及び運転支援方法 Download PDFInfo
- Publication number
- WO2023017291A1 WO2023017291A1 PCT/IB2021/000530 IB2021000530W WO2023017291A1 WO 2023017291 A1 WO2023017291 A1 WO 2023017291A1 IB 2021000530 W IB2021000530 W IB 2021000530W WO 2023017291 A1 WO2023017291 A1 WO 2023017291A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- user
- vehicle
- message
- driving
- controller
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims description 16
- 230000002159 abnormal effect Effects 0.000 claims abstract description 22
- 230000004044 response Effects 0.000 claims abstract description 15
- 208000019901 Anxiety disease Diseases 0.000 claims description 9
- 230000036506 anxiety Effects 0.000 claims description 8
- 238000012790 confirmation Methods 0.000 claims description 6
- 230000000694 effects Effects 0.000 claims description 5
- 230000006870 function Effects 0.000 description 69
- 238000004458 analytical method Methods 0.000 description 10
- 230000008859 change Effects 0.000 description 8
- 230000010365 information processing Effects 0.000 description 7
- 230000005856 abnormality Effects 0.000 description 5
- 238000001514 detection method Methods 0.000 description 5
- 230000008569 process Effects 0.000 description 5
- 230000001133 acceleration Effects 0.000 description 2
- 238000004590 computer program Methods 0.000 description 2
- 238000009434 installation Methods 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 230000002265 prevention Effects 0.000 description 2
- 230000003044 adaptive effect Effects 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000002349 favourable effect Effects 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
- G08G1/16—Anti-collision systems
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W40/00—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
- B60W40/08—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models related to drivers or passengers
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W50/00—Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
- B60W50/08—Interaction between the driver and the control system
- B60W50/10—Interpretation of driver requests or demands
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W50/00—Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
- B60W50/08—Interaction between the driver and the control system
- B60W50/14—Means for informing the driver, warning the driver or prompting a driver intervention
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
- G08G1/09—Arrangements for giving variable traffic instructions
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W50/00—Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
- B60W50/08—Interaction between the driver and the control system
- B60W50/14—Means for informing the driver, warning the driver or prompting a driver intervention
- B60W2050/143—Alarm means
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W50/00—Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
- B60W50/08—Interaction between the driver and the control system
- B60W50/14—Means for informing the driver, warning the driver or prompting a driver intervention
- B60W2050/146—Display means
Definitions
- the present invention relates to a driving support device and a driving support method.
- Patent Document 1 an invention that detects dangerous driving by a user and presents driving characteristics to the user.
- the system determines whether or not it is dangerous driving. It is up to the user himself whether or not to accept the judgment result. Therefore, driving assistance based on the determination result may not be effective.
- the present invention has been made in view of the above problems, and its object is to provide a driving assistance device and a driving assistance method that are capable of providing driving assistance that matches the user's senses.
- the driving support device 1 determines whether or not the vehicle is in an abnormal state based on the data acquired by the sensor, and if it is determined that the vehicle is in an abnormal state, the vehicle A message indicating that the vehicle is in an abnormal state is output to a user on board, and the user's response to the message is acquired via an input device.
- FIG. 1 is a configuration diagram of a driving assistance device 1 according to an embodiment of the present invention.
- FIG. 2 shows a conversation between user 51 and robot head 40 .
- FIG. 3 is an example of a method for determining dangerous driving.
- FIG. 4 is an example of data stored when the assistance function is performed.
- FIG. 5 is an example of a message output from the robot head 40.
- FIG. 6 shows a conversation between user 51 and robot head 40 .
- FIG. 7 shows a conversation between user 51 and robot head 40 .
- FIG. 8 is a flow chart for explaining an operation example of the driving assistance device 1 .
- FIG. 9 is a flow chart for explaining an operation example of the driving support device 1.
- FIG. FIG. 10 is a flow chart for explaining an operation example of the driving support device 1. As shown in FIG.
- the driving assistance device 1 includes a sensor 10, a microphone 11, a camera 12, a GPS receiver 13, a switch 14, a controller 20, a robot head 40, a speaker 41, and a display 42. , a steering actuator 43 , an accelerator pedal actuator 44 and a brake actuator 45 .
- the driving support device 1 is installed in a vehicle with an automatic driving function.
- the driving support device 1 may be installed in a vehicle capable of switching between automatic driving and manual driving.
- An example of an automatic driving function is to automatically control steering control, braking force control, driving force control, etc. to assist the user's driving.
- "user” means a user sitting in the driver's seat of the vehicle.
- the sensor 10 is used to acquire various data and information.
- the sensors 10 include sensors that acquire vehicle data and sensors that acquire information outside the vehicle. Examples of vehicle data are speed, acceleration, steering angle, brake hydraulic pressure, and accelerator opening. Sensors that acquire these data include a speed sensor, an acceleration sensor, a steering angle sensor, a gyro sensor, a brake oil pressure sensor, an accelerator opening sensor, and the like.
- Examples of information external to the vehicle include objects (pedestrians, bicycles, motorcycles, other vehicles, etc.) existing around the vehicle (self-vehicle), traffic lights, lane markings, signs, pedestrian crossings, and intersections.
- Sensors that acquire this information include laser range finders, radars, lidars, cameras 12, sonars, and the like.
- the sensor 10 and the camera 12 are separated for convenience of explanation, but the camera 12 is a kind of the sensor 10 .
- Data and information acquired by the sensor 10 are output to the controller 20 .
- the microphone 11 acquires the user's voice data. Audio data acquired by the microphone 11 is output to the controller 20 .
- the camera 12 has an imaging device such as a CCD (charge-coupled device) or a CMOS (complementary metal oxide semiconductor).
- the installation location of the camera 12 is not particularly limited, as an example, the camera 12 is installed in the front, side, and rear of the vehicle. Camera 12 images the surroundings of the vehicle. Image data captured by the camera 12 is output to the controller 20 .
- the camera 12 is also installed inside the vehicle. For example, the camera 12 is installed near the driver's seat and captures an image of the user's face. Face image data captured by the camera 12 is output to the controller 20 .
- the GPS receiver 13 detects the position information of the vehicle on the ground by receiving radio waves from satellites.
- the vehicle position information detected by the GPS receiver 13 includes latitude information and longitude information.
- the GPS receiver 13 outputs the detected vehicle position information to the controller 20 .
- the method of detecting the position information of the vehicle is not limited to the GPS receiver 13 .
- a vehicle's position may be estimated using a method called odometry.
- Odometry is a method of estimating the position of a vehicle by obtaining the amount and direction of movement of the vehicle according to the rotation angle and rotation angular velocity of the vehicle.
- a GNSS receiver may be used instead of the GPS receiver 13 .
- the switch 14 is used when the user uses the support function. If the user presses switch 14, the assist function is activated.
- the "support functions" in this embodiment include around view monitor (hereinafter referred to as AVM), automatic parking, side down view monitor, steering support, emergency stop braking, emergency avoidance steering, lane departure prevention, blind spot warning, AVM and a combination of ultrasonic sensors, and the like. Since these support functions are well known, detailed description thereof will be omitted.
- the switch 14 may be a physical switch or a virtual switch. A virtual switch is, for example, a switch displayed on the display 42 .
- Some of the assistance functions mentioned above are also autonomous driving functions. That is, the support function of this embodiment also includes the automatic driving function. Autonomous driving functions include ACC (Adaptive Cruise Control) and constant speed control.
- the controller 20 is a general-purpose microcomputer equipped with a CPU (central processing unit), memory, and input/output unit.
- a computer program for functioning as the driving support device 1 is installed in the microcomputer.
- the microcomputer functions as a plurality of information processing circuits included in the driving assistance device 1.
- FIG. Here, an example of realizing a plurality of information processing circuits provided in the driving assistance device 1 by software is shown, but of course, dedicated hardware for executing each information processing described below is prepared to configure the information processing circuit. It is also possible to Also, a plurality of information processing circuits may be configured by individual hardware.
- the controller 20 includes, as an example of a plurality of information processing circuits, a dangerous situation determination unit 21, a responsibility determination unit 22, a message creation unit 23, an utterance acquisition unit 24, an utterance analysis unit 25, a parameter change unit 26, a user A determination unit 27 , a position detection unit 28 , a support function selection unit 32 , a proposal unit 33 , and a support function execution unit 34 are provided.
- the robot head 40 is an ornament imitating the head of a robot.
- the robot head 40 is installed to enhance affinity between the user and the vehicle.
- the robot head 40 has a speaker function and an information processing function.
- dangerous situation determination unit 21 responsibility determination unit 22, message creation unit 23, speech acquisition unit 24, speech analysis unit 25, parameter change unit 26, user determination unit 27, position detection unit 28, the support function selection unit 32, the proposal unit 33, and the support function execution unit 34 will be described in detail.
- the scene shown in FIG. 2 is a scene in which the user 51 is having a conversation with the robot head 40. Conversations are indicated at 60-63.
- Reference numerals 60 and 62 indicate the utterances of the robot head 40 .
- Reference numerals 61 and 63 indicate user 51's utterances.
- the conversation flow is in the order 60, 61, 62, 63.
- Reference numeral 52 indicates a preceding vehicle.
- Reference numerals 60 to 63 are all described as voice, but are not limited to this.
- symbols 60, 62 may be displayed on display 42 as characters.
- Reference numerals 61 and 63 may indicate information selected by the user 51 from a plurality of options. In this case, multiple options are displayed on the display 42 . First, the conditions for outputting "What happened?" indicated by reference numeral 60 will be described. Installation of the robot head 40 is not essential. A virtual robot head may be displayed on the display 42 .
- the dangerous situation determination unit 21 uses the data acquired from the sensor 10 to determine whether or not the driving of the user 51 was dangerous. An example of the determination method will be described with reference to FIG. The scenes shown in FIG. 3 are classified into parking, narrow road, approaching preceding vehicle, lane departure, merging or changing lanes, right turn, and left turn. In "parking”, if the distance to the obstacle is 15 cm or less, or if the number of turns in the front-rear direction is 5 or more, the dangerous situation determination unit 21 determines that the driving of the user 51 was dangerous. The distance to obstacles is detected by the lidar. The number of turns is detected by a steering angle sensor.
- the dangerous situation determination unit 21 determines that the driving of the user 51 was dangerous.
- the Japanese traffic rules traffic on the left side
- “Left” or “right” can be read appropriately according to the traffic regulations of each country.
- the dangerous situation determination unit 21 determines that the driving of the user 51 was dangerous. .
- the speed of the own vehicle is detected by a speed sensor.
- the distance to the preceding vehicle is detected by the lidar.
- the dangerous situation determination unit 21 determines that the driving of the user 51 was dangerous.
- a camera 12 detects the distance to the marking line.
- the dangerous situation determination unit 21 determines that the driving of the user 51 was dangerous.
- the distance to the rear side vehicle is detected by the lidar.
- the steering direction is detected by a steering angle sensor.
- the dangerous situation determination unit 21 determines that the driving of the user 51 was dangerous.
- the TTC with the oncoming vehicle traveling straight is calculated using the speed of the oncoming vehicle traveling straight and the distance to the oncoming vehicle traveling straight.
- the dangerous situation determination unit 21 determines that the driving of the user 51 was dangerous.
- a signal indicating the determination result is sent to the responsibility determination unit 22 .
- the responsibility determination unit 22 determines who is responsible for the dangerous driving. In the scene of FIG. 3, it is determined that the user 51 is responsible for dangerous driving. In this case, a signal indicating the determination result is sent to the message creation unit 23 .
- the message creating section 23 creates a message indicated by reference numeral 60 .
- the message creating unit 23 outputs the created message to the robot head 40 .
- a message 60 is output by voice as shown in FIG.
- FIG. 2 in response to the question "What happened?", the user 51 replies "It was a little dangerous! This answer is an affirmative answer.
- a "positive answer” here means an answer by which the user 51 acknowledges that his or her driving was dangerous.
- the voice response of the user 51 is output to the speech acquisition unit 24 via the microphone 11 .
- the speech acquisition unit 24 outputs voice data of the user 51 to the speech analysis unit 25 .
- the speech analysis unit 25 analyzes voice data of the user 51 .
- a well-known technique is used for the speech analysis method.
- the utterance analysis unit 25 determines whether or not the answer of the user 51 is affirmative based on the analysis result.
- a signal indicating the determination result is output to the message creating unit 23 .
- the message creating unit 23 creates a message indicated by reference numeral 62 (Do you want to support us next time?). Message 62 is created because user 51 answered in the affirmative. As shown in FIG.
- a “positive answer” as used herein is an answer that accepts the support function.
- This answer 63 is also analyzed by the utterance analysis unit 25 in the same manner as the answer 61 . Then, it is determined whether or not the answer of the user 51 is affirmative.
- a signal indicating the determination result is output to the user database 31 .
- a user database 31 is stored in the storage device 30 .
- the storage device 30 is composed of an HDD (Hard Disk Drive), an SSD (Solid State Drive), and the like.
- the position detection unit 28 uses the data acquired from the GPS receiver 13 to detect the position (place) of the vehicle at that time.
- the detected position information is stored in the user database 31 .
- the "support function" in the present embodiment is implemented when the user 51 drives next time after the exchanges of codes 60-63.
- the camera 12 images the face of the user 51 .
- Face image data captured by the camera 12 is output to the user determination section 27 .
- the user determination unit 27 compares face image data acquired from the camera 12 with face image data pre-stored in the user database 31 to identify the user who has boarded the vehicle. In this embodiment, the user who got on board is identified as user 51 .
- the reason for specifying is to provide support suitable for the user. There are cases where one vehicle is shared by multiple users. In this case, it is necessary to identify the user who has boarded the vehicle in order to provide assistance suitable for the user. If the assistance is provided without specifying the user, the user may feel uncomfortable.
- the support function selection unit 32 selects support suitable for the scene from a plurality of support functions. As shown in FIG. 3, when the scene is "parking”, AVM or automatic parking is selected from multiple assistance functions. Note that AVM and automatic parking may be selected. If the scene is a "narrow road”, side down view monitor or steering assistance is selected from among the assistance functions. If the scene is "approaching preceding vehicle”, emergency stop braking or emergency avoidance steering is selected from a plurality of assistance functions. If the scene is "Lane Departure”, lane departure prevention is selected from among multiple assistance functions. If the scene is "merging or lane change”, blind spot warning or steering assistance is selected among multiple assistance functions. If the scene is a "right turn”, emergency stop braking is selected from among multiple assistance functions. If the scene is a "left turn”, the combination of AVM and ultrasonic sensor is selected from among multiple assist functions.
- the support unit 33 determines that the vehicle is in the parking lot A based on the position information acquired from the position detection unit 28 .
- the user 51 consented to the assistance function.
- the consent result is stored in the user database 31.
- FIG. The support function selection unit 32 selects a support function to be proposed to the user 51 by referring to the user database 31 . Assume that automatic parking is selected here.
- the support function selection unit 32 outputs the selection result to the proposal unit 33 .
- the proposal unit 33 confirms with the user 51 whether or not to implement the support function.
- the reason for the reconfirmation is to enhance the relationship of trust between the user 51 and the vehicle.
- This confirmation may be performed by voice through the speaker 41 or by displaying characters on the display 42 .
- the user 51 uses the switch 14 to respond to this confirmation. If the user 51 presses the switch 14, he consents to the assistance function.
- the proposal unit 33 Upon receiving the signal indicating that the switch 14 is on, the proposal unit 33 outputs a command for executing the support function to the support function execution unit 34 .
- the support function execution unit 34 that has received the command controls the steering actuator 43 , the accelerator pedal actuator 44 and the brake actuator 45 . This allows automatic parking.
- the data at that time is recorded in the user database 31.
- the data to be recorded are the user name, date and time, location, scene, and assistance function performed, as shown in FIG.
- the message creation unit 23 creates messages 60 and 62 and outputs them to the robot head 40, this is not the only option. Such messages may be pre-composed.
- the message created in advance may be stored in the storage device 30 .
- the message creation unit 23 can output an appropriate message to the robot head 40 .
- An example of a pre-written message is shown in FIG.
- the messages selected by the message creation unit 23 are "What happened?", "What happened”, "What happened?", "What did you do?", " What did you do?” is mentioned.
- These message examples 1 are intended to make the user 51 indirectly accept responsibility.
- message example 2 in FIG. 5 is intended to prompt the user 51 to give a specific answer.
- Message example 2 includes, as shown in FIG. 5, "Which is the fault?"
- the scene shown in FIG. 6 is a scene in which the user 51 is having a conversation with the robot head 40 .
- Conversations are indicated by reference numerals 70-73.
- Reference numerals 70 and 72 indicate the utterances of the robot head 40 .
- Reference numerals 71 and 73 indicate utterances of the user 51 .
- the flow of conversation is in the order 70, 71, 72, 73. Conditions for outputting "Are you scared?"
- a signal indicating the determination result is sent to the responsibility determination unit 22 .
- the responsibility determination unit 22 determines who is responsible for the dangerous driving. Here, it is determined that responsibility for dangerous driving lies with automatic driving (vehicle side).
- a signal indicating the determination result is sent to the message creation unit 23 .
- the message creating section 23 creates a message indicated by reference numeral 70 .
- the message creating unit 23 outputs the created message to the robot head 40 .
- a message 70 is output by voice as shown in FIG. As shown in FIG.
- the message creating unit 23 creates a message indicated by reference numeral 72 (do you want to make it longer from the next time?). Message 72 is created because user 51 answered in the affirmative. As shown in FIG. 6, the user 51 responds "Yes! to the question "Would you like to extend the next time?" (reference numeral 73). This answer is an affirmative answer.
- the "affirmative answer” here is an answer for changing the parameters of the support function (automatic driving function).
- the answer 73 is analyzed by the utterance analysis unit 25 . Then, it is determined whether or not the answer of the user 51 is affirmative. A signal indicating the determination result is output to the parameter changing unit 26 .
- the parameter changing unit 26 changes the inter-vehicle distance to be longer than that during the previous automatic driving.
- the changed parameters are output to the support function execution unit 34 .
- the support function execution unit 34 controls the inter-vehicle distance based on the changed parameters. As a result, it is possible to perform driving assistance that matches the user's senses.
- the messages 70 and 72 may also be created in advance, similar to the messages 60 and 62.
- the messages selected by the message creating unit 23 are "Did you feel dangerous?", “Was it dangerous driving?” ?” can be mentioned.
- Other messages include “Did you brake late?”, “Were you too close?”, “Was the preceding vehicle too close?”, “Have you stepped on the accelerator too much?” Was the steering too slow/too fast?)?”, “Did you lean too far to the right/left?”
- the cause of the occurrence of a dangerous situation is not limited to the user's 51 driving or automatic driving. For example, it may be caused by a user driving another vehicle, an automatic driving of another vehicle, a pedestrian, or the like.
- the scene shown in FIG. 7 is a scene in which a pedestrian 53 jumps out and the situation becomes dangerous.
- Reference numeral 80 indicates the utterance of the robot head 40 .
- Reference numeral 81 indicates the utterance of the user 51 .
- the flow of conversation is in the order of numerals 80,81.
- the message creation unit 23 creates a message 80 (Are you okay?) and outputs it to the robot head 40 .
- step S101 If the ignition is on (YES in step S101), the process proceeds to step S103. "Ignition on” means that the vehicle is powered on.
- steps S103 and S105 the dangerous situation determination unit 21 uses the data acquired from the sensor 10 to determine whether or not a dangerous situation has occurred. If a dangerous situation occurs (YES in step S105), the process proceeds to step S107.
- step S ⁇ b>107 the responsibility determination unit 22 estimates the cause of the occurrence of the dangerous situation based on the data acquired from the sensor 10 . For example, if the dangerous situation shown in FIG. 3 occurs due to the driving of the user 51, it is estimated that the user 51 is responsible for the dangerous driving (YES in step S113). When the dangerous situation shown in FIG.
- a message 80 is output (see FIG. 7).
- message 60 is output (see FIG. 2).
- the process proceeds to step S ⁇ b>117 , and the speech acquisition unit 24 acquires the user's 51 answer via the microphone 11 . If it is determined that the user's 51 answer is affirmative (YES in step S119), a message 62 for proposing a support function is output (see FIG. 2). If it is determined that the answer from user 51 is negative (NO in step S119), a message is output to the effect that user 51 is responsible for dangerous driving.
- a message 70 is output (see FIG. 6).
- the process proceeds to step S ⁇ b>129 , and the speech acquisition unit 24 acquires the user's 51 answer via the microphone 11 . If it is determined that the user's 51 answer is affirmative (YES in step S131), the parameter change unit 26 reflects the user's 51 desire in the control details when automatic operation is performed next time (step S133). ). Specifically, the parameter changing unit 26 changes parameters related to the automatic driving function. As an example of parameter change, the inter-vehicle distance is changed to be longer. If it is determined that the user's 51 answer is negative (NO in step S131), the parameters are not changed. The processing of steps S103 to S135 is repeated until the ignition is turned off.
- step S203 the proposal unit 33 acquires position information from the position detection unit 28 .
- step S205 the proposal unit 33 determines whether or not the current vehicle location is the location where the support function was previously proposed based on the position information. The “place where the support function was proposed last time” is called “place A”. If the vehicle's current location is location A (YES in step S205), the support function selection unit 32 refers to the user database 31 and proposes a support function to the user 51 (step S207). If the user 51 approves the support function (YES in step S209), the support function execution unit 34 executes the support function (step S211).
- step S213 If the user 51 does not accept the support function (NO in step S209), the support function execution unit 34 does not execute the support function (step S213).
- step S215 When the vehicle passes the place A (YES in step S215), data at that time is recorded in the user database 31 (step S217). The processing of steps S203 to S217 is repeated until the ignition is turned off.
- the driving assistance device 1 includes a sensor 10 that acquires data inside and outside the vehicle, and a controller 20 .
- Controller 20 determines whether the vehicle is in an abnormal state based on the data acquired by sensor 10 .
- the controller 20 determines that the vehicle is in an abnormal state, it outputs a message indicating that the vehicle is in an abnormal state to the user 51 riding in the vehicle.
- the controller 20 acquires the user's 51 reply to the message via the input device.
- An example of an input device is microphone 11 .
- the vehicle is in an abnormal state means a situation where there is a possibility of a vehicle collision or accident. It suffices to determine the “possibility of vehicle collision or accident” based only on the data acquired by the sensor 10 . It does not matter whether the driving subject is the user 51 or the automatic driving function. However, as will be described later, the driving entity may be taken into consideration.
- the answer of the user 51 relates to feelings of anxiety or danger about the state of the vehicle.
- the controller 20 determines that the user's 51 response to the message indicates anxiety or danger, it suggests using a support function to support the user 51 .
- An example of a response indicating anxiety or danger is message 61 shown in FIG. This makes it possible to perform driving assistance that matches the feeling of the user 51 .
- the controller 20 estimates the cause of the abnormal state based on the data acquired by the sensor 10. When the controller 20 presumes that the cause is caused by the driving of the user 51, it outputs a message confirming responsibility to the user 51.
- FIG. An example of a message confirming responsibility is "Which responsibility?" shown in FIG. If the controller 20 determines that the user's 51 response to the message is an admission of responsibility, then the controller 20 offers to use assistance functions to assist the user 51 . This makes it possible to perform driving assistance that matches the feeling of the user 51 . Also, by allowing the user 51 to accept responsibility, it is possible to make the relationship between the user 51 and the vehicle favorable.
- the messages include those confirming responsibility for the user 51, those regarding the state of the vehicle, those confirming the feelings of the user 51, or those confirming the safety of the user 51.
- An example of the "message for confirming responsibility to user 51" is "Which responsibility?" shown in FIG.
- An example of a "vehicle status message” is "Was the brake slow?” shown in FIG.
- An example of the "message confirming the feelings of the user 51" is "Are you scared?" shown in FIG.
- An example of the "message for confirming the safety of user 51” is "Did you feel dangerous?" shown in FIG.
- An example of a message confirming responsibility to the user 51 is a message confirming to the user 51 what happened in or around the vehicle (see FIG. 5).
- the controller 20 determines that the user 51's response to the message confirming responsibility does not admit responsibility, it outputs a message to the effect that the user 51 is responsible. This can provide the user 51 with an opportunity to change his or her way of thinking.
- the controller 20 estimates the cause of the abnormal state based on the data acquired by the sensor 10. When the controller 20 presumes that the cause is due to the automatic driving function, it outputs a message asking whether the user 51 feels uneasy.
- An example of the "message for confirming whether the user 51 felt uneasy" is "Are you scared?" shown in FIG. This makes it possible to confirm whether or not the user 51 feels uneasy about automatic driving.
- the controller 20 determines that the user's 51 response to the message confirming anxiety indicates that the user is feeling anxious
- the controller 20 changes the parameter settings for the automatic driving function. Specifically, the controller 20 changes the parameters of the automatic driving function to the safe side.
- An example of a parameter change is increasing the inter-vehicle distance. This makes it possible to perform driving assistance that matches the feeling of the user 51 . It will also increase confidence in automated driving.
- Processing circuitry includes programmed processing devices, such as processing devices that include electrical circuitry. Processing circuitry also includes devices such as application specific integrated circuits (ASICs) and circuit components arranged to perform the described functions.
- ASICs application specific integrated circuits
Landscapes
- Engineering & Computer Science (AREA)
- Automation & Control Theory (AREA)
- Transportation (AREA)
- Mechanical Engineering (AREA)
- Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- General Physics & Mathematics (AREA)
- Traffic Control Systems (AREA)
- Mathematical Physics (AREA)
Abstract
Description
以上説明したように、本実施形態に係る運転支援装置1によれば、以下の作用効果が得られる。
Claims (11)
- 車両の内部及び外部のデータを取得するセンサと、
コントローラと、を備え、
前記コントローラは、
前記センサによって取得されたデータに基づいて前記車両が異常な状態にあるか否かを判定し、
前記車両が異常な状態にあると判定した場合、前記車両に乗車しているユーザに対して前記車両が異常な状態にあることを示すメッセージを出力し、
前記メッセージに対する前記ユーザの回答を入力装置を介して取得する
ことを特徴とする運転支援装置。 - 前記車両が異常な状態にあるとは、前記車両の衝突または事故の可能性がある状況を意味する
ことを特徴とする請求項1に記載の運転支援装置。 - 前記ユーザの回答は、前記車両の状態に対する不安または危険を示す感情に関するものである
ことを特徴とする請求項1または2に記載の運転支援装置。 - 前記コントローラは、前記メッセージに対する前記ユーザの回答が不安または危険を示すものであると判定した場合、前記ユーザを支援するための支援機能の利用を提案する
ことを特徴とする請求項1~3のいずれか1項に記載の運転支援装置。 - 前記コントローラは、
前記センサによって取得されたデータに基づいて前記異常な状態が発生した原因を推定し、
前記原因が前記ユーザの運転に起因すると推定した場合、前記ユーザに対して責任を確認するメッセージを出力し、
前記メッセージに対する前記ユーザの回答が責任を認めるものであると判定した場合、前記ユーザを支援するための支援機能の利用を提案する
ことを特徴とする請求項4に記載の運転支援装置。 - 前記メッセージには、前記ユーザに対して責任を確認するもの、前記車両の状態に関するもの、前記ユーザの感情を確認するもの、または、前記ユーザの安全を確認するものが含まれる
ことを特徴とする請求項1~5のいずれか1項に記載の運転支援装置。 - 前記ユーザに対して責任を確認するメッセージは、前記車両または前記車両の周囲において何が起きたかを前記ユーザに確認するメッセージである
ことを特徴とする請求項6に記載の運転支援装置。 - 前記コントローラは、前記責任を確認するメッセージに対する前記ユーザの回答が前記責任を認めないものであると判定した場合、前記責任は前記ユーザにある旨のメッセージを出力する
ことを特徴とする請求項6または7に記載の運転支援装置。 - 前記コントローラは、
前記センサによって取得されたデータに基づいて前記異常な状態が発生した原因を推定し、
前記原因が自動運転機能に起因すると推定した場合、前記ユーザに対して不安を感じたかを確認するメッセージを出力する
ことを特徴とする請求項1~3のいずれか1項に記載の運転支援装置。 - 前記コントローラは、前記不安を確認するメッセージに対する前記ユーザの回答が前記不安を感じたことを示すものであったと判定した場合、前記自動運転機能のパラメータの設定を変更する
ことを特徴とする請求項9に記載の運転支援装置。 - 車両の内部及び外部のデータを取得し、
取得されたデータに基づいて前記車両が異常な状態にあるか否かを判定し、
前記車両が異常な状態にあると判定した場合、前記車両に乗車しているユーザに対して前記車両が異常な状態にあることを示すメッセージを出力し、
前記メッセージに対する前記ユーザの回答を入力装置を介して取得する
ことを特徴とする運転支援方法。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202180101467.3A CN117794802A (zh) | 2021-08-11 | 2021-08-11 | 驾驶辅助装置和驾驶辅助方法 |
EP21953025.0A EP4386712A1 (en) | 2021-08-11 | 2021-08-11 | Driving assistance device and driving assistance method |
JP2023540988A JPWO2023017291A1 (ja) | 2021-08-11 | 2021-08-11 | |
PCT/IB2021/000530 WO2023017291A1 (ja) | 2021-08-11 | 2021-08-11 | 運転支援装置及び運転支援方法 |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/IB2021/000530 WO2023017291A1 (ja) | 2021-08-11 | 2021-08-11 | 運転支援装置及び運転支援方法 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2023017291A1 true WO2023017291A1 (ja) | 2023-02-16 |
Family
ID=85199908
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/IB2021/000530 WO2023017291A1 (ja) | 2021-08-11 | 2021-08-11 | 運転支援装置及び運転支援方法 |
Country Status (4)
Country | Link |
---|---|
EP (1) | EP4386712A1 (ja) |
JP (1) | JPWO2023017291A1 (ja) |
CN (1) | CN117794802A (ja) |
WO (1) | WO2023017291A1 (ja) |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004291174A (ja) * | 2003-03-27 | 2004-10-21 | Denso Corp | ロボット、ロボット対話装置及びプログラム |
WO2009107185A1 (ja) * | 2008-02-25 | 2009-09-03 | パイオニア株式会社 | 車載ロボット |
JP2018147111A (ja) * | 2017-03-02 | 2018-09-20 | パナソニックIpマネジメント株式会社 | 運転支援方法およびそれを利用した運転支援装置、自動運転制御装置、車両、プログラム、運転支援システム |
JP2018144568A (ja) * | 2017-03-02 | 2018-09-20 | パナソニックIpマネジメント株式会社 | 運転支援方法およびそれを利用した運転支援装置、自動運転制御装置、車両、プログラム、運転支援システム |
JP2018144570A (ja) * | 2017-03-02 | 2018-09-20 | パナソニックIpマネジメント株式会社 | 運転支援方法およびそれを利用した運転支援装置、自動運転制御装置、車両、プログラム、運転支援システム |
JP2018144569A (ja) * | 2017-03-02 | 2018-09-20 | パナソニックIpマネジメント株式会社 | 運転支援方法およびそれを利用した運転支援装置、自動運転制御装置、車両、プログラム、運転支援システム |
JP6480143B2 (ja) | 2014-10-09 | 2019-03-06 | 株式会社日立製作所 | 運転特性診断装置、運転特性診断システム、運転特性診断方法 |
JP2019116185A (ja) * | 2017-12-27 | 2019-07-18 | クラリオン株式会社 | 運転支援装置、及び運転支援方法 |
-
2021
- 2021-08-11 EP EP21953025.0A patent/EP4386712A1/en active Pending
- 2021-08-11 WO PCT/IB2021/000530 patent/WO2023017291A1/ja active Application Filing
- 2021-08-11 CN CN202180101467.3A patent/CN117794802A/zh active Pending
- 2021-08-11 JP JP2023540988A patent/JPWO2023017291A1/ja active Pending
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004291174A (ja) * | 2003-03-27 | 2004-10-21 | Denso Corp | ロボット、ロボット対話装置及びプログラム |
WO2009107185A1 (ja) * | 2008-02-25 | 2009-09-03 | パイオニア株式会社 | 車載ロボット |
JP6480143B2 (ja) | 2014-10-09 | 2019-03-06 | 株式会社日立製作所 | 運転特性診断装置、運転特性診断システム、運転特性診断方法 |
JP2018147111A (ja) * | 2017-03-02 | 2018-09-20 | パナソニックIpマネジメント株式会社 | 運転支援方法およびそれを利用した運転支援装置、自動運転制御装置、車両、プログラム、運転支援システム |
JP2018144568A (ja) * | 2017-03-02 | 2018-09-20 | パナソニックIpマネジメント株式会社 | 運転支援方法およびそれを利用した運転支援装置、自動運転制御装置、車両、プログラム、運転支援システム |
JP2018144570A (ja) * | 2017-03-02 | 2018-09-20 | パナソニックIpマネジメント株式会社 | 運転支援方法およびそれを利用した運転支援装置、自動運転制御装置、車両、プログラム、運転支援システム |
JP2018144569A (ja) * | 2017-03-02 | 2018-09-20 | パナソニックIpマネジメント株式会社 | 運転支援方法およびそれを利用した運転支援装置、自動運転制御装置、車両、プログラム、運転支援システム |
JP2019116185A (ja) * | 2017-12-27 | 2019-07-18 | クラリオン株式会社 | 運転支援装置、及び運転支援方法 |
Also Published As
Publication number | Publication date |
---|---|
EP4386712A1 (en) | 2024-06-19 |
CN117794802A (zh) | 2024-03-29 |
JPWO2023017291A1 (ja) | 2023-02-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11673569B2 (en) | Alert control apparatus and alert control method | |
CN111361552B (zh) | 自动驾驶*** | |
JP7139717B2 (ja) | 車両用通信装置、車両用通信方法、及び制御プログラム | |
JP6515814B2 (ja) | 運転支援装置 | |
JP5987906B2 (ja) | 運転支援装置 | |
WO2018186127A1 (ja) | 走行支援装置 | |
US10843693B2 (en) | System and method for rear collision avoidance | |
WO2014148025A1 (ja) | 走行制御装置 | |
JP7119653B2 (ja) | 車両制御装置 | |
WO2013153660A1 (ja) | 運転支援装置 | |
JP2019142246A (ja) | 車両制御装置 | |
KR101552017B1 (ko) | 성능이 개선된 운전보조시스템 및 그 제어방법 | |
JP7382327B2 (ja) | 情報処理装置、移動体、情報処理方法及びプログラム | |
JP2006301842A (ja) | 車両の運転支援装置 | |
US20190232931A1 (en) | Method and driver assistance system for assisting a driver in driving a vehicle | |
US11130488B2 (en) | Vehicle control device and vehicle control method | |
US11904856B2 (en) | Detection of a rearward approaching emergency vehicle | |
US11613252B2 (en) | Driving assistance system and control method thereof | |
JP6631569B2 (ja) | 運転状態判定装置、運転状態判定方法及び運転状態判定のためのプログラム | |
JP6648551B2 (ja) | 自動運転装置 | |
KR20210157520A (ko) | 센서 각도 조절을 활용한 전방 충돌 방지 시스템 및 방법 | |
WO2023017291A1 (ja) | 運転支援装置及び運転支援方法 | |
JP5018102B2 (ja) | 衝突防止装置 | |
JP6648722B2 (ja) | 故障判定装置、方法およびプログラム | |
JP7441839B2 (ja) | 異常検出装置と異常検出方法およびプログラムと情報処理システム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21953025 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2023540988 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 202180101467.3 Country of ref document: CN |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 2021953025 Country of ref document: EP Effective date: 20240311 |