EP2793045A1 - Procédé de vérification d'un système de détection de l'environnement d'un véhicule - Google Patents
Procédé de vérification d'un système de détection de l'environnement d'un véhicule Download PDFInfo
- Publication number
- EP2793045A1 EP2793045A1 EP20140163655 EP14163655A EP2793045A1 EP 2793045 A1 EP2793045 A1 EP 2793045A1 EP 20140163655 EP20140163655 EP 20140163655 EP 14163655 A EP14163655 A EP 14163655A EP 2793045 A1 EP2793045 A1 EP 2793045A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- vehicle
- sensor type
- objects
- environment
- static
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 32
- 238000001514 detection method Methods 0.000 title claims abstract description 31
- 238000012360 testing method Methods 0.000 title description 2
- 230000007613 environmental effect Effects 0.000 claims abstract description 46
- 230000003068 static effect Effects 0.000 claims abstract description 46
- 238000004590 computer program Methods 0.000 claims abstract description 9
- 230000000007 visual effect Effects 0.000 claims abstract description 5
- 230000003287 optical effect Effects 0.000 claims description 36
- 230000001133 acceleration Effects 0.000 claims description 4
- 230000007547 defect Effects 0.000 claims description 4
- 230000004907 flux Effects 0.000 description 15
- 238000005259 measurement Methods 0.000 description 13
- 230000004888 barrier function Effects 0.000 description 7
- 238000002604 ultrasonography Methods 0.000 description 6
- 238000004364 calculation method Methods 0.000 description 3
- 230000008859 change Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000007257 malfunction Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000009897 systematic effect Effects 0.000 description 1
- 238000013024 troubleshooting Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S15/00—Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
- G01S15/88—Sonar systems specially adapted for specific applications
- G01S15/93—Sonar systems specially adapted for specific applications for anti-collision purposes
- G01S15/931—Sonar systems specially adapted for specific applications for anti-collision purposes of land vehicles
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S11/00—Systems for determining distance or velocity not using reflection or reradiation
- G01S11/12—Systems for determining distance or velocity not using reflection or reradiation using electromagnetic waves other than radio waves
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S15/00—Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
- G01S15/86—Combinations of sonar systems with lidar systems; Combinations of sonar systems with systems not using wave reflection
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S15/00—Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
- G01S15/87—Combinations of sonar systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/86—Combinations of radar systems with non-radar systems, e.g. sonar, direction finder
- G01S13/867—Combination of radar systems with cameras
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/87—Combinations of radar systems, e.g. primary radar and secondary radar
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/88—Radar or analogous systems specially adapted for specific applications
- G01S13/93—Radar or analogous systems specially adapted for specific applications for anti-collision purposes
- G01S13/931—Radar or analogous systems specially adapted for specific applications for anti-collision purposes of land vehicles
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/86—Combinations of lidar systems with systems other than lidar, radar or sonar, e.g. with direction finders
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/87—Combinations of systems using electromagnetic waves other than radio waves
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/88—Lidar systems specially adapted for specific applications
- G01S17/93—Lidar systems specially adapted for specific applications for anti-collision purposes
- G01S17/931—Lidar systems specially adapted for specific applications for anti-collision purposes of land vehicles
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/88—Radar or analogous systems specially adapted for specific applications
- G01S13/93—Radar or analogous systems specially adapted for specific applications for anti-collision purposes
- G01S13/931—Radar or analogous systems specially adapted for specific applications for anti-collision purposes of land vehicles
- G01S2013/9323—Alternative operation using light waves
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/88—Radar or analogous systems specially adapted for specific applications
- G01S13/93—Radar or analogous systems specially adapted for specific applications for anti-collision purposes
- G01S13/931—Radar or analogous systems specially adapted for specific applications for anti-collision purposes of land vehicles
- G01S2013/9324—Alternative operation using ultrasonic waves
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/88—Radar or analogous systems specially adapted for specific applications
- G01S13/93—Radar or analogous systems specially adapted for specific applications for anti-collision purposes
- G01S13/931—Radar or analogous systems specially adapted for specific applications for anti-collision purposes of land vehicles
- G01S2013/9327—Sensor installation details
- G01S2013/93274—Sensor installation details on the side of the vehicles
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S15/00—Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
- G01S15/88—Sonar systems specially adapted for specific applications
- G01S15/93—Sonar systems specially adapted for specific applications for anti-collision purposes
- G01S15/931—Sonar systems specially adapted for specific applications for anti-collision purposes of land vehicles
- G01S2015/937—Sonar systems specially adapted for specific applications for anti-collision purposes of land vehicles sensor installation details
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/217—Validation; Performance evaluation; Active pattern learning techniques
Definitions
- the invention relates to a method for checking a surroundings detection system of a vehicle, wherein the surroundings detection system comprises at least two different environmental sensor types, objects are detected in the environment of the vehicle with a first environment sensor type, the objects are categorized into static and dynamic objects and wherein the position of the static objects with the position determined by a second environment sensor type is compared. Furthermore, the invention relates to a computer program and a device which are set up for carrying out the method.
- Surround detection systems are used in modern vehicles to obtain data about the environment of the vehicle for use in driver assistance systems.
- Ultrasound, radar or lidar based sensors are used as sensors for the surroundings detection systems.
- optical systems such as cameras are suitable for environment detection.
- Stereo cameras allow a three-dimensional imaging of the vehicle environment. For a reliable three-dimensional detection of the environment, a stereo camera must first be calibrated. Further, the parameters of the stereo camera, such as the alignment of the two individual cameras to each other, must not change during operation, otherwise occur in the calculation of the depth of the captured image systematic errors, for example in the form of a distorted 3D survey.
- the image sensor system to be calibrated can be, for example, a stereo camera.
- the device comprises, in addition to the image sensor system, a further sensor system, for example a radar sensor or an ultrasound sensor, whose detection range overlaps that of the image sensor system.
- a calibration object is used, which is introduced into the detection range of the two different sensor systems.
- the measurement data of the various sensor systems are transmitted to an evaluation unit, which then determines the calibration data for calibrating the image sensor system.
- a method and apparatus for real-time obstacle detection is known based on a calibrated camera. It analyzes the optical flow in the camera images and selects a point around the vehicle that is easy to track. The movement of this point in the camera images is analyzed and compared with a movement of the point calculated via the known proper motion of the vehicle. To increase reliability, the method may be combined with other sensor systems, such as an ultrasonic sensor system.
- a method for detecting features in the surroundings of a vehicle wherein at least two images are taken with substantially the same orientations, but different locations of a camera.
- a stereo camera can be used, which typically contains two cameras.
- a laser, ultrasound or radar system can also be used in combination with the camera system.
- a disadvantage of the methods known from the prior art is that they require a further environmental sensor for checking or calibrating a camera, in particular a stereo camera, which supplies substantially redundant information. Since the environment sensors typically have only a limited field of view, additional sensors must therefore be installed so that the viewing areas of different sensor types at least partially overlap.
- a method for checking a surroundings detection system of a vehicle wherein the surroundings detection system comprises at least two different environmental sensor types, whereby objects in the surroundings of the vehicle are detected with a first environment sensor type, which objects are categorized with the data of the first environmental sensor type into static and dynamic objects, the relative position of the detected static object to the vehicle is determined, this position is compared with a position determined by a second environmental sensor type and an error is closed when a deviation above a threshold, wherein the field of view of the first environmental sensor type does not overlap with the field of view of the second environmental sensor type and the relative position the detected static objects after leaving the field of view of the first environmental sensor type is updated taking into account the movement of the vehicle.
- the surroundings detection system to be checked is arranged in a vehicle and comprises at least two different types of environment sensors.
- Each environment sensor has a viewing area in which these objects can recognize.
- the environmental sensors are arranged distributed on the vehicle such that their viewing areas can capture as much of the surroundings of the vehicle as possible.
- the different environmental sensor types differ both in the opening angle of their field of view, as well as in the range in which these objects can perceive.
- a sensor type with a long range is selected for detecting objects in the direction of travel, while sensor types with a short range are sufficient for the detection of objects located laterally next to the vehicle.
- sensor types are preferably selected whose viewing region has a large opening angle.
- a function check of a sensor can be done by comparing its measurement result with that of another sensor. To do this, two sensors must be set up to measure the same size. In the case of an environment detection system, this means that the position of an object relative to the vehicle has to be measured for a comparison of two different sensors. If it is then determined during the comparison that the difference is greater than a predefined limit value, an error is concluded in the surroundings detection system.
- the limit value is preferably selected such that usual statistical fluctuations in the position determinations are below the limit value.
- static objects are used to compare the positions determined. It is not necessary that the static object is simultaneously detected by two different environment sensors, that is, it is not necessary that the viewing areas of two different environment sensors overlap.
- objects in the surroundings of the vehicle are detected with an environment sensor of a first type, wherein the objects are categorized into static and dynamic objects with the aid of the data of this environment sensor.
- Static objects are immovable objects such as crash barriers, trees, traffic signs, lane markings or parked vehicles.
- Dynamic objects are moving objects such as other vehicles on the road.
- the relative position of the static objects to the vehicle is determined.
- the relative position of an object is then updated continuously. As long as the object is still within the field of view of the sensor of the first environment sensor type, its measurement data can be used. After leaving the field of view, the relative position of the static object to the vehicle is calculated taking into account the known movement of the vehicle.
- the static object enters the field of view of a sensor of a second type, in a third step of the method, its relative position to the vehicle is determined from the measured data of the second environmental sensor type.
- a fourth step of the method the relative position determined via a sensor of a second environmental sensor type is compared with the position which was determined on the basis of the measurement of the first sensor type in the second step of the method. If the difference between the two positions exceeds a predetermined limit value, an error is detected in the surroundings detection system of the vehicle.
- the limit value is preferably chosen such that only in the case of deviations greater than the fluctuations to be expected as a result of the measurement accuracy of the sensors are concluded to be a malfunction.
- the first environmental sensor type is a camera or a stereo camera.
- the categorization of detected objects into static and dynamic objects is done by comparing the measured optical flux in the camera image with the optical flux expected for static objects.
- the optical flow refers to a motion vector associated with each pixel of a camera image.
- the motion vector indicates the direction and speed into which the pixel assigned to it moves. If a camera is placed in a vehicle and aligned in the direction of travel and there are no moving objects in the field of view of the camera, pixels that represent distant objects or distant parts of the road have a low optical flux - they only move in the camera image slowly. On the other hand, pixels that belong to nearby objects or to a nearby part of the road have a large optical flux - they move quickly in the camera image.
- a hypothetical optical flux is calculated for each object on the assumption that the object is static. If the measured optical flux and the hypothetical optical flow agree except for a statistical error, the object is categorized as static.
- the calculation of the hypothetical optical flow value f stat of a static object is dependent on its position (x w , y w ) relative to the vehicle and the proper motion of the vehicle.
- ⁇ ' is the yaw rate and v f is the speed of the vehicle.
- (x 1 , y 1 ) is the position of the object in camera coordinates in the current camera image
- (x 2 , y 2 ) is the hypothetical position in the previous camera image
- the flow f is the object's movement by a certain number of columns camera image.
- T is the time that elapses between two camera images.
- the position (x 2, w , y 2, w ) can then be converted into the camera coordinate system.
- the calculated hypothetical optical flux f stat of each object is now compared with the optical flux measured by the camera. If the values agree except for a difference resulting from the measurement uncertainty, then the object is a static object. Otherwise, the object is categorized as a dynamic object. To reduce the measurement uncertainty, the calculation can be carried out over several camera images and an average can be formed.
- a categorization of the objects into static and dynamic objects can take place with the aid of an object recognition. If e.g. Warning beacons, traffic signs or crash barriers detected in an image region, the object can be classified as static.
- the distance to an object can also be estimated from a simple two-dimensional camera image. In the simplest case, this can be done by assigning a distance to each image line via the known orientation of the camera.
- the second environmental sensor type is an ultrasonic sensor.
- Ultrasonic sensors emit an ultrasonic signal and receive the signal echo reflected by an object. From the duration of the signal and the known speed of sound then the distance of the object to the ultrasonic sensor can be determined.
- the first surround sensor type is a stereo camera and the second surround sensor type is an ultrasound sensor, with two ultrasound sensors aligned with each side of the vehicle and the stereo camera facing forward.
- the position of a static object determined via the stereo camera and then updated via the known proper movement of the vehicle is compared with the positions determined via the respective ultrasound sensors when the vehicle passes by. If at least two ultrasonic sensors consistently detect different object positions from the stereo camera, it is concluded that the stereo camera has failed. If only one ultrasonic sensor indicates deviating values, an error is detected in this ultrasonic sensor.
- the driver is preferably informed of a detected error and the relevant sensor and / or the surroundings detection system are deactivated.
- movement of the vehicle is determined via wheel path sensors, wheel speed sensors, yaw rate sensors, acceleration sensors, and satellite navigation (e.g., GPS). Satellite navigation can improve the estimation of the movement of the vehicle, in particular by using the global heading of the vehicle.
- satellite navigation e.g., GPS
- a computer program is also proposed according to which one of the methods described herein is performed when the computer program is executed on a programmable computer device.
- the computer program may, for example, be a module for implementing a driver assistance system or a subsystem thereof in a vehicle or an application for driver assistance functions executable on a portable device, such as a smartphone or a tablet computer.
- the computer program may be stored on a machine-readable storage medium, such as on a permanent or rewritable storage medium or in association with a computer device or on a portable unit, such as a removable CD-ROM, DVD, memory card or USB stick.
- the computer program may be provided for download on a computing device, such as on a server, e.g. via a data network such as the Internet or a communication connection such as a telephone line or a wireless connection.
- a further aspect of the invention relates to a device for checking a surroundings detection system of a vehicle, wherein the surroundings detection system comprises at least two different environmental sensor types and wherein the device comprises a control device which is configured to detect objects with a first environment sensor type, the objects using the data of categorize the first environment sensor type into static and dynamic objects, to determine the relative position of the detected static objects relative to the vehicle and to compare these with the position determined via a second environment sensor type, wherein a deviation above a predetermined Limit is closed on an error and wherein the field of view of the first environmental sensor type does not overlap with the second environment sensor type vision area and the controller is adapted to update the relative position of the detected static objects after leaving the field of view of the first environmental sensor type taking into account the movement of the vehicle ,
- the first environmental sensor type is a camera or a stereo camera.
- the second environmental sensor type is an ultrasonic sensor. Also conceivable are versions of the second environmental sensor type as radar or lidar-based sensors.
- the device further comprises means for detecting movement of the vehicle, the means being selected from wheel-path sensors, yaw-rate sensors, wheel-speed sensors, acceleration sensors, as well as satellite navigation (e.g., GPS).
- means for detecting movement of the vehicle the means being selected from wheel-path sensors, yaw-rate sensors, wheel-speed sensors, acceleration sensors, as well as satellite navigation (e.g., GPS).
- a method for checking an environment detection system of a vehicle in which the measured values of two different environmental sensor types can be compared with one another without having to overlap the visual ranges of the sensors.
- the proposed method can also be easily integrated into existing systems.
- the method also allows to identify when a sensor fails. Troubleshooting is considerably simplified.
- FIG. 1 shows a vehicle with an environment detection system in a schematic representation.
- a vehicle 10 which comprises a plurality of environmental sensors 12, 14.
- a first environmental sensor type 12 is embodied as a stereo camera 13 in the illustrated embodiment.
- the stereo camera 13 is arranged in the front of the vehicle 10 and aligned in the direction of travel, so that the field of view 18 of the stereo camera 13 also points in the direction of travel. Within the field of view 18, the stereo camera 13 can detect objects in the vicinity of the vehicle 10.
- a second environmental sensor type 14 is arranged on the vehicle 10, which in the embodiment shown is embodied as ultrasonic sensors 15. Two ultrasonic sensors 15 are on each side of the vehicle 10 arranged. The viewing areas 20 of the ultrasonic sensors 15 also point to the respective sides.
- Both sensor types 12, 14 are connected to a control unit 16, which evaluates the measurement data of all sensors and generates a description of the surroundings of the vehicle 10 from these.
- deviating embodiments of the environment sensors 12, 14 used are conceivable.
- a normal optical camera may be used as the first environmental sensor type 12 and a radar or lidar based sensor type as the second environmental sensor type 14. It is also conceivable to use more than two different environmental sensor types.
- FIG. 2 shows a picture of a camera pointing in the direction of travel.
- FIG. 2 an image is shown, which was taken by a stereo camera, which is arranged in the front of a vehicle and aligned in the direction of travel.
- a road 20 with lane markings 28 can be seen.
- To the left side of the road is bounded by crash barriers 24, on the right side are limiting cone 26.
- In the right lane is another road user in the form of another vehicle 30th
- FIG. 3 is the measured and expected for static objects optical flow for the in FIG. 2 illustrated situation shown.
- FIG. 3 shows a diagram in which the optical axis F and on the X-axis, the column C in the camera image are plotted on the Y-axis.
- the optical flow F here is the number of columns by which an object in the camera image has moved from one image to the next camera image.
- a positive value indicates a movement to the left, a negative value a movement to the right in the camera image.
- the camera used here is arranged in a vehicle and aligned in the direction of travel.
- the traffic situation detected by the camera is already at the FIG. 2 been described.
- the optical flux detected by the camera is shown as curve 40 in FIG. 3 located.
- In the left area of the image are crash barriers that have no proper motion.
- the crash barriers are aligned parallel to the road. Pixels that represent parts of the guardrail that are farther away from the camera have a low optical flux and move slowly in the camera image.
- pixels that belong to close parts of the guardrail have a larger optical flux - they move quickly in the camera image.
- the optical flow is determined, Since the other vehicle 30 is traveling, with its speed is only slightly smaller than that of the vehicle 10, it does not appear to move in the camera image. The measured optical flux is almost zero.
- the optical flow measurement curve 40 is interrupted because there is no object in the middle of the road and the horizon is too far away for a measurement. Furthermore, from the FIG. 3 It can be seen that, although the vast majority of the measuring points for the optical flow 40 are on a curve, but occasionally deviate points due to measurement uncertainties, with most measuring errors accumulating around the measured value 0.
- a hypothetical optical flux is calculated for each object on the assumption that the object is static. If the measured optical flux and the hypothetical optical flow agree except for a statistical error, the object is categorized as static.
- the hypothetical optical flow is in FIG. 3 plotted as curve 42.
- the measurement data of the guardrail agree very well with the hypothetical optical flux 42, the guardrail is therefore classified as a static object.
- the optical flow associated with the further vehicle 30 deviates significantly from the hypothetical optical flow 42. The other vehicle is therefore classified as a dynamic object.
- FIG. 4 shows a plan view of the in FIG. 2 illustrated situation.
- FIG. 4 is that out FIG. 2 known traffic situation shown schematically as a plan view.
- the vehicle 10 is located on a road 20.
- the vision area 18 of the first environmental sensor type is oriented to the front, the vision areas 20 of the second sensor type are aligned to the sides of the vehicle 10.
- the vehicle 10 with the crash barrier 24 and the further vehicle 30 can detect two objects in the field of view 18 of the first sensor type.
- the guardrail 24 is classified as a static object and its movement is further tracked relative to the vehicle 10.
- the further vehicle 30 is classified as a dynamic object and ignored when monitoring the environment sensors.
- the limiting cone 26 are also static, but are not here in the field of view 18 of the first sensor type.
- FIG. 5 shows a plan view of the in FIG. 2 presented situation at a later date.
- FIG. 5 will be the same traffic situation as in FIG. 2 and in FIG. 4 but at a later date.
- the vehicle 10 is thereby to overtake the further vehicle 30, so that the further vehicle 30 is located to the right of the vehicle 10 and in the field of view 20 of the second environment sensor type. Since the vehicle 10 has moved away, the guard rail 24, which has already been classified as a static object, is now located to the left of the vehicle 10 and thus also in the field of view 20 of the second environment sensor type.
- the first environment sensor type since the first environment sensor type first detected the barrier 24, its relative position to the vehicle 10 has been constantly updated, with the self-movement of the vehicle 10 being used for those parts which are no longer in the field of vision of the first environment sensor type.
- the relative position of the guardrail 24 determined with the first surround sensor type is now compared to the position determined by the second surround sensor type, with a pass as shown in FIG FIG. 5 is shown, both sensors of the second environmental sensor type measure the position of the guardrail successively.
Landscapes
- Engineering & Computer Science (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Networks & Wireless Communication (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Electromagnetism (AREA)
- Traffic Control Systems (AREA)
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE201310206707 DE102013206707A1 (de) | 2013-04-15 | 2013-04-15 | Verfahren zur Überprüfung eines Umfelderfassungssystems eines Fahrzeugs |
Publications (2)
Publication Number | Publication Date |
---|---|
EP2793045A1 true EP2793045A1 (fr) | 2014-10-22 |
EP2793045B1 EP2793045B1 (fr) | 2019-12-04 |
Family
ID=50488978
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP14163655.5A Active EP2793045B1 (fr) | 2013-04-15 | 2014-04-07 | Procédé de vérification d'un système de détection de l'environnement d'un véhicule |
Country Status (2)
Country | Link |
---|---|
EP (1) | EP2793045B1 (fr) |
DE (1) | DE102013206707A1 (fr) |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2016065375A1 (fr) * | 2014-10-27 | 2016-05-06 | Fts Computertechnik Gmbh | Surveillance de vision artificielle pour système de vision artificielle |
EP3037842A1 (fr) * | 2014-12-26 | 2016-06-29 | Ricoh Company, Ltd. | Système et procédé de mesure et support d'enregistrement non transitoire |
EP3252501A1 (fr) * | 2016-06-03 | 2017-12-06 | Autoliv Development AB | Détection d'objet amélioré et estimation de l'état de mouvement pour un système de détection d'environnement de véhicule |
DE102017210112A1 (de) | 2017-06-16 | 2018-12-20 | Robert Bosch Gmbh | Verfahren und System zur Durchführung einer Kalibrierung eines Sensors |
CN109383518A (zh) * | 2017-08-03 | 2019-02-26 | 通用汽车环球科技运作有限责任公司 | 冗余主动控制***协调 |
CN112304356A (zh) * | 2019-07-31 | 2021-02-02 | 大众汽车股份公司 | 用于检查环境传感器的校准的方法和装置 |
US11003934B2 (en) | 2019-08-30 | 2021-05-11 | Here Global B.V. | Method, apparatus, and system for selecting sensor systems for map feature accuracy and reliability specifications |
WO2021109033A1 (fr) | 2019-12-04 | 2021-06-10 | Huawei Technologies Co., Ltd. | Appareil et procédé de collecte et de marquage automatique de données de mesure dans un scénario de trafic |
CN112955775A (zh) * | 2018-10-30 | 2021-06-11 | 戴姆勒公司 | 用于检查车辆的至少一个环境检测传感器的方法 |
CN114216483A (zh) * | 2021-12-14 | 2022-03-22 | 北京云迹科技股份有限公司 | 一种机器人的检测方法及装置 |
EP4296715A1 (fr) * | 2022-06-20 | 2023-12-27 | Robert Bosch GmbH | Procédé de détermination d'une position approximative d'un objet dynamique, programme informatique, dispositif et véhicule |
Families Citing this family (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE102015104965A1 (de) * | 2015-03-31 | 2016-10-06 | Valeo Schalter Und Sensoren Gmbh | Fahrerassistenzsystem zur Seitenüberwachung eines Kraftfahrzeugs, Kraftfahrzeug und Verfahren |
DE102015216494A1 (de) * | 2015-08-28 | 2017-03-02 | Robert Bosch Gmbh | Verfahren und Vorrichtung zur Erkennung wenigstens einer Sensorfunktion wenigstens eines ersten Sensor wenigstens eines ersten Fahrzeugs |
JP6569504B2 (ja) * | 2015-12-09 | 2019-09-04 | 株式会社デンソー | 物体認識装置 |
DE102016218213A1 (de) * | 2016-09-22 | 2018-03-22 | Zf Friedrichshafen Ag | Klassifizierung statischer Objekte mit beweglicher Kamera |
DE102016223107A1 (de) | 2016-11-23 | 2018-05-24 | Volkswagen Aktiengesellschaft | System und Verfahren zur Schätzung der Eigenbewegung eines Fahrzeuges |
US10452067B2 (en) | 2017-02-23 | 2019-10-22 | GM Global Technology Operations LLC | System and method for detecting improper sensor installation within a vehicle to mitigate hazards associated with object detection |
DE102017111318B4 (de) | 2017-08-03 | 2019-03-07 | Valeo Schalter Und Sensoren Gmbh | Ultraschallsensoreinheit mit Ultraschallsensor und Infrarotsensor |
DE102018202766A1 (de) * | 2018-02-23 | 2019-08-29 | Siemens Aktiengesellschaft | Verfahren und System zur Selbstdiagnose mindestens eines Sensors eines eine autonom bewegbare oder bewegliche Vorrichtung unterstützenden Systems |
DE102018128535B4 (de) * | 2018-11-14 | 2021-09-02 | Valeo Schalter Und Sensoren Gmbh | Trainieren eines künstlichen neuronalen Netzwerks mit durch Sensoren eines Fahrzeugs erfassten Daten |
DE102018221427B4 (de) * | 2018-12-11 | 2020-08-06 | Volkswagen Aktiengesellschaft | Verfahren zur Ermittlung einer vorliegenden Dejustage wenigstens eines Sensors innerhalb eines Sensorverbundes |
DE102019002487A1 (de) * | 2019-04-04 | 2020-10-08 | Daimler Ag | Verfahren zur Überprüfung eines Umfelderfassungssensors eines Fahrzeugs und Verfahren zum Betrieb eines Fahrzeugs |
DE102019212021B4 (de) | 2019-08-09 | 2024-02-08 | Volkswagen Aktiengesellschaft | Verfahren und Vorrichtung zum Feststellen eines Parallaxenproblems in Sensordaten zweier Sensoren |
DE102019134985B4 (de) | 2019-12-18 | 2022-06-09 | S.M.S, Smart Microwave Sensors Gmbh | Verfahren zum Erfassen wenigstens eines Verkehrsteilnehmers |
DE102020203633A1 (de) | 2020-03-20 | 2021-09-23 | Robert Bosch Gesellschaft mit beschränkter Haftung | Verfahren und Vorrichtung zur Bestimmung der Lufttemperatur und Computerprogrammprodukt |
DE102020117870B4 (de) | 2020-07-07 | 2023-01-12 | Dr. Ing. H.C. F. Porsche Aktiengesellschaft | Fahrzeug |
JP7213279B2 (ja) | 2021-02-03 | 2023-01-26 | 本田技研工業株式会社 | 運転支援装置 |
DE102021107793A1 (de) | 2021-03-29 | 2022-09-29 | Ford Global Technologies, Llc | Verfahren zum Betreiben einer Fahrerassistenzfunktion |
DE102022101737A1 (de) | 2022-01-26 | 2023-07-27 | Bayerische Motoren Werke Aktiengesellschaft | System und Verfahren zur Objekterkennung |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE10229336A1 (de) | 2002-06-29 | 2004-01-15 | Robert Bosch Gmbh | Verfahren und Vorrichtung zur Kalibrierung von Bildsensorsystemen |
WO2004072901A1 (fr) | 2003-02-05 | 2004-08-26 | Siemens Corporate Research Inc. | Detection d'obstacles en temps reel au moyen d'une camera etalonnee et d'un mouvement de camera connu |
EP1462820A1 (fr) * | 2003-03-24 | 2004-09-29 | DaimlerChrysler AG | Indication visuelle dans un système de collecte de données pour véhicule |
EP2101193A1 (fr) * | 2008-03-10 | 2009-09-16 | Sick Ag | Système de sécurité destiné à la mesure sans contact de positions, de voies et de vitesses |
DE102008038731A1 (de) * | 2008-08-12 | 2010-02-18 | Continental Automotive Gmbh | Verfahren zur Erkennung ausgedehnter statischer Objekte |
US20110116717A1 (en) * | 2009-11-17 | 2011-05-19 | Mando Corporation | Method and system for recognizing parking lot |
US20110133917A1 (en) * | 2009-12-03 | 2011-06-09 | Gm Global Technology Operations, Inc. | Cross traffic collision alert system |
DE102010051602A1 (de) | 2010-11-16 | 2012-05-16 | Hans-Heinrich Götting jun. | Verfahren zur Erfassung von Merkmalen in der Umgebung |
-
2013
- 2013-04-15 DE DE201310206707 patent/DE102013206707A1/de not_active Withdrawn
-
2014
- 2014-04-07 EP EP14163655.5A patent/EP2793045B1/fr active Active
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE10229336A1 (de) | 2002-06-29 | 2004-01-15 | Robert Bosch Gmbh | Verfahren und Vorrichtung zur Kalibrierung von Bildsensorsystemen |
WO2004072901A1 (fr) | 2003-02-05 | 2004-08-26 | Siemens Corporate Research Inc. | Detection d'obstacles en temps reel au moyen d'une camera etalonnee et d'un mouvement de camera connu |
EP1462820A1 (fr) * | 2003-03-24 | 2004-09-29 | DaimlerChrysler AG | Indication visuelle dans un système de collecte de données pour véhicule |
EP2101193A1 (fr) * | 2008-03-10 | 2009-09-16 | Sick Ag | Système de sécurité destiné à la mesure sans contact de positions, de voies et de vitesses |
DE102008038731A1 (de) * | 2008-08-12 | 2010-02-18 | Continental Automotive Gmbh | Verfahren zur Erkennung ausgedehnter statischer Objekte |
US20110116717A1 (en) * | 2009-11-17 | 2011-05-19 | Mando Corporation | Method and system for recognizing parking lot |
US20110133917A1 (en) * | 2009-12-03 | 2011-06-09 | Gm Global Technology Operations, Inc. | Cross traffic collision alert system |
DE102010051602A1 (de) | 2010-11-16 | 2012-05-16 | Hans-Heinrich Götting jun. | Verfahren zur Erfassung von Merkmalen in der Umgebung |
Non-Patent Citations (1)
Title |
---|
STILLER C ET AL: "MULTISENSOR OBSTACLE DETECTION AND TRACKING", IMAGE AND VISION COMPUTING, ELSEVIER, GUILDFORD, GB, vol. 18, no. 5, April 2000 (2000-04-01), pages 389 - 396, XP001179983, ISSN: 0262-8856, DOI: 10.1016/S0262-8856(99)00034-7 * |
Cited By (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170305438A1 (en) * | 2014-10-27 | 2017-10-26 | Fts Computertechnik Gmbh | Computer vision monitoring for a computer vision system |
WO2016065375A1 (fr) * | 2014-10-27 | 2016-05-06 | Fts Computertechnik Gmbh | Surveillance de vision artificielle pour système de vision artificielle |
EP3037842A1 (fr) * | 2014-12-26 | 2016-06-29 | Ricoh Company, Ltd. | Système et procédé de mesure et support d'enregistrement non transitoire |
JP2016125843A (ja) * | 2014-12-26 | 2016-07-11 | 株式会社リコー | 移動体、測定システム、測定方法及びプログラム |
US10151840B2 (en) | 2014-12-26 | 2018-12-11 | Ricoh Company, Ltd. | Measuring system, measuring process, and non-transitory recording medium |
EP3252501A1 (fr) * | 2016-06-03 | 2017-12-06 | Autoliv Development AB | Détection d'objet amélioré et estimation de l'état de mouvement pour un système de détection d'environnement de véhicule |
WO2017207727A1 (fr) * | 2016-06-03 | 2017-12-07 | Autoliv Development Ab | Détection améliorée d'objet et estimation de l'état de mouvement pour un système de détection d'environnement de véhicule |
CN109313258A (zh) * | 2016-06-03 | 2019-02-05 | 奥托立夫开发公司 | 用于车辆环境检测***的增强的物体检测和运动状态估计 |
JP2019526781A (ja) * | 2016-06-03 | 2019-09-19 | ヴィオニア スウェーデン エービー | 車両環境検知システム用に向上された物体検出及び運動状態推定 |
US10698084B2 (en) | 2017-06-16 | 2020-06-30 | Robert Bosch Gmbh | Method and system for carrying out a calibration of a sensor |
DE102017210112A1 (de) | 2017-06-16 | 2018-12-20 | Robert Bosch Gmbh | Verfahren und System zur Durchführung einer Kalibrierung eines Sensors |
JP2019002929A (ja) * | 2017-06-16 | 2019-01-10 | ロベルト・ボッシュ・ゲゼルシャフト・ミト・ベシュレンクテル・ハフツングRobert Bosch Gmbh | センサの校正の実行方法およびシステム |
CN109383518A (zh) * | 2017-08-03 | 2019-02-26 | 通用汽车环球科技运作有限责任公司 | 冗余主动控制***协调 |
CN112955775A (zh) * | 2018-10-30 | 2021-06-11 | 戴姆勒公司 | 用于检查车辆的至少一个环境检测传感器的方法 |
CN112955775B (zh) * | 2018-10-30 | 2023-12-22 | 戴姆勒公司 | 用于检查车辆的至少一个环境检测传感器的方法 |
CN112304356A (zh) * | 2019-07-31 | 2021-02-02 | 大众汽车股份公司 | 用于检查环境传感器的校准的方法和装置 |
US11645782B2 (en) | 2019-07-31 | 2023-05-09 | Volkswagen Aktiengesellschaft | Method and device for checking a calibration of environment sensors |
US11003934B2 (en) | 2019-08-30 | 2021-05-11 | Here Global B.V. | Method, apparatus, and system for selecting sensor systems for map feature accuracy and reliability specifications |
WO2021109033A1 (fr) | 2019-12-04 | 2021-06-10 | Huawei Technologies Co., Ltd. | Appareil et procédé de collecte et de marquage automatique de données de mesure dans un scénario de trafic |
CN113330331A (zh) * | 2019-12-04 | 2021-08-31 | 华为技术有限公司 | 用于交通场景中收集和自动标记测量数据的装置和方法 |
EP4058825A4 (fr) * | 2019-12-04 | 2023-01-04 | Huawei Technologies Co., Ltd. | Appareil et procédé de collecte et de marquage automatique de données de mesure dans un scénario de trafic |
CN114216483A (zh) * | 2021-12-14 | 2022-03-22 | 北京云迹科技股份有限公司 | 一种机器人的检测方法及装置 |
EP4296715A1 (fr) * | 2022-06-20 | 2023-12-27 | Robert Bosch GmbH | Procédé de détermination d'une position approximative d'un objet dynamique, programme informatique, dispositif et véhicule |
Also Published As
Publication number | Publication date |
---|---|
DE102013206707A1 (de) | 2014-10-16 |
EP2793045B1 (fr) | 2019-12-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2793045B1 (fr) | Procédé de vérification d'un système de détection de l'environnement d'un véhicule | |
DE102016219455B4 (de) | Verfahren und aktive Infrastruktur zur Überprüfung einer Umfelderfassung eines Kraftfahrzeugs | |
DE102017105305A1 (de) | Verfahren zur automatischen bestimmung einer sensorstellung | |
EP1690730B1 (fr) | Système d'assistance au conducteur avec unité de décision rdondante | |
DE102014111127A1 (de) | Verfahren zum Erzeugen einer Umgebungskarte sowie Fahrerassistenzsystem | |
EP3714286B1 (fr) | Procédé et dispositif de détermination d'un angle d'installation entre une piste sur laquelle un véhicule circule et une direction de mesure d'un capteur de mesure ou radar | |
EP1159638A1 (fr) | Procede et dispositifs de detection de desalignement dans un systeme radar pour vehicules ou un systeme capteur pour vehicules | |
DE10149115A1 (de) | Objekterfassungsvorrichtung | |
DE102019206021A1 (de) | Verfahren zur Erkennung einer Funktionsfähigkeit eines Umgebungssensors, Steuergerät und Fahrzeug | |
WO2016177727A1 (fr) | Procédé de diagnostic pour un capteur de vision d'un véhicule et véhicule muni d'un capteur de vision | |
DE102018103803A1 (de) | System und verfahren für das erkennen einer fehlerhaften sensorinstallation innerhalb eines fahrzeugs, um die mit der objekterkennung verbundenen gefahren zu verringern | |
DE102007020264A1 (de) | Verfahren und Messvorrichtung zum Bestimmen einer Relativgeschwindigkeit | |
DE102014201158A1 (de) | Verfahren und Vorrichtung zum Überprüfen eines von einer Objekterkennung erkannten relevanten Objekts | |
WO2022012923A1 (fr) | Procédé d'implantation d'un ordinateur pour détermination de la validité d'une estimation de position d'un véhicule | |
DE102019102769A1 (de) | Verfahren und eine Vorrichtung zur Sensordatenfusion für ein Fahrzeug | |
EP4139709A1 (fr) | Procédé et dispositif d'identification de l'efflorescence dans une mesure lidar | |
DE10326431A1 (de) | Vorrichtung und Verfahren zur Bestimmung der Position von Objekten im Umfeld eines Fahrzeuges | |
DE102016109850B4 (de) | Verfahren zum Erkennen einer Neigung in einer Fahrbahn eines Kraftfahrzeugs, Fahrerassistenzsystem sowie Kraftfahrzeug | |
DE102015205087A1 (de) | Verfahren zum Ermitteln der Dejustage eines Fahrersassistenzsensors | |
DE102009052560A1 (de) | Verfahren und Vorrichtung zur Bestimmung der 3D-Position von Objekten im Umfeld eines Kraftfahrzeugs | |
DE102021205804A1 (de) | Verfahren zum Überwachen mindestens eines Sensors | |
EP3658953B1 (fr) | Dispositif et procédé de reconnaissance de la hauteur d'un objet | |
DE102020125448A1 (de) | Kartenplausibilitätsprüfungsverfahren | |
DE102022201199B3 (de) | Verfahren und Vorrichtung zur Überwachung eines Spurerkennungssystems, Spurerkennungssystem | |
DE102016006381B3 (de) | Verfahren zum Betreiben einer Verarbeitungseinrichtung für Sensordaten eines in einem Kraftfahrzeug angeordneten Sensors und Kraftfahrzeug |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
17P | Request for examination filed |
Effective date: 20140407 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
AX | Request for extension of the european patent |
Extension state: BA ME |
|
R17P | Request for examination filed (corrected) |
Effective date: 20150422 |
|
RBV | Designated contracting states (corrected) |
Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
GRAP | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOSNIGR1 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: GRANT OF PATENT IS INTENDED |
|
RIC1 | Information provided on ipc code assigned before grant |
Ipc: G01S 11/12 20060101AFI20190731BHEP Ipc: G06K 9/00 20060101ALI20190731BHEP Ipc: G01S 17/02 20060101ALI20190731BHEP |
|
INTG | Intention to grant announced |
Effective date: 20190823 |
|
GRAS | Grant fee paid |
Free format text: ORIGINAL CODE: EPIDOSNIGR3 |
|
GRAA | (expected) grant |
Free format text: ORIGINAL CODE: 0009210 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE PATENT HAS BEEN GRANTED |
|
AK | Designated contracting states |
Kind code of ref document: B1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: FG4D Free format text: NOT ENGLISH |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: EP |
|
REG | Reference to a national code |
Ref country code: AT Ref legal event code: REF Ref document number: 1210069 Country of ref document: AT Kind code of ref document: T Effective date: 20191215 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R096 Ref document number: 502014013162 Country of ref document: DE |
|
REG | Reference to a national code |
Ref country code: IE Ref legal event code: FG4D Free format text: LANGUAGE OF EP DOCUMENT: GERMAN |
|
REG | Reference to a national code |
Ref country code: NL Ref legal event code: MP Effective date: 20191204 |
|
REG | Reference to a national code |
Ref country code: LT Ref legal event code: MG4D |
|
RAP2 | Party data changed (patent owner data changed or rights of a patent transferred) |
Owner name: ROBERT BOSCH GMBH |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 Ref country code: LV Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 Ref country code: BG Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20200304 Ref country code: FI Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 Ref country code: NO Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20200304 Ref country code: GR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20200305 Ref country code: ES Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 Ref country code: SE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: RS Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 Ref country code: HR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: AL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: NL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 Ref country code: EE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 Ref country code: PT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20200429 Ref country code: RO Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 Ref country code: CZ Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: IS Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20200404 Ref country code: SM Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 Ref country code: SK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R097 Ref document number: 502014013162 Country of ref document: DE |
|
PLBE | No opposition filed within time limit |
Free format text: ORIGINAL CODE: 0009261 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: DK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 |
|
26N | No opposition filed |
Effective date: 20200907 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: SI Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 Ref country code: PL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 Ref country code: MC Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: PL |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LI Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20200430 Ref country code: CH Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20200430 Ref country code: LU Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20200407 |
|
REG | Reference to a national code |
Ref country code: BE Ref legal event code: MM Effective date: 20200430 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: BE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20200430 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: IE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20200407 |
|
REG | Reference to a national code |
Ref country code: AT Ref legal event code: MM01 Ref document number: 1210069 Country of ref document: AT Kind code of ref document: T Effective date: 20200407 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: AT Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20200407 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: TR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 Ref country code: MT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 Ref country code: CY Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: MK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20191204 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: IT Payment date: 20230428 Year of fee payment: 10 Ref country code: FR Payment date: 20230417 Year of fee payment: 10 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: GB Payment date: 20230420 Year of fee payment: 10 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: DE Payment date: 20240619 Year of fee payment: 11 |