WO2021136643A1 - Verfahren zum unterdrücken von mehrdeutigen messdaten von umfeldsensoren - Google Patents
Verfahren zum unterdrücken von mehrdeutigen messdaten von umfeldsensoren Download PDFInfo
- Publication number
- WO2021136643A1 WO2021136643A1 PCT/EP2020/085902 EP2020085902W WO2021136643A1 WO 2021136643 A1 WO2021136643 A1 WO 2021136643A1 EP 2020085902 W EP2020085902 W EP 2020085902W WO 2021136643 A1 WO2021136643 A1 WO 2021136643A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- hypotheses
- sensor
- measurement data
- hypothesis
- control device
- Prior art date
Links
- 238000005259 measurement Methods 0.000 title claims abstract description 48
- 238000000034 method Methods 0.000 title claims abstract description 45
- 230000003068 static effect Effects 0.000 claims abstract description 29
- 238000001514 detection method Methods 0.000 claims abstract description 10
- 238000004590 computer program Methods 0.000 claims abstract description 8
- 230000007613 environmental effect Effects 0.000 description 3
- 230000001133 acceleration Effects 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 2
- 230000004927 fusion Effects 0.000 description 2
- 230000004807 localization Effects 0.000 description 2
- 230000001419 dependent effect Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000008030 elimination Effects 0.000 description 1
- 238000003379 elimination reaction Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 230000001629 suppression Effects 0.000 description 1
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W40/00—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models
- B60W40/10—Estimation or calculation of non-directly measurable driving parameters for road vehicle drive control systems not related to the control of a particular sub unit, e.g. by using mathematical models related to vehicle motion
- B60W40/105—Speed
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B29/00—Checking or monitoring of signalling or alarm systems; Prevention or correction of operating errors, e.g. preventing unauthorised operation
- G08B29/18—Prevention or correction of operating errors
- G08B29/185—Signal analysis techniques for reducing or preventing false alarms or for enhancing the reliability of the system
- G08B29/188—Data fusion; cooperative systems, e.g. voting among different detectors
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/02—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S13/00
- G01S7/41—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S13/00 using analysis of echo signal for target characterisation; Target signature; Target cross-section
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/02—Systems using reflection of radio waves, e.g. primary radar systems; Analogous systems
- G01S13/50—Systems of measurement based on relative movement of target
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/86—Combinations of radar systems with non-radar systems, e.g. sonar, direction finder
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/88—Radar or analogous systems specially adapted for specific applications
- G01S13/93—Radar or analogous systems specially adapted for specific applications for anti-collision purposes
- G01S13/931—Radar or analogous systems specially adapted for specific applications for anti-collision purposes of land vehicles
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2420/00—Indexing codes relating to the type of sensors based on the principle of their operation
- B60W2420/40—Photo, light or radio wave sensitive means, e.g. infrared sensors
- B60W2420/403—Image sensing, e.g. optical camera
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2420/00—Indexing codes relating to the type of sensors based on the principle of their operation
- B60W2420/40—Photo, light or radio wave sensitive means, e.g. infrared sensors
- B60W2420/408—Radar; Laser, e.g. lidar
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2554/00—Input parameters relating to objects
- B60W2554/20—Static objects
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2554/00—Input parameters relating to objects
- B60W2554/40—Dynamic objects, e.g. animals, windblown objects
- B60W2554/404—Characteristics
- B60W2554/4041—Position
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60W—CONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
- B60W2554/00—Input parameters relating to objects
- B60W2554/40—Dynamic objects, e.g. animals, windblown objects
- B60W2554/404—Characteristics
- B60W2554/4049—Relationship among other objects, e.g. converging dynamic objects
Definitions
- the invention relates to a method for eliminating sensor errors, in particular ambiguities in the detection of dynamic objects, and a method for eliminating sensor errors, in particular ambiguities in the detection of static objects.
- the invention also relates to a control device, a computer program and a machine-readable storage medium.
- environment sensors such as radar sensors, LIDAR sensors or camera sensors
- Other road users or dynamic objects as well as static objects in the vehicle environment can be detected and measured with environment sensors in order to enable an automated reaction of the driving function.
- the currently available environment sensors and the evaluation of the corresponding measurement data can, however, result in incorrect results, such as false-positive or false-negative results, which can cause incomprehensible or dangerous reactions of the automated driving function.
- incorrect results can be suppressed if necessary, since the supporting driving function can be deactivated and the vehicle control can be handed over to a driver at any time.
- automatic emergency braking cannot be carried out.
- Such a suppression of erroneous results is at Automated and, in particular, driverless driving functions are not possible due to the lack of a driver.
- the ghost targets can arise in particular from ambiguities and can be measured, for example, at an angle of, for example, 0 ° and 30 °. If the wrong decision is made in signal processing, this can lead to false detection and thus to security issues.
- the object on which the invention is based can be seen in proposing a method for reducing error rates, in particular false-positive and false-negative rates, of an environment sensor.
- a method for eliminating sensor errors, in particular ambiguities when detecting dynamic objects is provided by a control device.
- the ambiguities can in particular result from false-positive results and thus cause several detected objects or object hypotheses.
- Dynamic objects are preferably moving road users.
- the objects can, for example, be measurable in the surroundings of a mobile unit, the control device and the sensors being installed, for example, on the mobile unit.
- measurement data are received from at least one first sensor and object hypotheses are formed from the measurement data received.
- the at least one first sensor can be, for example, a radar sensor, a LIDAR sensor, ultrasonic sensor and the like.
- determined ambiguous object hypotheses for example in the form of angle hypotheses, are determined or received.
- objects measured in real terms can provide ambiguous angle measurements in which only one of the angle hypotheses is correct.
- a further step data of at least one reference object is received, which is detected on the basis of measurement data from at least one second sensor.
- a so-called reference object method can be used to eliminate ambiguities.
- data from other environmental sensors or the at least one second sensor are used. If an object is detected stably and unambiguously by the second sensor, this object can be specified as a reference object.
- the object hypotheses formed are then compared with the at least one detected reference object. Object hypotheses that do not match the detected reference object are preferably discarded. Using a first sensor configured, for example, as a radar sensor, such an object hypothesis or angle hypothesis can be selected which, in particular, corresponds locally to the reference object. All other object hypotheses are discarded and can therefore no longer lead to ghost goals.
- the method allows additional information from further sensors to be used to correct sensor errors, such as ambiguities, at an early stage. In this way, the occurrence of both false positive and false negative results or object hypotheses can be avoided or at least reduced.
- the control device and the sensors can be arranged in a mobile unit which, according to the BASt standard, can be assisted, partially automated, highly automated and / or fully automated or operated without a driver.
- the mobile unit can be configured as a vehicle, a robot, a drone, a watercraft, a rail vehicle, a robotic taxi, an industrial robot, a utility vehicle, a bus, an airplane, a helicopter and the like
- the method can be implemented without high technical effort.
- the reliability of the measurement data provided can be increased without additional costs, since the necessary sensors, such as LIDAR sensors or radar sensors, are installed in an increasing number of mobile units.
- a vehicle following the method or within the scope of the method, can be controlled based on the corrected or adjusted measurement data from the sensors. By eliminating sensor errors from the measurement data, safety can be increased for all road users involved.
- a method for eliminating sensor errors, in particular ambiguities when detecting static objects is provided by a control device.
- Static objects are preferably immobile objects.
- static objects can be designed in the form of parked vehicles, trees, buildings and the like.
- measurement data are received from at least one first sensor and object hypotheses are formed from the measurement data received.
- Static real measured objects can also deliver ambiguous object hypotheses or angle hypotheses in which only one of the object hypotheses is correct and the remaining object hypotheses can be traced back to ambiguities.
- a so-called stationary target acceptance can thus take place.
- the speed of the first sensor or the mobile unit on which the first sensor is attached must be known. Based on the speed speeds of the generated object hypotheses can be calculated. The speeds can preferably be calculated as absolute speeds.
- At least one object hypothesis depicts a static object based on the calculated speed. Subsequently, in the case of at least one determined object hypothesis depicting a static object, all other object hypotheses are discarded. Static objects can also generate ambiguous object hypotheses or, in the case of a radar measurement, angle hypotheses. If the wrong object hypothesis is selected, real static objects can be assigned incorrect speeds and the objects can be classified as dynamic. Moving or dynamic objects have a high relevance for the driving functions, since they are usually other road users. Moving object hypotheses caused by stationary targets or static objects are particularly critical for this reason, since both the position and the speed can be incorrect. It is therefore advantageous to accept a determined object hypothesis or location as a stationary target if one of the object hypotheses or angle hypotheses speak for it. This assumption can be made because a stationary target hypothesis or static object hypothesis is very unlikely in the case of a dynamic object.
- a reference object can also be used here to eliminate ambiguities.
- the method can also be supported by the method for eliminating sensor errors, in particular ambiguities in the detection of dynamic objects. If, for example, there is a reference object, an object hypothesis superimposed on the reference object can be selected and all other object hypotheses can be discarded.
- the two methods according to the invention can be used, for example, when recognizing landmarks for landmark-based vehicle localization.
- a control device is provided, the control device being set up to carry out the method.
- the control device can be, for example, a vehicle-side control device, a vehicle-external control device or a vehicle-external server unit, such as a cloud system.
- the control device can preferably receive and process measurement data from the at least one measurement antenna and / or measurement data from sensors in the at least one mobile unit.
- a computer program which comprises commands which, when the computer program is executed by a computer or a control device, cause the computer or a control device to execute the method according to the invention.
- a machine-readable storage medium is provided on which the computer program according to the invention is stored.
- rejected object hypotheses are marked as “faulty” and / or positions at which the rejected object hypotheses are determined are marked as “unreliable”.
- This information and markings can be stored in a central server unit or in the control unit and made available to other participants and mobile units. In this way, the rejected object hypotheses can be used to mark other, nearby locations or object hypotheses as "unreliable”. These unreliable object hypotheses can then be treated more restrictively in a downstream object tracking or rejected.
- Such a procedure makes sense because a reference object is not always available to correctly resolve all ambiguities. For example, if more than one object hypothesis is confirmed by a reference object, the most likely object hypothesis is selected and the less likely object hypothesis is discarded.
- the object hypotheses are formed as angle hypotheses from measurement data from at least one radar sensor.
- the at least one reference object is determined from measurement data from at least one second sensor which differs from the at least one first sensor.
- the reference object can be determined by evaluating measurement data from a second sensor, which is designed as an additional radar sensor, LIDAR sensor, camera sensor, ultrasonic sensor and the like. The elimination of ambiguities can be particularly robust based on measurements of different sensor types or sensor classes.
- the at least one object hypothesis maps a static object if a lower speed than a limit value is calculated for the object hypothesis.
- a limit value For example, in the case of an angle measurement by a radar sensor, the following condition, which is simplified for driving straight ahead, can be checked with regard to each of the angle hypotheses: v r + Vego * COS (thetai)
- v r corresponds to a measured relative speed
- v ego corresponds to the determined speed of the mobile unit with the first sensor, theta, to the determined angle of an angle hypothesis, and t to the limit value or the threshold. If a speed assigned to the object hypothesis is determined to be below the limit value of, for example, 1 m / s, then the object hypothesis is defined as an object hypothesis of a static object.
- the corresponding probability for the object hypotheses is calculated for at least two object hypotheses each depicting a static object, the object hypothesis with the lower probability being discarded. In this way, additional ambiguities can be eliminated particularly efficiently.
- all object hypotheses except at least one object hypothesis depicting a static object are discarded if no reference object is received. This allows a Technically simple decision-making can be implemented, in which a stationary target hypothesis or an object hypothesis depicting a static object is selected if there are no reference objects. All other object hypotheses are rejected here.
- FIG. 1 shows a schematic traffic situation with a dynamic object to illustrate a method according to one embodiment.
- FIG. 2 shows a schematic traffic situation with static objects to illustrate a method according to a further embodiment
- FIG. 3 shows a schematic flow diagram to illustrate a method.
- FIG. 1 shows a schematic traffic situation 1 with a dynamic object 2 to illustrate a method according to one embodiment.
- the method is used in particular to eliminate sensor errors, such as, for example, ambiguities in the detection of dynamic objects 2, by a control device 4.
- the control device 4 is installed in a mobile unit 6, which is designed as a motor vehicle that can be operated automatically.
- the mobile unit 6 has a first sensor 8 and a second sensor 10.
- the first sensor 8 is configured, for example, as a radar sensor and the second sensor 10 as a LIDAR sensor.
- the control device 4 can receive and evaluate measurement data from the sensors 8, 10.
- the control device 4 is connected to the sensors 8, 10 in a data-conducting manner.
- information from the second sensor 10 can be used.
- a stably determined reference object 11 and in particular a position of the reference object 11 can be used to confirm one of the two object hypotheses 12.
- FIG. 2 shows a schematic traffic situation 1 with static objects 3 to illustrate a method according to a further embodiment.
- the method is used to eliminate sensor errors, in particular ambiguities in the detection of static objects 3.
- measurement data from the first sensor 8 are evaluated and a number of object hypotheses 12, 14 are formed. There are no reference objects 11 that can be used by the control device 4.
- Moving objects 2 are highly relevant for the driving function, since they are normally other road users.
- Stationary targets 3 classified as moving objects 2 are particularly critical since both their position and their speed are incorrect. It is therefore advantageous to assume a location as a stationary target or as a static object 3 if one of the object hypotheses 12 or angle hypotheses speak for it. All other object hypotheses 14 are discarded.
- FIG. 3 shows a schematic flow chart to illustrate a method according to a further embodiment.
- Measurement data are received from at least one first sensor 8.
- the first sensor 8 can be a radar sensor, for example.
- Angle hypotheses are also formed and transmitted with the measurement data.
- measurement data from an inertial measurement unit 13 can be received.
- the measurement data can include, for example, a speed, acceleration values and yaw rates of the vehicle 6.
- an object hypothesis is selected 20 and ambiguities are thus eliminated.
- the selection 20 of an object hypothesis can be implemented by one of the methods according to the invention, so that only correct object hypotheses 12 are forwarded for further processing, such as object tracking and measurement data fusion 22, for example.
- Measurement data from a second sensor 10, such as a LIDAR sensor 10, for example, can be used both for the sensor data fusion 22 and for the selection 20 of the object hypothesis in order to provide reference objects 11, for example.
- the merged measurement data can then be used to implement driving functions 24.
- the driving function 24 can have direct or indirect access to a vehicle actuator 26, such as braking functions, acceleration functions and steering functions.
- a vehicle actuator 26 such as braking functions, acceleration functions and steering functions.
- the data determined and forwarded by the selection 20 of the object hypothesis can also be used for a landmark-based localization.
Landscapes
- Engineering & Computer Science (AREA)
- Remote Sensing (AREA)
- Radar, Positioning & Navigation (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Networks & Wireless Communication (AREA)
- Electromagnetism (AREA)
- Automation & Control Theory (AREA)
- Mathematical Physics (AREA)
- Transportation (AREA)
- Mechanical Engineering (AREA)
- Computer Security & Cryptography (AREA)
- Radar Systems Or Details Thereof (AREA)
- Traffic Control Systems (AREA)
Abstract
Description
Claims
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020227022251A KR20220104250A (ko) | 2019-12-30 | 2020-12-14 | 환경 인지 센서의 모호한 측정 데이터 억제 방법 |
JP2022540797A JP7451718B2 (ja) | 2019-12-30 | 2020-12-14 | 環境センサーからの曖昧な測定データを排除する方法 |
CN202080091202.5A CN114902071A (zh) | 2019-12-30 | 2020-12-14 | 用于抑制环境传感器的不确定性测量数据的方法 |
US17/789,832 US20230059090A1 (en) | 2019-12-30 | 2020-12-14 | Method for suppressing ambiguous measurement data from environmental sensors |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE102019135851.5A DE102019135851A1 (de) | 2019-12-30 | 2019-12-30 | Verfahren zum Unterdrücken von mehrdeutigen Messdaten von Umfeldsensoren |
DE102019135851.5 | 2019-12-30 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021136643A1 true WO2021136643A1 (de) | 2021-07-08 |
Family
ID=74104062
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/EP2020/085902 WO2021136643A1 (de) | 2019-12-30 | 2020-12-14 | Verfahren zum unterdrücken von mehrdeutigen messdaten von umfeldsensoren |
Country Status (6)
Country | Link |
---|---|
US (1) | US20230059090A1 (de) |
JP (1) | JP7451718B2 (de) |
KR (1) | KR20220104250A (de) |
CN (1) | CN114902071A (de) |
DE (1) | DE102019135851A1 (de) |
WO (1) | WO2021136643A1 (de) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11709260B2 (en) * | 2021-04-30 | 2023-07-25 | Zoox, Inc. | Data driven resolution function derivation |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE102008013366A1 (de) * | 2008-03-10 | 2009-09-17 | Bayerische Motoren Werke Aktiengesellschaft | Verfahren zur Bereitstellung von Information für Fahrerassistenzsysteme |
US20150009031A1 (en) * | 2013-07-03 | 2015-01-08 | Honeywell International Inc. | Multilayer perimeter instrusion detection system for multi-processor sensing |
Family Cites Families (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2563509B2 (ja) * | 1988-08-29 | 1996-12-11 | 沖電気工業株式会社 | レーダの偽像除去装置 |
JP2000230974A (ja) | 1999-02-09 | 2000-08-22 | Toyota Motor Corp | レーダ装置 |
JP2002098754A (ja) | 2000-09-22 | 2002-04-05 | Fujitsu Ten Ltd | 車両用レーダ装置 |
JP2004117071A (ja) | 2002-09-24 | 2004-04-15 | Fuji Heavy Ind Ltd | 車外監視装置、及び、この車外監視装置を備えた走行制御装置 |
JP2004333282A (ja) * | 2003-05-07 | 2004-11-25 | Optex Co Ltd | マイクロウエーブセンサ |
JP4082442B2 (ja) * | 2007-07-24 | 2008-04-30 | 株式会社デンソー | 方位検出装置 |
CN103026395A (zh) * | 2010-11-15 | 2013-04-03 | 图像传感系统有限公司 | 混合交通传感器系统和相关的方法 |
JP5835243B2 (ja) | 2013-02-07 | 2015-12-24 | 株式会社デンソー | 物標認識装置 |
DE102016205139B4 (de) * | 2015-09-29 | 2022-10-27 | Volkswagen Aktiengesellschaft | Vorrichtung und Verfahren zur Charakterisierung von Objekten |
DE102016212326A1 (de) * | 2016-07-06 | 2018-01-11 | Robert Bosch Gmbh | Verfahren zur Verarbeitung von Sensordaten für eine Position und/oder Orientierung eines Fahrzeugs |
US10151830B2 (en) * | 2016-09-14 | 2018-12-11 | Caterpillar Inc. | Systems and methods for detecting objects proximate to a machine utilizing a learned process |
JP6548708B2 (ja) * | 2017-11-21 | 2019-07-24 | ディスペース デジタル シグナル プロセッシング アンド コントロール エンジニアリング ゲゼルシャフト ミット ベシュレンクテル ハフツングdspace digital signal processing and control engineering GmbH | 画像処理システムのための低レイテンシの試験機 |
DE102018200011A1 (de) * | 2018-01-02 | 2019-07-04 | Ford Global Technologies, Llc | Testsystem und Verfahren zum Testen einer Steuerung eines zumindest teilweise autonom fahrenden Fahrzeugs in einer virtuellen Umgebung |
US11102492B2 (en) * | 2018-02-20 | 2021-08-24 | Arlo Technologies, Inc. | Multi-sensor motion detection |
-
2019
- 2019-12-30 DE DE102019135851.5A patent/DE102019135851A1/de active Pending
-
2020
- 2020-12-14 JP JP2022540797A patent/JP7451718B2/ja active Active
- 2020-12-14 US US17/789,832 patent/US20230059090A1/en active Pending
- 2020-12-14 CN CN202080091202.5A patent/CN114902071A/zh active Pending
- 2020-12-14 WO PCT/EP2020/085902 patent/WO2021136643A1/de active Application Filing
- 2020-12-14 KR KR1020227022251A patent/KR20220104250A/ko unknown
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE102008013366A1 (de) * | 2008-03-10 | 2009-09-17 | Bayerische Motoren Werke Aktiengesellschaft | Verfahren zur Bereitstellung von Information für Fahrerassistenzsysteme |
US20150009031A1 (en) * | 2013-07-03 | 2015-01-08 | Honeywell International Inc. | Multilayer perimeter instrusion detection system for multi-processor sensing |
Also Published As
Publication number | Publication date |
---|---|
JP2023509154A (ja) | 2023-03-07 |
US20230059090A1 (en) | 2023-02-23 |
CN114902071A (zh) | 2022-08-12 |
KR20220104250A (ko) | 2022-07-26 |
JP7451718B2 (ja) | 2024-03-18 |
DE102019135851A1 (de) | 2021-07-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2019223931A1 (de) | Verfahren zum steuern eines zur durchführung eines automatisierten fahrbetriebs eingerichteten fahrzeugsystems eines fahrzeugs und vorrichtung zur durchführung des verfahrens | |
WO2001050154A1 (de) | Verfahren und vorrichtung zur dejustageerkennung bei einem kraftfahrzeug-radarsystem bzw. einem kraftfahrzeug-sensorsystem | |
DE102013206707A1 (de) | Verfahren zur Überprüfung eines Umfelderfassungssystems eines Fahrzeugs | |
DE102015217386A1 (de) | Verfahren und System zum Betreiben eines Kraftfahrzeugs | |
WO2022012924A1 (de) | Verfahren zum betreiben eines fahrerassistenzsystems eines fahrzeugs und verarbeitungseinrichtung zum steuern eines fahrerassistenzsystems | |
DE102018103803A1 (de) | System und verfahren für das erkennen einer fehlerhaften sensorinstallation innerhalb eines fahrzeugs, um die mit der objekterkennung verbundenen gefahren zu verringern | |
WO2022012923A1 (de) | Computerimplementiertes verfahren zum bestimmen der validität einer geschätzten position eines fahrzeugs | |
DE102018218492A1 (de) | Steuergerät, Verfahren und Sensoranordnung zur selbstüberwachten Lokalisierung | |
DE102018006503A1 (de) | Digitale Karte | |
WO2021130066A1 (de) | Training von neuronalen netzen durch ein neuronales netz | |
WO2019211293A1 (de) | Verfahren zum betreiben eines fahrerassistenzsystems eines egofahrzeugs mit wenigstens einem umfeldsensor zum erfassen eines umfelds des egofahrzeugs, computer-lesbares medium, system, und fahrzeug | |
EP3374798B1 (de) | System zum plausibilisieren von satellitensignalen globaler navigationssysteme | |
WO2021136643A1 (de) | Verfahren zum unterdrücken von mehrdeutigen messdaten von umfeldsensoren | |
EP3545330A1 (de) | Verfahren und vorrichtung zum bestimmen einer genauen position eines fahrzeugs anhand von radarsignaturen der fahrzeugumgebung | |
DE102020203829A1 (de) | Verfahren zum Überwachen eines Fahrzeugsystems zum Erkennen einer Umgebung eines Fahrzeugs | |
WO2023148339A1 (de) | Fahrassistenzsystem und verfahren zum betreiben eines fahrassistenzsystems | |
DE102020200133A1 (de) | Verfahren und Vorrichtung zum Betreiben eines automatisierten Fahrzeugs | |
EP3673234A1 (de) | Verfahren und vorrichtung zum bestimmen einer hochgenauen position und zum betreiben eines automatisierten fahrzeugs | |
DE102016223107A1 (de) | System und Verfahren zur Schätzung der Eigenbewegung eines Fahrzeuges | |
WO2020233957A1 (de) | Verfahren zum validieren einer kartenaktualität | |
DE102019211006B4 (de) | Auswerten von Sensordaten eines Fahrzeugs | |
WO2022002564A1 (de) | Ermitteln einer ausgangsposition eines fahrzeugs für eine lokalisierung | |
DE102020118628A1 (de) | Verfahren und Vorrichtung zum Bestimmen einer Position eines Fahrzeugs | |
DE102019131334A1 (de) | Verfahren zur Spur verfolgen von ausgedehnten Objekten mit wenigstens einer Detektionsvorrich-tung und Detektionsvorrichtung | |
DE102019212892A1 (de) | Erkennung von Detektorfehlern |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20830121 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 20227022251 Country of ref document: KR Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2022540797 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20830121 Country of ref document: EP Kind code of ref document: A1 |