EP3938946A1 - Method for providing a training data set quantity, method for training a classifier, method for controlling a vehicle, computer-readable storage medium and vehicle - Google Patents
Method for providing a training data set quantity, method for training a classifier, method for controlling a vehicle, computer-readable storage medium and vehicleInfo
- Publication number
- EP3938946A1 EP3938946A1 EP20700599.2A EP20700599A EP3938946A1 EP 3938946 A1 EP3938946 A1 EP 3938946A1 EP 20700599 A EP20700599 A EP 20700599A EP 3938946 A1 EP3938946 A1 EP 3938946A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- training data
- data set
- vehicle
- training
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000012549 training Methods 0.000 title claims abstract description 120
- 238000000034 method Methods 0.000 title claims abstract description 46
- 230000003287 optical effect Effects 0.000 claims abstract description 48
- 238000012545 processing Methods 0.000 claims abstract description 19
- 238000013528 artificial neural network Methods 0.000 claims abstract description 16
- 238000009434 installation Methods 0.000 claims abstract description 4
- 238000012546 transfer Methods 0.000 claims description 5
- 238000013507 mapping Methods 0.000 claims description 4
- 238000012512 characterization method Methods 0.000 abstract 1
- 238000004519 manufacturing process Methods 0.000 description 10
- 230000006870 function Effects 0.000 description 7
- 239000011521 glass Substances 0.000 description 6
- 230000005540 biological transmission Effects 0.000 description 3
- 238000001514 detection method Methods 0.000 description 3
- 238000013527 convolutional neural network Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000002372 labelling Methods 0.000 description 2
- 230000010287 polarization Effects 0.000 description 2
- 238000002310 reflectometry Methods 0.000 description 2
- 230000001133 acceleration Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 239000000356 contaminant Substances 0.000 description 1
- 238000006073 displacement reaction Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000001454 recorded image Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/28—Determining representative reference patterns, e.g. by averaging or distorting; Generating dictionaries
Definitions
- Method for providing a set of training data sets a method for training a classifier, a method for controlling a vehicle, a computer-readable storage medium and a vehicle
- the invention relates to a method for providing a set of training data sets, a method for training a classifier, a method for controlling a vehicle, a computer-readable storage medium and a vehicle.
- an image sensor For a number of driver assistance functions it is common in modern vehicles for an image sensor to be arranged behind a windshield, for example as part of a camera. This image sensor records the vehicle environment, in particular in the area in front of and behind the vehicle.
- the recorded image data are analyzed by a computing device in the vehicle and control functions are carried out based on this analysis. For example, if the analysis has determined that a stop sign is arranged in front of the vehicle, the vehicle can show the driver a warning that a vehicle stop should be carried out. In principle, however, it is also possible for the vehicle to independently carry out this stop.
- Classifiers are usually used to analyze the image data, which determine for each pixel of an image to which object or to which class of objects this pixel belongs. Object recognition can thus be carried out as a pixel-based classification.
- Artificial neural networks in particular “deep convolutional nets”, are often used as classifiers. These neural networks accept either a complete image or a section of an image as input parameters and specify the associated object class for each pixel.
- Neural networks can achieve a very high level of accuracy with a high hit rate (precision, recall).
- training data In order to train a classifier, it is necessary to provide annotated training data.
- These annotated training data contain a large amount of Image data, with an identification or a so-called label being stored for each image and there each pixel, to which object or which object class the pixel belongs.
- training data can contain an indication that a particular pixel belongs to a stop sign.
- Training data are usually generated by first driving a vehicle or several vehicles along a large number of different routes and thereby recording a large amount of image data.
- the image data is then annotated manually, i. H. performed by a human. This process is partially carried out automatically, but at least verified and corrected by a human. This is already necessary for legislative reasons.
- Annotating the training data is therefore a very complex and expensive process.
- a disadvantage of using the machine learning methods described is that a large amount of training data is required for this in order to be able to achieve the accuracy required for practical use. This is problematic because, as described, creating the training data is very complex and expensive.
- US 2017/0236013 A1 describes the generation of synthetic training data for an artificial neural network using a graphics engine. Objects can be placed anywhere in a three-dimensional space using the graphics engine. In this way, rare situations can be created in a targeted manner that only very rarely occur when recording with cameras in real situations. In this respect, an artificial neural network to be trained can be trained in such a way that it delivers improved results in the trained situations.
- the influence of the manufacturing tolerances reduces the accuracy of the classifier in the classification during operation. It is therefore the object of the invention to reduce the effort involved in generating training data. It is a particular object of the invention to reduce the influence of manufacturing tolerances on the classification. Another particular object of the invention is to increase the safety during the operation of a vehicle with assistance functions. The object is achieved by a method according to claim 1, a method according to claim 8, a method according to claim 9, a computer-readable storage medium according to claim 10 and a vehicle according to claim 11.
- the object is achieved by a method for providing a set of training data sets, in particular for an artificial neural network, comprising the following steps:
- Providing a training data set comprising the basic training data set and the initial training data set.
- a core of the invention is that the basic training data set is processed using an optical filter.
- the existing data is doubled. This allows a classifier to be trained better.
- the resources to be used to create the training data are also significantly reduced.
- Image data represent, in particular, a data structure in which individual images are stored in a chronologically ordered manner.
- Using an optical filter means in particular that image data of a training data set are modified by using the optical filter.
- the basic training data set is assigned properties of an optically transparent reference medium, in particular a reference windshield
- the output training data set is assigned properties of an optically transparent training medium, in particular a training windshield.
- optically transparent means in particular that the medium is permeable to visible light, in particular in the range from 400 nm to 800 nm.
- the basic training data set is thus assigned to an optically transparent reference medium.
- the training data can be recorded by an image sensor which is arranged behind a reference windshield.
- the initial training data set is in turn assigned to the properties of an optically transparent training medium, for example a training windshield. This means that by providing the training dataset, two different transparent media are now taken into account. This improves the accuracy of a classification when used with another windshield. This also makes it possible to take manufacturing tolerances into account in the production of optical media.
- the at least one optical filter can indicate an analytical mapping from the basic training data set to the initial training data set.
- the optical filter indicates an analytical image. Because an analytical mapping is specified, the processing of the basic training data set is comprehensible or predictable. It is thus possible in particular to provide information about which pixel in an image of the basic training data set corresponds to which pixel in an image of the output training data set. Correspondingly, taking into account the analytical mapping, an identification of the pixels of image data of the initial training data set can be adapted in accordance with the identification of the pixels of image data of the basic training data set. A method is thus specified in which the identification of the initial training data set can be carried out particularly efficiently.
- the image data can be stored as a set of pixels with assigned brightness values, preferably in each case for a multiplicity of color channels, wherein an assignment of image data to identifications can specify an associated object class for each pixel.
- An identifier can be an indication of an object class. For example, an identification can indicate that a certain pixel of an image is assigned to the object class “stop sign”. Ultimately, such a designation makes it possible to segment an image, with each pixel storing the object class to which it belongs. In one embodiment it is possible for an identifier to be stored as a data structure in which a coordinate of the pixel is stored as a first property and the assigned object class is stored as a second property.
- the optical filter can be determined by measuring properties of at least one optically transparent reference medium.
- An optical filter can in particular be designed as a Gaussian blurring, as an offset filter or as a color filter.
- the determination of the optical filter can be carried out efficiently by measuring properties of at least one optically transparent reference medium.
- the variance in the production of optically transparent media, for example windshields can thus be simulated by optical filters.
- the optical filter can be determined by determining a modulation transfer function. Determining a modulation transfer function is a particularly efficient implementation for determining the optical filter.
- the optical filter is determined taking into account an installation position of the optically transparent reference medium with respect to an image sensor. In one embodiment it is also possible that the optical filter is determined taking into account geometric properties of the optically transparent reference medium, in particular using a ray tracing-based method.
- Geometric properties can indicate a reflectivity, a thickness, a refractive power, a transmission and / or a polarization of an optically transparent medium.
- an image sensor with respect to the optically transparent medium is also taken into account. This can be exploited in particular when using ray tracing-based methods. All in all, an optically transparent medium can be simulated very precisely by using ray tracing-based methods. This improves the accuracy of the optical filters.
- the method can comprise the following steps:
- Image sensors generally have a characteristic noise which can turn out differently depending on the image sensor used.
- the noise can be measured, and optical filters can be designed to reduce the measured noise. It is therefore helpful to restore the noise ratio, which has been artificially changed by an optical filter, with a sensor-specific filter and to adapt it to the real conditions to be expected. Since the image data of the basic training data set and the initial training data set are recorded using the same image sensor, the same sensor filter can be used for all training data sets. It is of course also conceivable to use different sensor filters for the different training data sets. In particular, a sensor filter can be determined taking into account the image sensor used to record the image data of the corresponding training data sets.
- the object is also achieved in particular by a method for training an artificial neural network, comprising the following steps:
- the image sensor can be a CMOS or CCD sensor, for example.
- the acquisition of the reference image data can be carried out, for example, using a test vehicle on which an image sensor is arranged.
- the assignment of the identification to pixels can be carried out manually.
- the object is also achieved in particular by a method for controlling a vehicle, comprising the following steps:
- a control instruction can be, for example, an indication of a steering angle, an acceleration indication, a speed indication, a braking indication or a similar indication.
- a control instruction can be, for example, an indication of a steering angle, an acceleration indication, a speed indication, a braking indication or a similar indication.
- FIGS. 1 a and 1 b a schematic representation of a vehicle in a top view and a side view;
- FIG. 2 a representation of image data
- FIG. 3 a detailed view of an image section
- FIG. 4 a schematic representation of an assignment of pixels to
- FIG. 5 an illustration of the use of an optical filter
- FIG. 6 is an illustration showing the generation of a set of training data sets
- FIG. 7 an illustration of a light beam which is an optically transparent
- FIG. 8 a flow chart showing the generation of a
- FIG. 1A shows a vehicle 1.
- a camera 3 is arranged in the driver's cab of vehicle 1.
- the camera 3 supplies image data to a processing device 4, which is also arranged in the vehicle 1.
- the camera 3 is arranged in the area of a rearview mirror 7 of the vehicle 1.
- the camera 3 is arranged and aligned in such a way that the camera 3 can record the area in front of the vehicle 1.
- the camera 3 has an image sensor which can be designed as a CMOS or CCD sensor, for example.
- the driver 2 is shown symbolically in FIG. 1A and the steering wheel 5 in FIG. 1B.
- Light rays that are recorded by the image sensor of the camera 3 first pass a windshield 6 and then a lens of the camera 3.
- the effective passage area of the windshield 6 can have an area of 7 cm x 7 cm or preferably 40 cm x 20 cm .
- the image data recorded by the image sensor are sent to the processing device 4 via a bus system.
- the bus system can be, for example, an Ethernet-based communication system. It is also conceivable that a CAN bus or a similar data connection is used. In particular, it is conceivable that a wireless connection is used.
- the processing device 4 is designed to generate control instructions based on the image data for the vehicle 1.
- the processing device 4 can use an artificial neural network or another classifier.
- the image data serve as input parameters for the classifier.
- a classifier can be used that recognizes objects in the front area of vehicle 1.
- FIG. 2 shows an image section 10 of image data at a specific point in time.
- the classifier which is executed by the processing device 4 of the vehicle 1, is designed to determine the individual objects with pixel accuracy. This means that an object class can be specified for each pixel. Thereby it is possible, on the one hand, to segment the image detail 10 and, on the other hand, to determine which objects are in front of the vehicle 1.
- a control instruction can then be derived by the processing device 4.
- the position of objects 1 1, 12 can also be included as a parameter.
- the processing device 4 can be designed to output a warning to the driver 2 of the vehicle 1 when a stop sign 11 is arranged in front of the vehicle 1.
- a light in the vehicle interior can light up or a warning message can be projected into the field of view of the driver 2 by means of a head-up display.
- FIG. 4 shows that a white pixel 14 is assigned the object class 16, ie. H. "Background” (bg). However, object class 16 ‘is assigned to pixel 14‘ by means of assignment 15 Zu whatsoever. The object class 16 ‘indicates that the pixel 14‘ is part of a “stop sign” (obj 1). Correspondingly, those pixels of the image section 10 are also assigned to the object class 16 ”that are part of an object“ tree ”(obj 3).
- the camera 3 is arranged in the vehicle 1 behind a windshield 6.
- the windshield 6 has an influence on the recording of the surroundings of the vehicle 1.
- the windshield 6 can cause distortion.
- This is particularly disadvantageous because manufacturing tolerances occur in the manufacture of windshields, so that the representation of the same scene with different windshields 6 leads to different image data. If a classifier is then trained with the data of only one windshield 6, the manufacturing tolerances or different vehicle models are not taken into account. This leads to unsatisfactory results in the classification as described in connection with FIGS. 1 and 2.
- the effect that a windshield 6 has on the light which is transmitted from an object to the image sensor of the camera 3 can be approximated by means of optical filters.
- Such an optical filter 19 is shown as an example in FIG. In the example of FIG. 5, an original image detail 17 is shown, which was recorded using a reference windshield. The optical filter 19 now defines an image for each pixel 14 ‘of the original image section 17 on pixel 14 ′′ of a processed image section 18.
- FIG. 5 shows that the pixel 14 Pixel, which in the exemplary embodiment shown is arranged in the third line at the fourth position from the left, is arranged in the processed image section 18 in the fourth line at the third position from the left.
- An offset is therefore defined for each pixel 14 ‘.
- a number of other possible optical filters are of course conceivable. For example, different windshields can differ in their light transmission. As a result, the brightness values of the individual pixels have different strengths. This can be emulated with an optical filter. It is also conceivable that individual image areas are shown distorted by a slight curvature in the pane. Such a behavior can also be represented by an optical filter 19.
- an optical filter 19 can include an analytical representation, so that it is possible to understand which pixels in the output image correspond to which pixels in the processed image. As a result, an identification or a label of corresponding pixels can also be transmitted.
- FIG. 6 once again illustrates the advantage of the present invention.
- FIG. 6 shows that a training data set 31, which contains image data and a corresponding identifier, can be processed with an optical filter so that a training data set 30 is generated which includes the original training data set 31 and the processed training data set 3T. The number of training data was thus doubled, with different windshields now being covered by the training data.
- FIG. 7 shows how the properties of a pane 20 can be approximated with the aid of a ray tracing-based method.
- a light source 21 emits a light beam linearly in the direction of the pane 20.
- part of the light is reflected, so that a reflected light beam 22 is reflected away from the glass entry plane 24.
- Another part of the light beam is refracted and passed through the pane 20.
- the light beam is refracted again and directed in the direction of the camera 3. Before the light beam can strike an image sensor 23, it is refracted again by an objective 26 of the camera 3.
- the parameters of the pane 20 therefore include, on the one hand, the thickness B of the pane 20, the reflectivity, the refractive power, the transmission and / or the polarization. These parameters can also represent parameters of an optical filter 19, so that different slices can be emulated by adjusting the parameters of the optical filter 19.
- An optical filter 19 can be modeled by a large number of standard filters, for example a Gaussian blurring filter or a displacement filter.
- FIG. 8 is a flow chart which once again describes the entire method 40.
- image data 41 are recorded and the objects shown in the image data 41 are manually assigned to corresponding object classes in a labeling step 42.
- Annotated or labeled image data 43 are now processed in a processing step 44 using an optical filter 19.
- Different optical filters 19 are used in order to simulate a large number of different optically transparent media. For example, a large number of different windshields 6 can be simulated in this step.
- the processing step 44 generates a training dataset set 45, which is provided to a training algorithm for a classifier 47 in a training step 46.
- a training algorithm for a classifier 47 for example, it can be an artificial neural network, for example a convolutional neural network.
- the trained classifier 47 is transferred to a processing device 4 for a vehicle 1 in a transfer step 48.
- image data 50 are fed to the classifier during the operation of the vehicle 1, so that the classifier 47 classifies the objects stored in the image data 50.
- the classified image data 52 ie data which contain information about the objects shown in the image data, are analyzed by the processing device 4 in a control step 53, corresponding control instructions for actuators of the vehicle 1 being derived. These control instructions are also implemented in control step 53, so that, for example, a warning is displayed for a user.
- the detection step 51 and the control step 53 are carried out alternately until the vehicle 1 comes to a stop or is switched off.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
Description
Claims
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP19157049 | 2019-02-14 | ||
PCT/EP2020/050913 WO2020164841A1 (en) | 2019-02-14 | 2020-01-15 | Method for providing a training data set quantity, method for training a classifier, method for controlling a vehicle, computer-readable storage medium and vehicle |
Publications (1)
Publication Number | Publication Date |
---|---|
EP3938946A1 true EP3938946A1 (en) | 2022-01-19 |
Family
ID=65440803
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP20700599.2A Pending EP3938946A1 (en) | 2019-02-14 | 2020-01-15 | Method for providing a training data set quantity, method for training a classifier, method for controlling a vehicle, computer-readable storage medium and vehicle |
Country Status (4)
Country | Link |
---|---|
EP (1) | EP3938946A1 (en) |
CN (1) | CN111837125A (en) |
MA (1) | MA55272A (en) |
WO (1) | WO2020164841A1 (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3982328A1 (en) | 2020-10-08 | 2022-04-13 | Saint-Gobain Glass France | Method for simulating the effects of the optical distortions of a windshield on the image recording quality of a digital image recording device |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9373160B2 (en) | 2013-12-18 | 2016-06-21 | New York University | System, method and computer-accessible medium for restoring an image taken through a window |
US9996771B2 (en) | 2016-02-15 | 2018-06-12 | Nvidia Corporation | System and method for procedurally synthesizing datasets of objects of interest for training machine-learning models |
US10007269B1 (en) * | 2017-06-23 | 2018-06-26 | Uber Technologies, Inc. | Collision-avoidance system for autonomous-capable vehicle |
-
2020
- 2020-01-15 MA MA055272A patent/MA55272A/en unknown
- 2020-01-15 WO PCT/EP2020/050913 patent/WO2020164841A1/en unknown
- 2020-01-15 CN CN202080000259.XA patent/CN111837125A/en active Pending
- 2020-01-15 EP EP20700599.2A patent/EP3938946A1/en active Pending
Also Published As
Publication number | Publication date |
---|---|
MA55272A (en) | 2022-01-19 |
CN111837125A (en) | 2020-10-27 |
WO2020164841A1 (en) | 2020-08-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP3291172B1 (en) | Method for calculating final image data | |
WO2009127271A1 (en) | Traffic object detection system, method for detecting a traffic object, and method for setting up a traffic object detection system | |
DE102017217072B4 (en) | Method for detecting a weather condition in an area surrounding a motor vehicle, and control device and motor vehicle | |
WO2022128014A1 (en) | Correction of images from a panoramic-view camera system in the case of rain, incident light and contamination | |
EP3393875B1 (en) | Method for the improved detection of objects by a driver assistance system | |
EP2402827A1 (en) | Method and device for testing the functionality of an object recognition device of a motor vehicle | |
WO2022128013A1 (en) | Correction of images from a camera in case of rain, incident light and contamination | |
DE102020208008A1 (en) | Image classification and related training for security-related classification tasks | |
EP3938946A1 (en) | Method for providing a training data set quantity, method for training a classifier, method for controlling a vehicle, computer-readable storage medium and vehicle | |
DE102019219734B4 (en) | Method for operating an evaluation system for measurement data from multiple domains | |
EP3948649A1 (en) | Masking of objects contained in an image | |
DE102006037600A1 (en) | Method for resolution based representation of ambience of motor vehicle for use in parking system, involves capturing ambience information from ambience of motor vehicle by image sensor | |
DE102011121473A1 (en) | A method for displaying images on a display device of a motor vehicle, driver assistance device, motor vehicle and computer program | |
DE102008057671A1 (en) | Vehicle's e.g. commercial passenger car, surrounding area monitoring method, involves illuminating vehicle and/or surrounding area of vehicle depending on captured real illumination in overall image | |
DE102018201909A1 (en) | Method and device for object recognition | |
DE102016119592A1 (en) | Method for detecting objects in an environmental region of a motor vehicle taking into account sensor data in the infrared wavelength range, object recognition device, driver assistance system and motor vehicle | |
DE102020206990A1 (en) | Device for processing sensor data and training methods | |
DE102009009904A1 (en) | Object i.e. road sign, identifying method, involves adding images and classifiers, which contain stored samples and sequences, as another set of samples, if sequences correspond to another set of sequences | |
DE102019103192A1 (en) | Method for generating training data for a digital, adaptive camera system | |
DE102018221625A1 (en) | Transfer of additional information between camera systems | |
DE102020000490A1 (en) | Device and method for object recognition in a vehicle environment | |
DE102019004586A1 (en) | Process for the detection of disturbing light influences | |
DE102022212480A1 (en) | Matrix light guidance for object detection for automated driving | |
WO2024013162A1 (en) | Method for determining deficiencies in the detection of objects in data acquired by an environment sensor of a motor vehicle | |
DE102022205505A1 (en) | Method for creating a data set |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: UNKNOWN |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE |
|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE |
|
17P | Request for examination filed |
Effective date: 20210916 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
DAX | Request for extension of the european patent (deleted) | ||
RAV | Requested validation state of the european patent: fee paid |
Extension state: MA Effective date: 20210916 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: EXAMINATION IS IN PROGRESS |
|
17Q | First examination report despatched |
Effective date: 20231108 |