WO2000068910A1 - Verfahren und vorrichtung zur überwachung des innenraums und des umfeldes eines fahrzeugs - Google Patents
Verfahren und vorrichtung zur überwachung des innenraums und des umfeldes eines fahrzeugs Download PDFInfo
- Publication number
- WO2000068910A1 WO2000068910A1 PCT/DE2000/001426 DE0001426W WO0068910A1 WO 2000068910 A1 WO2000068910 A1 WO 2000068910A1 DE 0001426 W DE0001426 W DE 0001426W WO 0068910 A1 WO0068910 A1 WO 0068910A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- vehicle
- interior
- camera device
- driver
- camera
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/18—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
- G08B13/189—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
- G08B13/194—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
- G08B13/196—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
- G08B13/19695—Arrangements wherein non-video detectors start video recording or forwarding but do not generate an alarm themselves
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/18—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
- G08B13/189—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
- G08B13/194—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
- G08B13/196—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
- G08B13/19639—Details of the system layout
- G08B13/19645—Multiple cameras, each having view on one of a plurality of scenes, e.g. multiple cameras for multi-room surveillance or for tracking an object by view hand-over
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/18—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
- G08B13/189—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
- G08B13/194—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
- G08B13/196—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
- G08B13/19639—Details of the system layout
- G08B13/19647—Systems specially adapted for intrusion detection in or around a vehicle
Definitions
- the invention is based on a method according to the preamble of the main claim.
- Bosch-Zünder, October 1998 edition, page 8 the article "The new eyes of the car, learn to read limousines" describes a process in which the space in front of the driver in the vehicle's surroundings is monitored by two video cameras The image determined by the cameras is then evaluated with respect to traffic signs recognizable in the image, which are then displayed to the driver in a display unit
- the method according to the invention with the features of the main claim has the advantage that both the interior and the surroundings of a vehicle are captured with only one camera device. This is possible, in particular, by alternately detecting the interior and the surroundings. In the case of a sufficiently rapid change between the detection of the interior and the surroundings, on the one hand a loss of information due to the change can be neglected, while on the other hand only a single camera device and not two are required for the interior and the surroundings of the vehicle. Likewise, only one computing unit is required to process the image information obtained.
- the interior of the vehicle is illuminated with a radiation source that is at least largely invisible to the human eye.
- a radiation source that is at least largely invisible to the human eye.
- an infrared radiation source preferably one or more infrared light-emitting diodes. This prevents the driver from being disturbed compared to a visible source.
- Infrared radiation source and a detection of the image of the interior by an infrared filter an image of the interior is essentially only detected when the infrared radiation source is activated. An alternate observation of the interior and the surroundings is thus possible by successively switching the infrared radiation source on and off if the camera device has a further beam path which leads into the surroundings of the vehicle and can detect the surroundings.
- Image lines or by groups of pixels occur because the image information must also be transmitted to the processing unit and processed there, this procedure has the advantage that a faster change between a detection of the interior and the exterior is possible, so that the shift between two captured images eg the outside space, which is based on the vehicle movement, is reduced
- the observation by the camera does not require the application of electrodes to the driver's body. Since such electrodes can not only be a hindrance since they restrict the freedom of movement of the driver, but can also be forgotten by the driver before the start of the journey or cannot be intentionally carried out for convenience, the use of a sleep warning can thus be facilitated and made more pleasant for the driver become.
- the method according to the invention also has the advantage that, in addition to observing the interior, it is also possible to detect the traffic signs in the vicinity of the vehicle, and thus the driver e.g. warning signs or maximum speeds can be specifically indicated by an optical or acoustic output unit.
- This information can be used, for example, to regulate a chassis, by means of which an uneven load on the vehicle can be compensated if, for example, there are only people on the left side of the vehicle, the driver and one person behind the driver.
- This information can also be used to control a seat heater that is only activated when someone is actually using the seat.
- a determination as to whether a seat is occupied or occupied by a child seat is advantageous in that the deployment of an airbag is blocked. if a seat is unoccupied or occupied by a child seat. In this way, on the one hand, the unnecessary deployment of an airbag in an unoccupied seat and also an injury to a child by an airbag when occupying a seat with a child seat can be avoided.
- a voice input can be checked on the basis of an evaluation of the driver's lip movements, which are recorded by the camera device. This is e.g. possible by analyzing the lip movements to determine whether the command understood by the voice input unit also contains the syllables that correspond to the detected lip movements. If the speech input unit cannot make an unambiguous assignment based solely on what is understood, this may be possible through a comparison with the lip movements.
- a device in such a way that the surroundings of the vehicle and the interior of the vehicle can be detected.
- a camera device in such a way that a beam path points in the direction of the interior and a beam path points in the direction of the roadway, preferably in the direction of travel, since these are the most important for a driver
- Information about the vehicle's environment is usually the road or the lane edge as well as objects in your own lane.
- a beam path e.g. enter the camera device from the interior through reflection and another beam path through transmission through the semi-transparent mirror. This eliminates the need for mechanical adjustment between the two beam paths.
- the camera is also advantageous to design the camera as a CCD or a CMOS camera.
- the camera device according to the invention can hereby be carried out particularly cheaply.
- Integrate camera device into the roof of the vehicle To a position at least near the vehicle roof on the one hand, the overview of the vehicle surroundings, and on the other hand, the overview of the vehicle interior is particularly well guaranteed.
- At least one deflecting mirror so that it can be aligned by means of an adjusting device such that at least the driver's eyes and / or lips can be detected by the camera.
- This is particularly advantageous for changing drivers who on the one hand have a different body size and on the other hand can prefer a different seat setting.
- movements of the driver while driving are to be observed.
- a deflecting mirror in such a way that the detected visual range can be tracked, in particular the eyes and / or lips of a driver can be kept constantly in the detection range of the camera device. A function of the sleep control as well as a control of a voice input is thus guaranteed, especially while driving.
- FIG. 1 shows an arrangement of the device according to the invention in a motor vehicle
- FIG. 2 shows a flowchart of the method according to the invention
- FIGS. 2a and 2b show details of the method according to the invention
- FIG. 2c shows an evaluation method according to the invention
- FIG. 3 shows a flowchart for a further embodiment of the method according to the invention
- FIG 4 an embodiment of the device according to the invention
- FIG. 5 another Implementation of the device according to the invention
- FIG. 6 shows another embodiment of the device according to the invention
- FIG. 7 shows another embodiment of the device according to the invention
- FIG. 1 shows a camera device 10 according to the invention in a motor vehicle on the upper edge 11
- Windshield 12 arranged.
- the camera device has a first optical opening 13, a first beam path 14 leading to a driver 15 of the vehicle.
- the center beam of the beam path is shown.
- the camera device 10 has a second optical opening 16, which is arranged on the side of the camera device 10 facing away from the first optical opening 13 and is therefore not visible from the perspective shown.
- the second optical opening 16 is therefore only shown in broken lines.
- a second beam path 17 is outlined, which leads from the second optical opening 16 of the camera device 10 through the windshield 12 into the vehicle surroundings into the area in front of the vehicle.
- the driver's gaze which is outlined as a third beam path 18, also leads in the same direction.
- the cockpit 19 of the vehicle is shown with a steering wheel 20 and a display unit 21.
- the display unit 21 is preferably designed as an instrument cluster, in which a large number of displays are integrated into an electronic unit.
- a freely programmable instrument cluster is also possible, in which a display of different display instruments in one Screen, for example in the form of a liquid crystal display.
- a computing unit by which the image information recorded by the camera device 10 is processed is not shown separately in the figure.
- the computing unit can either be arranged in the housing of the camera device 10 shown, in the roof of the vehicle beyond the upper edge 11 of the windshield or in the cockpit 19 of the vehicle.
- the computing unit is arranged in an area of the display unit 21 which is not visible to the driver 15. Since the display unit 21 is used for the output of optical warning signals, which are based on the evaluation of the image information recorded by the camera device 10 by the computing unit, for example if the driver threatens to fall asleep or if the driver falls asleep
- the camera device 10 is so close to the vehicle roof, not shown, in the upper region of the
- the first and second optical openings 13, 16 can be designed in different ways.
- An embodiment as a filter, as an opening, as an objective or a combination in the form of a series connection of the named components is possible.
- FIG. 1 A sequence of the method according to the invention is shown in FIG.
- first image information 32 of the vehicle surroundings is acquired and evaluated by the computing unit, with a first output 33 taking place via optical and / or acoustic output media as a function of the first image information 32.
- the first output is therefore dependent on the vehicle environment.
- second method step 34 image information is obtained from second image information 35, in which image information of the vehicle environment and the vehicle interior is superimposed, taking into account the previously determined first image information 32 of the vehicle environment by subtracting the first image information 32 from the second image information 35 of the vehicle interior is determined, so that, depending on the image information determined, a second output 36 also takes place via optical and / or acoustic output media, the second output being dependent in particular on the image information of the vehicle interior.
- the method is terminated if the camera device is deactivated, in particular if the vehicle is parked. This decision-making path is marked in the drawing with the letter "Y".
- the method ends with a deactivation of the camera device in a final method step 38. If the vehicle is not deactivated, the method branches back to the first method step 31. This decision-making path is identified in FIG. 2 with the letter “N”.
- the first method step 31 is shown in detail in FIG. 2a.
- a first sub-step 40 the camera device is activated and the first
- Image information 32 recorded.
- the first image information 32 is transmitted to the computing unit for further processing.
- the second method step 34 is shown in FIG. 2b
- a first sub-step 42 the radiation source which is not visible to the human eye is activated by being supplied with an electrical voltage.
- the camera device 10 is activated and an overlaid image of the interior and the vehicle surroundings is acquired as second image information 35.
- exposure must be regulated, e.g. a variable aperture or regulation of the current applied to the light-sensitive sensors of the camera device.
- the second image information 35 is stored and transmitted to the computing unit for further processing.
- a fourth sub-step 45 that which is not visible to the human eye is shown
- FIG. 2c shows an evaluation method by the computing unit, which consists of processing the image information recorded by the camera device and the first output 33 and the second output 36, respectively.
- a sleep warning by observation of the driver 15 is given, an observation of the vehicle interior being necessary and consequently the second output 36 taking place.
- a method for vehicle surroundings recognition can also be carried out in a comparable manner, e.g. for recognizing traffic signs and / or road markings, the first output 33 taking place.
- a first initialization step 50 an image of the driver's eye area is determined from the first and second image information 32 and 35 by the computing unit.
- a first decision step 52 the recorded image is compared with previously stored image information 51 of the driver's eye area.
- the image information 51 is an empty image if the vehicle has just started and no image information has yet been stored. If it is determined that the driver's eyes are open, that is to say the driver is not sleeping, or if the image information 51 is the empty image, the decision path N is followed and the newly recorded partial image is stored in a method step 53 filed that the driver is awake at the time of recording.
- a final step 54 the evaluation process completed. The evaluation process is started again the next time the first and second image information 32 and 35 are transmitted to the computing unit. The restart takes place every time the evaluation process is ended, unless the vehicle or the camera device is not deactivated.
- the decision path Y becomes a second from the first decision step 52
- Step 55 it is checked whether the driver's eyes were already closed when the last picture was taken. If this is not the case, a branch is made to a sub-step 56 in which it is stored that the driver's eyes are closed at the time of the current recording. The evaluation process is ended in a final step 57. If the driver's eyes were already closed when the last picture was taken, the decision path Y is followed from the second decision step 55 to a first warning step 58. This warning is an acoustic warning or a visual warning, preferably via the display unit 21. The fact that only a warning is given after a second recording and thus after the second decision step 55 prevents a warning from being given by accidental ones
- Step 59 in which image information 67 of a further image of the driver's face area is taken into account. If the driver's eyes are now open again, a branch is made in decision path Y to a method step 60 in which the newly recorded image information 67 is stored. It is also stored in a memory that the driver's eyes are open. In a subsequent final step 61, the evaluation process is ended. However, if the driver's eyes are still closed, a decision path N to a second warning step 62 is followed in the third decision step 59. In the second warning step 62, an acoustic warning is given clearly louder than in the first warning step 58.
- a fourth decision step 63 image information 68 of the part of the driver's face is recorded again and the state 69 of a switch is queried. If it is ascertained that the driver's eyes are now open, or if the driver actuates the switch, the decision path Y is continued. In a first sub-step 64, it is now stored that the driver's eyes are open, and the evaluation process is ended in a final step 65. If it is not ascertained that the driver's eyes are open or that the switch has been triggered, a branch is made in decision path N to a third warning step 66. A loud, acoustic warning now follows, and the vehicle is decelerated by switching on the hazard warning lights and the brake lights, so that driverless driving is avoided.
- the sequence shown in FIG. 2c can be deactivated. It is also possible to increase the number of queries of the image information from the driver's eye area to the implementation of a respective warning step in order to avoid accidental warnings. The number of queries depends on the frequency of the acquisition of image information of the interior.
- the method shown in FIG. 2c can also be based on the monitoring position of the vehicle
- Road markings are transmitted if the detection of the image information of the part of the driver's face is replaced by a detection of the image information of the road marking and the position of the vehicle relative to the road markings is evaluated.
- FIG. 3 shows a further method according to the invention for monitoring the surroundings of the interior of a motor vehicle.
- the same reference numerals stand for the same method elements as in FIG. 2.
- a first method step 80 determines a first image information 81 of the vehicle's surroundings, forwards it to the computing unit and, depending on the first image information 81, the first output 33 occurs second method step 82, second image information 83 of the interior is captured by the camera device and forwarded to the computing unit.
- the second output 36 occurs depending on the captured image information.
- an electro-optical light valve is opened in the direction of the vehicle surroundings.
- an electronic light valve to the vehicle interior is opened.
- the second method step 82 After the second method step 82, it comes to a decision step 37. If the camera device is deactivated, the decision path Y is followed and in a final method step 38 the camera device is deactivated. Otherwise, the decision path labeled N is branched back to the first method step 80. In a preferred exemplary embodiment, the respective light valve is only opened for 90% of the duration of the respective method step both during the first and during the second method step 80, 82. This will create a
- the evaluation method described in FIG. 2c can be transferred directly to the first edition or the second edition 36 in FIG. 3.
- FIG. 4 shows an embodiment of a camera device 10 according to the invention with a computing unit 110.
- the camera device 10 is located in a housing in which a camera 100, which is designed as a CCD or a CMOS camera, is arranged with a first lens 101.
- a first deflecting mirror 102 strikes the first lens 101.
- the first deflecting mirror 102 is designed to be semi-transparent, so that on the one hand a first beam path 103 from the vehicle surroundings through an opening 109 in the housing of the camera device 10 through the first deflecting mirror 102 and the first Objective 101 runs to camera 100.
- a second beam path 108 runs from a second deflecting mirror 104 to the first deflecting mirror 102.
- the second beam path 108 is deflected by the first deflecting mirror 102 to the camera 100.
- the second beam path 108 starts from the vehicle interior and enters the camera device 10 through a second lens 107. Before it reaches the second deflecting mirror 104, it crosses an infrared filter 106.
- the camera 100 is connected to the computing unit 110 via a first data connection 111.
- the computing unit 110 consists of a control unit 112 and an evaluation unit 113, which uses a second
- Data connection 114 are interconnected.
- the evaluation unit 113 is connected to sensors 116 via a third data connection 117 and at least to acoustic and / or optical display elements 119 via a fourth data connection 118.
- the control unit 112 is also connected via a fifth data connection 120 to the camera 100 and via a sixth data connection 122 to a radiation source 121 which emits radiation which is not visible to the human eye.
- the radiation source 121 is arranged in a housing, which is preferably designed as a reflector 123.
- the first beam path 103 and the second beam path 108 are each characterized by the optical axis of the beam path. Here and in the following FIGS. 5-7, only this center beam is shown as representative of the entire beam path. In front of the first objective 101, the optical axis runs the same for both beam paths. For better clarity of the drawing, the two beam paths are in the
- the computing unit 110 and the camera device 10 can also be arranged in a single housing near the vehicle roof or the upper limit of the windshield 12. However, it is also possible to arrange the computing unit 110 and the camera device 10 at different locations on the vehicle. In a preferred exemplary embodiment, the computing unit 110 is integrated in the display unit 21.
- the radiation is emitted into the interior of the vehicle.
- the radiation emitted is not visible to a person.
- the radiation source is preferably in the form of an infrared radiation diode or an infrared radiation diode array of a multiplicity of
- Infrared radiation diodes executed. If the interior of the vehicle is consequently illuminated by the radiation source 121, so occurs the reflected in the vehicle interior
- the filtered infrared radiation strikes the second deflecting mirror 104, the first deflecting mirror 102, the first objective 101 and the camera 100.
- the second deflecting mirror 104 is provided with an adjusting device 30. Only one holder 130 of this actuating device is shown in the figure. An electric motor and a controller as well as a power supply are not shown. With the adjusting device, the second deflecting mirror 104 can be rotated about an axis of rotation 131 within a certain angular range. This allows the area of
- the interior can be changed, which is imaged into the camera 100 by the second lens 107 and via the second deflection mirror. This is particularly advantageous if a driver changes his seating position while driving and nevertheless his face area is to be captured by the camera device 10.
- Seat sensors are to be mentioned as sensors 116, which provide information as to whether a seat is occupied. If a seat sensor reports that a seat is unoccupied, the camera can be used to check whether this is really the case or whether there is movement in the seat, for example takes place and for this reason the seat is occupied. In this case, the airbag is not deactivated and / or the seat heating is deactivated.
- the sensors are also to be understood as input elements with which, for example, a sleep warning can be deactivated when the driver is wearing sunglasses that do not make his eyes visible to the camera 100.
- the output units are to be understood as acoustic and / or optical warning elements which can be designed as loudspeakers, warning lamps or liquid crystal displays.
- Evaluation unit 113 and control unit 112 can also be embodied integrated in one device.
- the control unit 112 controls the position of the second deflecting mirror 104 via a connection (not shown) as a function of instructions transmitted by the evaluation unit 113 via the second data connection 114. If an object that is being observed by the camera device 10 threatens to move out of the visible area, the computing unit can adjust the visible area in this way by regulating the second deflection mirror.
- a connection between the camera device 10 and the computing unit 110 takes place via the first data connection 111 and the fifth data connection 120.
- the first data connection 111 is used to transmit image information from the camera 100 to the computing unit 110, in particular to the evaluation unit 113.
- the fifth data connection 120 is used to control the camera 100 by the computing unit 110, in particular by the control unit 112.
- the first data connection 111 and the fifth data connection 120 can also be combined in one data line.
- FIG. 5 shows a further exemplary embodiment of the device according to the invention for monitoring the vehicle surroundings and the vehicle interior.
- the same reference symbols also denote the same components.
- the second beam path 108 already leaves the housing of the camera device 10 after the infrared filter 106. To distinguish it from the infrared filter 106, the housing of the camera device 10 is only shown in broken lines in FIG. The one shown in Figure 5
- Design allows the camera device with the cutting plane to be arranged in parallel on the one hand, but also perpendicular to the vehicle roof in the vehicle.
- the area of the camera is completely accommodated in the vehicle roof except for the opening 109, while the area of the second deflecting mirror projects into the vehicle interior, that is to say the sectional plane of the drawing is arranged perpendicular to the vehicle roof .
- Deflecting mirror 104 essentially uses the optical properties of the first lens 101 to generate an image in the camera 100.
- FIG. 6 shows another embodiment of the device according to the invention for monitoring the surroundings of a vehicle interior.
- the camera 100 is arranged on a different side of the first deflecting mirror 102 than in FIGS. 4 and 5.
- the light following the first beam path 103 is reflected by the first deflecting mirror 102 to the camera 100.
- the second beam path 108 following radiation deflected by the second deflecting mirror 104 in such a way that the radiation crosses the first deflecting mirror 102, which is designed as a semi-transparent mirror, and finally reaches the camera 100.
- the reflector 123 is integrated in the housing of the camera device 10, as a result of which space can be saved.
- the radiation source 121 can also be arranged in a favorable place far away from the camera device 10 in the motor vehicle.
- Radiation sources possible in the vehicle to ensure optimal illumination of the vehicle interior.
- FIG. 7 shows a device for carrying out the method according to the invention described in FIG. 3.
- an electro-optical light valve in the form of a first liquid crystal cell 151 is inserted into the first beam path 103.
- the first liquid crystal cell 151 can be controlled by the control unit 112 via a control line 150 in such a way that it is possible to switch between a transmissive and absorbing state of the first liquid crystal cell 151.
- a detailed structure of the liquid crystal cell and a power supply are not shown in the drawing.
- the first liquid crystal cell 151 can be controlled by the control unit 112 via a control line 150 in such a way that it is possible to switch between a transmissive and absorbing state of the first liquid crystal cell 151.
- a detailed structure of the liquid crystal cell and a power supply are not shown in the drawing.
- Liquid crystal cell 151 can be designed in such a way that a liquid crystal is arranged between two glass substrates between two transparent electrodes and influences the direction of polarization of the light differently depending on an applied electric field. By arranging polarizing foils on the glass substrates, depending on the transparent electrodes applied an absorption or a maximum transmission of the light predetermined by the glass substrate, the polarizers and the liquid crystal possible.
- a second liquid crystal cell 153 is also designed, which can be switched by the control unit 112 via a control line 152 and which is arranged in the second beam path 108.
- the first liquid crystal cell 151 is now transparent and the second liquid crystal cell 153 is switched to be absorbent. In this case, only the light extending from the vehicle surroundings along the first beam path 103 enters the camera 100.
- the second method step 34 the first liquid crystal cell 151 is now absorbent and the second liquid crystal cell 153 is switched transmissively.
- the second method step 34 the first liquid crystal cell 151 is now absorbent and the second liquid crystal cell
- Liquid crystal cells 151 and 153 are connected to be absorbent. This is particularly advisable at low temperatures, since in this case a switching of the liquid crystal can be delayed and maximum absorption or transmission is only achieved after the electric field has been applied for some time. In contrast to those shown in Figures 4 to 6
- visible light also enters the camera 100 along the second beam path 108.
- a suitable calculation by the evaluation unit 113 can be used to draw conclusions about the distances of individual objects from the stereoscopic image acquisition. This is advantageous, for example, when recognizing objects, such as traffic signs.
- FIGS. 8a and 8b Exemplary embodiments for the second deflecting mirror 104 are shown in FIGS. 8a and 8b.
- a second deflection mirror 1041 is concave in FIG. 8a and a second deflection mirror 1042 is convex in FIG. 8b. Both the deflecting mirror 1041 and the deflecting mirror 1042 can be used as a second deflecting mirror 104.
- Such a design makes it possible to change the area visible to the camera. In FIG. 8b, widening is achieved, and in FIG. 8a, the beam area is restricted by the different curvature of the mirror.
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Fittings On The Vehicle Exterior For Carrying Loads, And Devices For Holding Or Mounting Articles (AREA)
- Traffic Control Systems (AREA)
- Closed-Circuit Television Systems (AREA)
- Emergency Alarm Devices (AREA)
- Auxiliary Drives, Propulsion Controls, And Safety Devices (AREA)
Abstract
Description
Claims
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2000617420A JP2002544043A (ja) | 1999-05-08 | 2000-05-05 | 車両の室内空間及び周辺環境の監視のための方法及び装置 |
DE50007823T DE50007823D1 (de) | 1999-05-08 | 2000-05-05 | Verfahren und vorrichtung zur überwachung des innenraums und des umfeldes eines fahrzeugs |
US09/743,305 US6920234B1 (en) | 1999-05-08 | 2000-05-05 | Method and device for monitoring the interior and surrounding area of a vehicle |
EP00941906A EP1297511B1 (de) | 1999-05-08 | 2000-05-05 | Verfahren und vorrichtung zur überwachung des innenraums und des umfeldes eines fahrzeugs |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE19921488.3 | 1999-05-08 | ||
DE19921488A DE19921488A1 (de) | 1999-05-08 | 1999-05-08 | Verfahren und Vorrichtung zur Überwachung des Innenraums und des Umfeldes eines Fahrzeugs |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2000068910A1 true WO2000068910A1 (de) | 2000-11-16 |
Family
ID=7907575
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/DE2000/001426 WO2000068910A1 (de) | 1999-05-08 | 2000-05-05 | Verfahren und vorrichtung zur überwachung des innenraums und des umfeldes eines fahrzeugs |
Country Status (5)
Country | Link |
---|---|
US (1) | US6920234B1 (de) |
EP (1) | EP1297511B1 (de) |
JP (1) | JP2002544043A (de) |
DE (2) | DE19921488A1 (de) |
WO (1) | WO2000068910A1 (de) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005523200A (ja) * | 2002-04-23 | 2005-08-04 | オートリブ ディヴェロプメント アクチボラゲット | 夜間視界装置 |
US7995095B2 (en) | 2001-10-18 | 2011-08-09 | Autoliv Development Ab | Night vision device for a vehicle |
US8471911B2 (en) | 2000-10-26 | 2013-06-25 | Autoliv Development Ab | Night vision arrangement |
DE102016001054A1 (de) | 2016-01-30 | 2016-07-21 | Daimler Ag | Verfahren zur Eingabe eines Befehls in ein Steuergerät eines Fahrzeuges |
DE102019210144A1 (de) * | 2019-07-10 | 2021-01-14 | Zf Friedrichshafen Ag | Kombination von Sensorsystemen im Fahrzeug zum Verbessern des Erkennens von Nutzerbefehlen |
Families Citing this family (32)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6583730B2 (en) | 2000-07-28 | 2003-06-24 | Lang-Mekra North America, Llc | Surveillance apparatus for a vehicle |
DE10036875A1 (de) | 2000-07-28 | 2002-02-28 | Mekra Lang Gmbh & Co Kg | Rückspiegel, insbesondere für Nutzfahrzeuge mit Kamera und Monitor |
DE10062977A1 (de) | 2000-12-16 | 2002-06-27 | Bayerische Motoren Werke Ag | Verfahren zur Überwachung des Innenraums eines Kraftfahrzeuges |
KR100910670B1 (ko) * | 2001-03-05 | 2009-08-04 | 지멘스 악티엔게젤샤프트 | 승객 보호 시스템을 위한 이미지 보정 방법 및 장치 |
DE10216111B4 (de) * | 2002-04-12 | 2009-12-03 | GM Global Technology Operations, Inc., Detroit | Fahrzeug mit einem im Fond- oder Frontbereich angeordneten Kindersitz |
DE10353001A1 (de) * | 2003-11-13 | 2005-06-16 | Conti Temic Microelectronic Gmbh | Vorrichtung und Verfahren zur Objekterkennung für eine Kraftfahrzeug-Sicherheitseinrichtung |
DE102004005163B3 (de) * | 2004-02-02 | 2005-06-02 | Braun, Uwe Peter, Dipl.-Ing. | Vorrichtung zur Erfassung der Fahrtüchtigkeit eines Fahrers in einem Fahrzeug |
FR2885860B1 (fr) * | 2005-05-20 | 2007-08-17 | Valeo Vision Sa | Dispositif de detection d'obstacles comportant un systeme d'imagerie pour vehicule automobile |
US7597489B2 (en) * | 2005-09-12 | 2009-10-06 | Honeywell International Inc. | Apparatus and method for providing pointing capability for a fixed camera |
US20070115343A1 (en) * | 2005-11-22 | 2007-05-24 | Sony Ericsson Mobile Communications Ab | Electronic equipment and methods of generating text in electronic equipment |
DE102006004260A1 (de) | 2006-01-31 | 2007-08-02 | Bayerische Motoren Werke Ag | Kamerasystem für ein Kraftfahrzeug |
JP4882499B2 (ja) * | 2006-05-10 | 2012-02-22 | 株式会社デンソー | 速度情報提供装置及びプログラム |
SE532317C2 (sv) * | 2007-07-05 | 2009-12-15 | Svenska Utvecklings Entrepreno | Anordning för att väcka förare och operatörer |
DE102008040149A1 (de) | 2008-07-03 | 2010-01-07 | Robert Bosch Gmbh | Vorrichtung und Verfahren zur Freigabe einer automatischen Führung eines Fahrzeugs |
WO2010050012A1 (ja) * | 2008-10-29 | 2010-05-06 | 京セラ株式会社 | 車載用カメラモジュール |
DE102012205012A1 (de) * | 2011-07-12 | 2013-01-17 | Robert Bosch Gmbh | Kamerasystem für den Einsatz in einem Fahrzeug sowie Fahrzeug mit einem derartigen Kamerasystem |
DE102012018345A1 (de) | 2012-09-17 | 2014-03-20 | Volkswagen Aktiengesellschaft | Verfahren und Vorrichtung zur Überwachung für ein Fahrzeug |
US9128354B2 (en) | 2012-11-29 | 2015-09-08 | Bendix Commercial Vehicle Systems Llc | Driver view adapter for forward looking camera |
JP5590747B2 (ja) * | 2012-11-29 | 2014-09-17 | 株式会社ホンダアクセス | 運転者危険運転報知装置 |
DE102012023941A1 (de) * | 2012-12-06 | 2014-06-12 | GM Global Technology Operations LLC (n. d. Gesetzen des Staates Delaware) | Kraftfahrzeug-Cockpit mit einer Anzeigeneinheit und einem Schattenbereich |
DE102013002686B4 (de) * | 2013-02-15 | 2017-03-23 | Audi Ag | Verfahren zum Betreiben eines Kraftwagens sowie Kraftwagen |
DE102013010019B3 (de) * | 2013-06-14 | 2014-10-23 | Audi Ag | Verfahren zum Betreiben eines Verkehrsschilderfassungssystems eines Kraftwagens und Kraftwagen mit einem Verkehrsschilderfassungssystem |
KR102381140B1 (ko) * | 2015-02-25 | 2022-03-31 | 엘지전자 주식회사 | 디지털 디바이스 및 그의 운전자 모니터링 방법 |
DE102015117610A1 (de) * | 2015-10-16 | 2017-04-20 | Connaught Electronics Ltd. | Verfahren zum Überprüfen einer Wahrnehmung eines visuellen Warnhinweises durch einen Fahrzeugnutzer, Fahrerassistenzsystem und Kraftfahrzeug |
DE102016225518A1 (de) | 2016-12-20 | 2018-06-21 | Robert Bosch Gmbh | Kameraanordnung für eine Überwachungsvorrichtung zum Überwachen eines Innenraums und eines Außenbereichs, Überwachungsvorrichtung mit der Kameraanordnung und Verwendung der Kameraanordnung zum Aufnehmen eines Außenbereichs und Innenraums |
US10272807B2 (en) | 2017-05-02 | 2019-04-30 | Ford Global Technologies, Llc | Efficient control of temperature altering systems within a vehicle seating assembly |
US20190012552A1 (en) * | 2017-07-06 | 2019-01-10 | Yves Lambert | Hidden driver monitoring |
JP7247499B2 (ja) * | 2018-09-20 | 2023-03-29 | いすゞ自動車株式会社 | 車両用監視装置 |
DE102018217869A1 (de) | 2018-10-18 | 2020-04-23 | Robert Bosch Gmbh | Innenraumerfassungssystem |
DE102019200099A1 (de) * | 2019-01-07 | 2020-07-09 | Conti Temic Microelectronic Gmbh | Sensorvorrichtung für ein Ego-Fahrzeug, Fahrerassistenzvorrichtung und Fahrzeug mit einer solchen Sensorvorrichtung |
CN111722609B (zh) * | 2019-03-18 | 2021-12-21 | 纬湃科技投资(中国)有限公司 | 针对车辆环境信号的诊断方法 |
JP7369796B2 (ja) * | 2019-06-13 | 2023-10-26 | ジェンテックス コーポレイション | 切替可能なマルチビュー撮像システム |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5293427A (en) * | 1990-12-14 | 1994-03-08 | Nissan Motor Company, Ltd. | Eye position detecting system and method therefor |
DE19546391A1 (de) * | 1994-12-12 | 1996-06-13 | Hisatsugu Nakamura | Bewegliche interakitv eingesetzte Arbeitsstation |
US5598145A (en) * | 1993-11-11 | 1997-01-28 | Mitsubishi Denki Kabushiki Kaisha | Driver photographing apparatus |
DE19736774A1 (de) * | 1997-08-23 | 1999-02-25 | Bosch Gmbh Robert | Verfahren zur Informationsanzeige in einem Kraftfahrzeug |
Family Cites Families (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5845000A (en) * | 1992-05-05 | 1998-12-01 | Automotive Technologies International, Inc. | Optical identification and monitoring system using pattern recognition for use with vehicles |
GB8825446D0 (en) * | 1988-10-31 | 1988-11-30 | Lawrence M J | Vehicle security camera |
DE4013125A1 (de) * | 1989-06-19 | 1990-12-20 | Iveco Magirus | Verfahren zur messung einer physikalischen groesse mit einem faseroptischen sensor |
DE4022055C2 (de) * | 1990-07-11 | 1994-03-10 | Wolfgang Caspar | Vorrichtung zur Farb- und Helligkeitsregelung |
US5065236A (en) * | 1990-11-02 | 1991-11-12 | The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration | Stereoscopic camera and viewing systems with undistorted depth presentation and reduced or eliminated erroneous acceleration and deceleration perceptions, or with perceptions produced or enhanced for special effects |
DE4115202C2 (de) * | 1991-05-10 | 1994-01-20 | Duerrwaechter E Dr Doduco | Verfahren und Einrichtung zum Scharfmachen der Auslöseschaltung eines Airbags oder eines Sicherheitsgurt-Strammers in einem Fahrzeug |
FR2679357B1 (fr) * | 1991-07-19 | 1997-01-31 | Matra Sep Imagerie Inf | Dispositif embarque et procede de reperage et de suivi de la position d'un vehicule sur la route et dispositif d'aide a la conduite en comportant application. |
JPH05155291A (ja) * | 1991-12-03 | 1993-06-22 | Mitsubishi Electric Corp | 車両用警報装置 |
TW333368U (en) | 1992-04-21 | 1998-06-01 | Toshiba Co Ltd | Image tube apparatus |
US5642093A (en) * | 1995-01-27 | 1997-06-24 | Fuji Jukogyo Kabushiki Kaisha | Warning system for vehicle |
JP3316725B2 (ja) * | 1995-07-06 | 2002-08-19 | 三菱電機株式会社 | 顔画像撮像装置 |
JPH1035320A (ja) * | 1996-07-24 | 1998-02-10 | Hitachi Ltd | 車両状況認識方法、車載用画像処理装置および記憶媒体 |
JP3702978B2 (ja) * | 1996-12-26 | 2005-10-05 | ソニー株式会社 | 認識装置および認識方法、並びに学習装置および学習方法 |
AU4184399A (en) * | 1998-05-13 | 1999-11-29 | Infinite Pictures Inc. | Panoramic movies which simulate movement through multidimensional space |
US6690374B2 (en) * | 1999-05-12 | 2004-02-10 | Imove, Inc. | Security camera system for tracking moving objects in both forward and reverse directions |
-
1999
- 1999-05-08 DE DE19921488A patent/DE19921488A1/de not_active Ceased
-
2000
- 2000-05-05 DE DE50007823T patent/DE50007823D1/de not_active Expired - Fee Related
- 2000-05-05 US US09/743,305 patent/US6920234B1/en not_active Expired - Fee Related
- 2000-05-05 EP EP00941906A patent/EP1297511B1/de not_active Expired - Lifetime
- 2000-05-05 JP JP2000617420A patent/JP2002544043A/ja not_active Withdrawn
- 2000-05-05 WO PCT/DE2000/001426 patent/WO2000068910A1/de active IP Right Grant
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5293427A (en) * | 1990-12-14 | 1994-03-08 | Nissan Motor Company, Ltd. | Eye position detecting system and method therefor |
US5598145A (en) * | 1993-11-11 | 1997-01-28 | Mitsubishi Denki Kabushiki Kaisha | Driver photographing apparatus |
DE19546391A1 (de) * | 1994-12-12 | 1996-06-13 | Hisatsugu Nakamura | Bewegliche interakitv eingesetzte Arbeitsstation |
DE19736774A1 (de) * | 1997-08-23 | 1999-02-25 | Bosch Gmbh Robert | Verfahren zur Informationsanzeige in einem Kraftfahrzeug |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8471911B2 (en) | 2000-10-26 | 2013-06-25 | Autoliv Development Ab | Night vision arrangement |
US7995095B2 (en) | 2001-10-18 | 2011-08-09 | Autoliv Development Ab | Night vision device for a vehicle |
US8228379B2 (en) | 2001-10-18 | 2012-07-24 | Autoliv Development Ab | Night vision device for a vehicle |
JP2005523200A (ja) * | 2002-04-23 | 2005-08-04 | オートリブ ディヴェロプメント アクチボラゲット | 夜間視界装置 |
US7592592B2 (en) | 2002-04-23 | 2009-09-22 | Autoliv Development Ab | Night vision arrangement |
DE102016001054A1 (de) | 2016-01-30 | 2016-07-21 | Daimler Ag | Verfahren zur Eingabe eines Befehls in ein Steuergerät eines Fahrzeuges |
DE102019210144A1 (de) * | 2019-07-10 | 2021-01-14 | Zf Friedrichshafen Ag | Kombination von Sensorsystemen im Fahrzeug zum Verbessern des Erkennens von Nutzerbefehlen |
Also Published As
Publication number | Publication date |
---|---|
DE50007823D1 (de) | 2004-10-21 |
EP1297511B1 (de) | 2004-09-15 |
DE19921488A1 (de) | 2000-11-16 |
US6920234B1 (en) | 2005-07-19 |
EP1297511A1 (de) | 2003-04-02 |
JP2002544043A (ja) | 2002-12-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP1297511B1 (de) | Verfahren und vorrichtung zur überwachung des innenraums und des umfeldes eines fahrzeugs | |
DE102016111783B4 (de) | Anzeigevorrichtung zur Einblendung eines virtuellen Bildes in das Blickfeld eines Benutzers | |
DE102004050181B4 (de) | Aktives Nachtsichtsystem mit adaptiver Bildgebung | |
DE10339314B3 (de) | Verfahren zur Anzeigesteuerung von unterschiedlichen Informationen in einem Fahrzeug und optoakustische Informationseinheit | |
EP1182089B1 (de) | Verfahren zur Warnung eines Fahrers eines Fahrzeugs | |
DE102015219111B4 (de) | Fahrerassistenzsystem und Verfahren zur Anpassung der Beleuchtung von Bedienelementen bei Kraftfahrzeugen | |
DE60300919T2 (de) | Bildaufnahmegerät der Fahrzeuginsassen in einem Fahrzeug | |
DE10261291A1 (de) | Anzeigeeinrichtung für ein Fahrzeug | |
DE102009010614A1 (de) | Fahrzeug-Anzeigevorrichtung | |
DE102013021150A1 (de) | Verfahren zur Anzeige von optischen Informationen in Fahrzeugen | |
EP1508476A2 (de) | Fahrerassistenzsystem für Kraftfahrzeuge | |
DE19802220A1 (de) | Anzeigevorrichtung | |
EP3621848A1 (de) | Verfahren und system zur steuerung bzw. zum trainieren des fahrverhaltens eines fahrers über dessen blicklenkung | |
EP3409551A1 (de) | Verfahren und vorrichtung zur unterstützung eines in einem fahrzeug befindlichen fahrzeuginsassen | |
DE10318741A1 (de) | Fahrerassistentensystem | |
EP3790768A1 (de) | Vorrichtung und verfahren zum betreiben einer objekterkennung für den innenraum eines kraftfahrzeugs sowie ein kraftfahrzeug | |
DE102018210160A1 (de) | Am Fahrzeugdach montiertes System | |
DE102014019420A1 (de) | Kraftfahrzeugscheinwerfersystem, Kraftfahrzeug, Verfahren zum Betrieb eines Kraftfahrzeugscheinwerfersystems sowie Computerprogrammprodukt | |
DE112019007195T5 (de) | Anzeigesteuerungseinrichtung, anzeigesteuerungsverfahren und anzeigesteuerungsprogramm | |
DE102019004692B3 (de) | Vorrichtung und Verfahren zur Ermittlung von Bilddaten der Augen, von Augenpositionen und/oder einer Blickrichtung eines Fahrzeugnutzers in einem Fahrzeug | |
DE4304005A1 (de) | Überwachungsvorrichtung für Fahrzeuge | |
DE102019220012A1 (de) | Verfahren zum Betreiben einer Anzeigeeinheit eines Fahrzeugs und Anzeigeeinheit | |
EP1394756B1 (de) | Vorrichtung zur Überwachung und Beeinflussung der Wachsamkeit eines Fahrzeugführers | |
EP3682431B1 (de) | Verfahren zum betreiben eines kraftfahrzeugs mit zumindest einer aussenkamera sowie kraftfahrzeug mit zumindest einer aussenkamera | |
DE112021002765T5 (de) | Auf einer Konvexlinse/Konkavlinse basierende durchsichtige virtuelle Anzeigevorrichtung einer toten Zone direkt vor einem Fahrzeug und zugehöriges Verfahren |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 2000941906 Country of ref document: EP |
|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): JP US |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE |
|
ENP | Entry into the national phase |
Ref country code: JP Ref document number: 2000 617420 Kind code of ref document: A Format of ref document f/p: F |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 09743305 Country of ref document: US |
|
WWP | Wipo information: published in national office |
Ref document number: 2000941906 Country of ref document: EP |
|
WWG | Wipo information: grant in national office |
Ref document number: 2000941906 Country of ref document: EP |