US20170006276A1 - Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle - Google Patents

Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle Download PDF

Info

Publication number
US20170006276A1
US20170006276A1 US15/248,294 US201615248294A US2017006276A1 US 20170006276 A1 US20170006276 A1 US 20170006276A1 US 201615248294 A US201615248294 A US 201615248294A US 2017006276 A1 US2017006276 A1 US 2017006276A1
Authority
US
United States
Prior art keywords
image
lens
sensor
processing unit
imaging system
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US15/248,294
Inventor
Christoph Arndt
Urs Christen
Heike Kleinschmidt
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ford Global Technologies LLC
Original Assignee
Ford Global Technologies LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ford Global Technologies LLC filed Critical Ford Global Technologies LLC
Priority to US15/248,294 priority Critical patent/US20170006276A1/en
Publication of US20170006276A1 publication Critical patent/US20170006276A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/207Image signal generators using stereoscopic image cameras using a single 2D image sensor
    • H04N13/218Image signal generators using stereoscopic image cameras using a single 2D image sensor using spatial multiplexing
    • H04N13/0217
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R1/00Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R1/00Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
    • B60R1/20Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
    • B60R1/22Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle
    • B60R1/23Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle with a predetermined field of view
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R1/00Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
    • B60R1/20Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
    • B60R1/31Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles providing stereoscopic vision
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R11/00Arrangements for holding or mounting articles, not otherwise provided for
    • B60R11/04Mounting of cameras operative during drive; Arrangement of controls thereof relative to the vehicle
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B13/00Optical objectives specially designed for the purposes specified below
    • G02B13/06Panoramic objectives; So-called "sky lenses" including panoramic objectives having reflecting surfaces
    • G06K9/00201
    • G06K9/00798
    • G06T7/002
    • G06T7/2006
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/588Recognition of the road, e.g. of lane markings; Recognition of the vehicle driving pattern in relation to the road
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/64Three-dimensional objects
    • H04N13/0246
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/246Calibration of cameras
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R2300/00Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
    • B60R2300/10Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used
    • B60R2300/107Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used using stereoscopic cameras
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R2300/00Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
    • B60R2300/10Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used
    • B60R2300/108Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used using 'non-standard' camera systems, e.g. camera sensor used for additional purposes i.a. rain sensor, camera sensor split in multiple image areas
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R2300/00Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
    • B60R2300/40Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the details of the power supply or the coupling to vehicle components
    • B60R2300/402Image calibration
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R2300/00Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
    • B60R2300/80Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement
    • B60R2300/804Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement for lane monitoring
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60RVEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
    • B60R2300/00Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
    • B60R2300/80Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement
    • B60R2300/8093Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement for obstacle warning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • G06T2207/10012Stereo images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30248Vehicle exterior or interior
    • G06T2207/30268Vehicle interior

Definitions

  • the invention relates to an imaging system for providing an image of the surroundings of a motor vehicle, such that this image can be displayed to the driver and simultaneously used for further evaluations by image analysis.
  • U.S. Pat. No. 6,922,292 B2 discloses a wide angle imaging system for providing an image of the surroundings of a vehicle (for example a reversing image display device).
  • a camera inserted into the front side or rear side of the vehicle is used together with a lens arrangement downstream thereof, in order to present the driver with a view of the surroundings imaged by the camera on a display screen in the vehicle's interior.
  • the image is prepared by refracting and diffracting lenses together with distortion compensation by way of correction.
  • the imaging device used has a non-uniform array of electro-optical pixels, and the display screen is provided with a uniform pixilation.
  • DE 10 2004 053 416 A1 discloses a stereoscopic distance measurement system for use in a motor vehicle which has the following elements arranged in the specified sequence in the beam path of a virtual beam: a) a digital camera having an imaging optics, and having an aperture solid angle range that delimits the virtual beam emanating from the camera; b) at least one deflecting mirror element entirely or partially deflecting the virtual beam emanating from the camera; and c) at least one image field divider element having at least one divider mirror element deflecting at least a first component beam of the virtual beam emanating from the camera, the first component beam overlapping at least a second component beam of the virtual beam emanating from the camera in a restricted spatial region.
  • the corresponding CCD element is split up into two separate regions lying one above another, of which a lower region is used for the driver display. This region and an upper region (invisible to the driver) are used together for a stereoscopic evaluation, both regions having the same resolution.
  • an imaging system for a motor vehicle comprises a digital camera having a sensor, a first lens directing a first image onto a first region of the sensor, and a second lens directing a second image onto a second region of the sensor.
  • a display screen displays to a driver of the vehicle the first image and a processing unit processes data originating from the second region.
  • the lens system is configured with at least two separate lenses in order to direct separate images of the surroundings on the imaging sensor as two optically different regions, of which a first region is associated with the driver display, and a second region is associated with the data processing unit for the purpose of evaluation. It is therefore possible to adapt the regions to the different requirements, and to optimize them.
  • the lenses can be optimized independently of one another for the respective use of the regions.
  • the region for the driver display can be a large central region of the virtual image, directed on the sensor, of the surroundings in high resolution.
  • the driver can be provided with a central, more effectively resolved image, and the other region can be optimized for digital image processing, such as detection of relative movement between the vehicle and objects exterior to the vehicle.
  • the lower image segment which is preferably used for this purpose, of the image of the surroundings on the sensor permits this in a favorable way, since the observed region comprises the substratum of the road that, in turn, includes structures (lane markings, etc.) which permit simple determination of the path and/or a deviation therefrom.
  • the evaluation of this image region also permits the determination of the vehicle's longitudinal and transverse speeds.
  • the digital camera further comprises a third lens directing a third image onto a third region of the sensor, and the processing unit processes data originating from the second and third regions of the sensor and performing stereoscopic image evaluation.
  • the lens system is configured with three independent lenses directing three separate images of the surroundings onto the imaging sensor in three different regions, of which the second and third regions are associated with the data processing unit for the purpose of stereoscopic evaluation
  • the second and third lenses and image regions can be optimized for stereoscopic evaluation, in particular as concerns their distortion and spacing from one another. The spacing between the lenses is important for the stereoscopic evaluation with the aid of sequentially stored images of the surroundings or their data from the sensor.
  • the lens system is configured even with a further lens in order to direct the surroundings on the imaging sensor with the aid of an additional (fourth) optical region, and this additional region is associated with the data processing unit for the purpose of evaluations intended to detect movement
  • a direction detection unit based on an image processing unit in addition to the driver display and the stereoscopic evaluation of the determination of distance.
  • the processing unit may detects movement and/or direction of the vehicle by analyzing lane markings, curbs, or other structures appearing in the fourth image.
  • the left and right hand lateral edge regions are suitable for the stereoscopic evaluation, and the lower edge region is suitable for detecting relative movement.
  • the middle region important to the driver therefore remains in association with the driver display.
  • the first region lies in the middle region of the image, for which the driver usually has the most interest, and so said region is associated and used with the driver display on a display screen. Additionally, information of interest to the driver such as, for example, data on path and distance etc, can be overlaid onto the visual display. Again, this region can be used for the image calibration and even detection of collision with an object, without the main function of the driver display being influenced.
  • the second region preferably lies at the middle lower edge.
  • This region has also, if appropriate, already been used in the first region, or been overlapped therewith.
  • the road is included directly in front of or behind the vehicle so that the structures present there (compare above) can be used to evaluate movement, that is to say to detect the path of the vehicle by image processing.
  • the third and fourth regions preferably lie respectively at the left and right hand lateral edge outside the middle region. As already described above, these regions can be used for the stereoscopic evaluation. Moreover, they are also suitable for monitoring the dead angle as well as for determining a rotation (angle and angular velocity) of the vehicle by means of image processing methods.
  • the lens system and/or respective lens may therefore comprise prisms and/or mirrors for increasing the effective spacing of the optical regions for the purpose of evaluation, and for reducing the distortions that occur.
  • An increase in the resolution of the stereoscopy may also be achieved.
  • the data processing unit may have a buffer for the sensor data such that the data can be processed sequentially, for example during the image processing for the purpose of detecting movement and path or for stereoscopic determination of distances.
  • FIG. 1 is an overview schematic of a vehicle with an embodiment of a wide angle imaging system
  • FIG. 2 is a schematic of the division of an imaging sensor
  • FIG. 3 is a schematic of the design of an embodiment of a lens system
  • FIG. 4 is a schematic of the design of an alternative embodiment of a lens system.
  • FIG. 1 is a schematic of a vehicle 1 having an imaging system 2 for providing an image of the vehicle surroundings U.
  • the imaging system is able to display this image to the driver on a screen and to simultaneously digitally process the image for further evaluations.
  • the imaging system 2 comprises a camera 3 for imaging the surroundings U, here in particular the road S in front of or behind the vehicle 1 .
  • the camera 3 comprises a lens system 4 and a sensor 5 , the lens system 4 directing multiple images of the surroundings onto the sensor 5 .
  • Sensor 5 is preferably an electronic optical sensor and may be a charge coupled device (CCD) or a complimentary metal-oxide semiconductor (CMOS) image sensor, as is well known in the field of artificial vision.
  • CCD charge coupled device
  • CMOS complimentary metal-oxide semiconductor
  • the imaging system 2 further comprises a visual display screen 6 for displaying to the driver of the vehicle the surroundings, as imaged by the camera 3 and directed onto the sensor 5 by the lens system 4 .
  • the driver is thereby enabled when driving backward to search the surroundings for objects, etc., visible on the display screen in a known way.
  • the imaging system 2 additionally comprises a data processing unit 7 for processing the electronic data originating from the sensor 5 for the purpose of further evaluations.
  • the data processing unit 7 is “switched” logically between the sensor 5 and display 6 , and can be embodied, with appropriate suitability, by a microprocessor, which is present in any case in the onboard electronics.
  • the lens system 4 is configured with four lenses 4 a , 4 b , 4 c , 4 d in order to direct images of the surroundings onto the imaging sensor 5 as four optically different regions 5 a , 5 b , 5 c , 5 d (compare FIG. 2 ).
  • the first region 5 a is associated with the visual display screen 6 .
  • the other regions 5 b , 5 c , 5 d are, by contrast, associated with the data processing unit 7 for digital image evaluation purposes.
  • the lenses 4 a , 4 b , 4 c , 4 d can be optimized independently of one another depending on the respective use of each particular region. Characteristics that may be optimized may, for example, relate to their distortion, resolution, etc.
  • the region for the driver's visual display can be a large middle region allowing for an image of the surroundings in high resolution. The driver can thereby be provided with a central, more effectively resolved image, and it is possible for the other regions to be optimized for detection of direction and movement and/or for stereoscopic evaluation by image processing.
  • a vehicle 11 located behind the driver's own vehicle is shown with little or no distortion in region 5 a .
  • the roadway directly behind the driver's own vehicle is illustrated in region 5 b with height distortion for the purpose of detecting obstacles by processing unit 7 .
  • the region behind the driver's own vehicle is illustrated in regions 5 c and 5 d with width distortion, with in each case the same resolution for the purpose of stereoscopic evaluation performed by processing unit 7 .
  • a quasi-stereoscopic evaluation is being carried out here, by means of a single camera with the aid of different camera positions (on the basis of the movement of the vehicle).
  • the differences between the respective images on the two regions of the sensor can be used to determine angular differences between the observed objects, and to calculate a pixel displacement from them.
  • the mutual inter-axis spacing between the two regions is known per se, and is used for the further evaluation by the processing unit.
  • the two lens portions 4 a , 4 b have different optical characteristics to produce the two different images in regions 5 a and 5 b respectively.
  • the first region 5 a is located in this embodiment in the middle region of the sensor 5 which is usually of most interest for the driver, and so is associated with the driver's visual display on the display screen 6 .
  • the display of this image on display screen 6 can be overlaid with information of interest to the driver such as, for example, vehicle path and distance data, etc.
  • This region 5 a can also be used for the image calibration, and even for detection of collision with objects, without the main function of the driver display being influenced.
  • the data of region 5 a can also be processed, analyzed, and/or conditioned by the data processing unit 7 .
  • the second region 5 b lies at the middle lower edge, below the first region 5 a .
  • This region may, if appropriate, also already be used for the driver display together with the first region 5 a.
  • the region 5 b of may include the road directly in front of or behind the vehicle, and so the structures present there (compare above) may be used to evaluate movement, that is to say to detect the path of the vehicle, by image processing.
  • the lower section of the image of the surroundings, which is preferably used for this purpose, on the sensor 5 a permits this in a favorable way, since the observed region comprises the surface of the road which, in turn, includes structure (lane markings, lines, curbs, etc.) that allow a determination of the desired vehicle path or a deviation therefrom, and/or the determination of the longitudinal and transverse speeds.
  • the third and fourth regions 5 c , 5 d lie respectively at the left and right lateral edge outboard of the middle regions 5 a and 5 b .
  • the regions 5 c , 5 d may be used for stereoscopic evaluation, performed by processing unit 7 . Furthermore, they are also suitable for monitoring a blind spot as well as for determining a rotation (angle and angular velocity) of the vehicle by means of image processing methods.
  • the lens system is configured with independent lenses 4 a , 4 b , 4 c , 4 d in order to direct multiple images of the surroundings onto the three or four optically different regions 5 a , 5 b , 5 c , 5 d of the imaging sensor 5 .
  • the inter-axis spacing A between the respective central axes of lenses 4 c and 4 d and between the respective regions 5 c and 5 d is significant for the stereoscopic evaluation with the aid of sequentially stored images of the surroundings, or their data from the sensor.
  • a quasi-stereoscopic evaluation may be carried out here by means of a single camera with the aid of different camera points-of-view (on the basis of the movement of the vehicle).
  • the greater this inter-axis spacing A the more effective or more accurate is the subsequent evaluation.
  • the spacing between the two regions 5 c , 5 d of the image of the surroundings U is used in this case to determine angular differences between the objects observed in the regions, and to calculate a pixel displacement from them.
  • the spacing A between the two regions is known per se and is used for the further evaluation by processing unit 7 .
  • each lens 4 ′ c , 4 ′ d comprises a pair of mirrors 9 c , 10 c and 9 d , 10 d respectively.
  • the mirror pairs 9 c , 10 c and 9 d , 10 d are optically aligned to direct and focus the images onto regions 5 c and 5 d respectively.
  • Other focusing lenses and/or prisms may also be employed along with the mirrors to provide the desired optical characteristics.
  • the outboard placement of mirrors 9 c , 9 d results in an increase in the effective spacing A between the optical regions for the purpose of stereoscopic evaluation. An increase in the resolution of the stereoscopy may be achieved thereby.

Abstract

An imaging system includes a digital camera having a sensor (such as a charge coupled device), a first lens directing a first image onto a first region of the sensor, a second lens directing a second image onto a second region of the sensor, and a third lens directing a third image onto a third region of the sensor. A display screen displays to a driver of the vehicle the first image, and a processing unit performs stereoscopic image analysis on data originating from the second and third regions. A fourth lens may be used to direct a fourth image onto a fourth region of the sensor, and the processing unit performs calculations on data from the fourth region for the detection of movement of the vehicle.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a Divisional claiming priority on U.S. application Ser. No. 12/695,554, filed on Jan. 28, 2010, which claims foreign priority benefits under 35 U.S.C. §119(a)-(d) to DE 10 2009 000 550.1, filed on Feb. 2, 2009, which is hereby incorporated by reference in its entirety.
  • BACKGROUND
  • 1. Technical Field
  • The invention relates to an imaging system for providing an image of the surroundings of a motor vehicle, such that this image can be displayed to the driver and simultaneously used for further evaluations by image analysis.
  • 2. Background Art
  • U.S. Pat. No. 6,922,292 B2 discloses a wide angle imaging system for providing an image of the surroundings of a vehicle (for example a reversing image display device). In this case, a camera inserted into the front side or rear side of the vehicle is used together with a lens arrangement downstream thereof, in order to present the driver with a view of the surroundings imaged by the camera on a display screen in the vehicle's interior. The image is prepared by refracting and diffracting lenses together with distortion compensation by way of correction. The imaging device used has a non-uniform array of electro-optical pixels, and the display screen is provided with a uniform pixilation.
  • Another wide angle imaging system is disclosed in DE 10 2004 053 416 A1 which discloses a stereoscopic distance measurement system for use in a motor vehicle which has the following elements arranged in the specified sequence in the beam path of a virtual beam: a) a digital camera having an imaging optics, and having an aperture solid angle range that delimits the virtual beam emanating from the camera; b) at least one deflecting mirror element entirely or partially deflecting the virtual beam emanating from the camera; and c) at least one image field divider element having at least one divider mirror element deflecting at least a first component beam of the virtual beam emanating from the camera, the first component beam overlapping at least a second component beam of the virtual beam emanating from the camera in a restricted spatial region. The corresponding CCD element is split up into two separate regions lying one above another, of which a lower region is used for the driver display. This region and an upper region (invisible to the driver) are used together for a stereoscopic evaluation, both regions having the same resolution.
  • SUMMARY
  • In a disclosed embodiment of the invention, an imaging system for a motor vehicle comprises a digital camera having a sensor, a first lens directing a first image onto a first region of the sensor, and a second lens directing a second image onto a second region of the sensor. A display screen displays to a driver of the vehicle the first image and a processing unit processes data originating from the second region. The lens system is configured with at least two separate lenses in order to direct separate images of the surroundings on the imaging sensor as two optically different regions, of which a first region is associated with the driver display, and a second region is associated with the data processing unit for the purpose of evaluation. It is therefore possible to adapt the regions to the different requirements, and to optimize them. Thus, the lenses can be optimized independently of one another for the respective use of the regions. Thus, the region for the driver display can be a large central region of the virtual image, directed on the sensor, of the surroundings in high resolution. Thus, the driver can be provided with a central, more effectively resolved image, and the other region can be optimized for digital image processing, such as detection of relative movement between the vehicle and objects exterior to the vehicle.
  • The lower image segment, which is preferably used for this purpose, of the image of the surroundings on the sensor permits this in a favorable way, since the observed region comprises the substratum of the road that, in turn, includes structures (lane markings, etc.) which permit simple determination of the path and/or a deviation therefrom. The evaluation of this image region also permits the determination of the vehicle's longitudinal and transverse speeds.
  • In another embodiment, the digital camera further comprises a third lens directing a third image onto a third region of the sensor, and the processing unit processes data originating from the second and third regions of the sensor and performing stereoscopic image evaluation. When the lens system is configured with three independent lenses directing three separate images of the surroundings onto the imaging sensor in three different regions, of which the second and third regions are associated with the data processing unit for the purpose of stereoscopic evaluation, the second and third lenses and image regions can be optimized for stereoscopic evaluation, in particular as concerns their distortion and spacing from one another. The spacing between the lenses is important for the stereoscopic evaluation with the aid of sequentially stored images of the surroundings or their data from the sensor.
  • When the lens system is configured even with a further lens in order to direct the surroundings on the imaging sensor with the aid of an additional (fourth) optical region, and this additional region is associated with the data processing unit for the purpose of evaluations intended to detect movement, it is further possible to provide a direction detection unit based on an image processing unit in addition to the driver display and the stereoscopic evaluation of the determination of distance. The processing unit may detects movement and/or direction of the vehicle by analyzing lane markings, curbs, or other structures appearing in the fourth image.
  • In particular, the left and right hand lateral edge regions are suitable for the stereoscopic evaluation, and the lower edge region is suitable for detecting relative movement. The middle region important to the driver therefore remains in association with the driver display.
  • In the disclosed embodiment of the invention, the first region lies in the middle region of the image, for which the driver usually has the most interest, and so said region is associated and used with the driver display on a display screen. Additionally, information of interest to the driver such as, for example, data on path and distance etc, can be overlaid onto the visual display. Again, this region can be used for the image calibration and even detection of collision with an object, without the main function of the driver display being influenced.
  • The second region preferably lies at the middle lower edge. This region has also, if appropriate, already been used in the first region, or been overlapped therewith. In this region of the image, the road is included directly in front of or behind the vehicle so that the structures present there (compare above) can be used to evaluate movement, that is to say to detect the path of the vehicle by image processing.
  • The third and fourth regions preferably lie respectively at the left and right hand lateral edge outside the middle region. As already described above, these regions can be used for the stereoscopic evaluation. Moreover, they are also suitable for monitoring the dead angle as well as for determining a rotation (angle and angular velocity) of the vehicle by means of image processing methods.
  • It may therefore be advantageous for the lens system and/or respective lens to comprise prisms and/or mirrors for increasing the effective spacing of the optical regions for the purpose of evaluation, and for reducing the distortions that occur. An increase in the resolution of the stereoscopy may also be achieved.
  • For evaluation of the data supplied by the sensor, the data processing unit may have a buffer for the sensor data such that the data can be processed sequentially, for example during the image processing for the purpose of detecting movement and path or for stereoscopic determination of distances.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Further details, features and advantages of the invention emerge from the following description of exemplary embodiments with the aid of the single drawing, in which:
  • FIG. 1 is an overview schematic of a vehicle with an embodiment of a wide angle imaging system;
  • FIG. 2 is a schematic of the division of an imaging sensor;
  • FIG. 3 is a schematic of the design of an embodiment of a lens system; and
  • FIG. 4 is a schematic of the design of an alternative embodiment of a lens system.
  • DETAILED DESCRIPTION
  • FIG. 1 is a schematic of a vehicle 1 having an imaging system 2 for providing an image of the vehicle surroundings U. The imaging system is able to display this image to the driver on a screen and to simultaneously digitally process the image for further evaluations.
  • The imaging system 2 comprises a camera 3 for imaging the surroundings U, here in particular the road S in front of or behind the vehicle 1. The camera 3 comprises a lens system 4 and a sensor 5, the lens system 4 directing multiple images of the surroundings onto the sensor 5. Sensor 5 is preferably an electronic optical sensor and may be a charge coupled device (CCD) or a complimentary metal-oxide semiconductor (CMOS) image sensor, as is well known in the field of artificial vision.
  • The imaging system 2 further comprises a visual display screen 6 for displaying to the driver of the vehicle the surroundings, as imaged by the camera 3 and directed onto the sensor 5 by the lens system 4. The driver is thereby enabled when driving backward to search the surroundings for objects, etc., visible on the display screen in a known way.
  • The imaging system 2 additionally comprises a data processing unit 7 for processing the electronic data originating from the sensor 5 for the purpose of further evaluations. The data processing unit 7 is “switched” logically between the sensor 5 and display 6, and can be embodied, with appropriate suitability, by a microprocessor, which is present in any case in the onboard electronics.
  • As seen in FIG. 3, the lens system 4 is configured with four lenses 4 a, 4 b, 4 c, 4 d in order to direct images of the surroundings onto the imaging sensor 5 as four optically different regions 5 a, 5 b, 5 c, 5 d (compare FIG. 2). The first region 5 a is associated with the visual display screen 6. The other regions 5 b, 5 c, 5 d are, by contrast, associated with the data processing unit 7 for digital image evaluation purposes.
  • Thus, the lenses 4 a, 4 b, 4 c, 4 d can be optimized independently of one another depending on the respective use of each particular region. Characteristics that may be optimized may, for example, relate to their distortion, resolution, etc. Thus, the region for the driver's visual display can be a large middle region allowing for an image of the surroundings in high resolution. The driver can thereby be provided with a central, more effectively resolved image, and it is possible for the other regions to be optimized for detection of direction and movement and/or for stereoscopic evaluation by image processing.
  • As seen in FIG. 2, a vehicle 11 located behind the driver's own vehicle is shown with little or no distortion in region 5 a. The roadway directly behind the driver's own vehicle is illustrated in region 5 b with height distortion for the purpose of detecting obstacles by processing unit 7. The region behind the driver's own vehicle is illustrated in regions 5 c and 5 d with width distortion, with in each case the same resolution for the purpose of stereoscopic evaluation performed by processing unit 7.
  • A quasi-stereoscopic evaluation is being carried out here, by means of a single camera with the aid of different camera positions (on the basis of the movement of the vehicle). The greater this inter-axis spacing between the respective central axes of the lenses, the more effective or accurate is the subsequent evaluation. The differences between the respective images on the two regions of the sensor can be used to determine angular differences between the observed objects, and to calculate a pixel displacement from them. The mutual inter-axis spacing between the two regions is known per se, and is used for the further evaluation by the processing unit.
  • It is also possible to combine two of more of the lenses so as, as illustrated here in the exemplary embodiment, to combine the two central lenses 4 a and 4 b into a combined central lens 8 a/b that consists of an upper lens portion 4 a and a lower lens portion 4 b. The two lens portions 4 a, 4 b have different optical characteristics to produce the two different images in regions 5 a and 5 b respectively.
  • The first region 5 a is located in this embodiment in the middle region of the sensor 5 which is usually of most interest for the driver, and so is associated with the driver's visual display on the display screen 6. The display of this image on display screen 6 can be overlaid with information of interest to the driver such as, for example, vehicle path and distance data, etc. This region 5 a can also be used for the image calibration, and even for detection of collision with objects, without the main function of the driver display being influenced. The data of region 5 a can also be processed, analyzed, and/or conditioned by the data processing unit 7.
  • The second region 5 b lies at the middle lower edge, below the first region 5 a. This region may, if appropriate, also already be used for the driver display together with the first region 5 a.
  • The region 5 b of may include the road directly in front of or behind the vehicle, and so the structures present there (compare above) may be used to evaluate movement, that is to say to detect the path of the vehicle, by image processing. The lower section of the image of the surroundings, which is preferably used for this purpose, on the sensor 5 a permits this in a favorable way, since the observed region comprises the surface of the road which, in turn, includes structure (lane markings, lines, curbs, etc.) that allow a determination of the desired vehicle path or a deviation therefrom, and/or the determination of the longitudinal and transverse speeds.
  • The third and fourth regions 5 c, 5 d lie respectively at the left and right lateral edge outboard of the middle regions 5 a and 5 b. As already described above, the regions 5 c, 5 d may be used for stereoscopic evaluation, performed by processing unit 7. Furthermore, they are also suitable for monitoring a blind spot as well as for determining a rotation (angle and angular velocity) of the vehicle by means of image processing methods.
  • As best seen in FIGS. 3 and 4, the lens system is configured with independent lenses 4 a, 4 b, 4 c, 4 d in order to direct multiple images of the surroundings onto the three or four optically different regions 5 a, 5 b, 5 c, 5 d of the imaging sensor 5.
  • The inter-axis spacing A between the respective central axes of lenses 4 c and 4 d and between the respective regions 5 c and 5 d is significant for the stereoscopic evaluation with the aid of sequentially stored images of the surroundings, or their data from the sensor. A quasi-stereoscopic evaluation may be carried out here by means of a single camera with the aid of different camera points-of-view (on the basis of the movement of the vehicle). The greater this inter-axis spacing A, the more effective or more accurate is the subsequent evaluation. The spacing between the two regions 5 c, 5 d of the image of the surroundings U is used in this case to determine angular differences between the objects observed in the regions, and to calculate a pixel displacement from them. The spacing A between the two regions is known per se and is used for the further evaluation by processing unit 7.
  • Referring now to FIG. 4, each lens 4c, 4d comprises a pair of mirrors 9 c, 10 c and 9 d, 10 d respectively. The mirror pairs 9 c, 10 c and 9 d, 10 d are optically aligned to direct and focus the images onto regions 5 c and 5 d respectively. Other focusing lenses and/or prisms (not shown) may also be employed along with the mirrors to provide the desired optical characteristics. The outboard placement of mirrors 9 c, 9 d results in an increase in the effective spacing A between the optical regions for the purpose of stereoscopic evaluation. An increase in the resolution of the stereoscopy may be achieved thereby.

Claims (19)

What is claimed is:
1. An imaging system comprising:
an electro-optical sensor;
a first lens directing an image onto the sensor;
a display screen displaying the image;
second and third inter-axially spaced lenses directing respective second and third images onto the sensor on opposite sides of the image, the second and third lenses producing purposeful width distortion of the respective images adapted for stereoscopic analysis; and
a processing unit performing stereoscopic analysis on the second and third images.
2. The imaging system of claim 1, further comprising a fourth lens inter-axially spaced from the first, the second, and the third lenses to direct a fourth image onto a fourth region of the sensor located adjacent a third edge of the sensor, the fourth image exhibiting purposeful height distortion.
3. The imaging system of claim 2, wherein the processing unit performs calculations on data from the fourth region to detect lateral movement of a host vehicle.
4. The imaging system of claim 3, wherein the processing unit detects lateral movement of the vehicle by analyzing lane markings appearing in the fourth image.
5. The imaging system of claim 1, wherein at least one of the lenses comprises at least two mirrors optically aligned with one another.
6. The imaging system of claim 1, further comprising a data buffer associated with the processing unit.
7. An imaging system for a motor vehicle comprising:
an electro-optical sensor;
a first lens directing a first image onto a central region of the sensor;
a display screen for displaying the first image to a driver of the vehicle;
a second lens inter-axially spaced from the first lens to direct a second image onto a second region of the sensor located adjacent a first lateral edge of the sensor;
a third lens inter-axially spaced from the first lens and the second lens to direct a third image onto a third region of the sensor located adjacent a second lateral edge of the sensor, the second lateral edge positioned opposite from the first lateral edge, the second and third images having substantially equal amounts of purposeful width distortion adapted for stereoscopic image analysis; and
a processing unit performing stereoscopic image analysis on data originating from the second and third regions.
8. The imaging system of claim 7, further comprising a fourth lens inter-axially spaced from the lens, the second lens, and the third lens to direct a fourth image onto the sensor, the fourth image exhibiting height distortion.
9. The imaging system of claim 8, wherein the processing unit performs analysis on the fourth image to detect lateral movement of a host vehicle.
10. The imaging system of claim 9, wherein the processing unit detects lateral movement of the vehicle by analyzing lane markings appearing in the fourth image.
11. The imaging system of claim 7, further comprising a data buffer associated with the processing unit.
12. Apparatus comprising:
a first lens directing a first image onto a first region of an electro-optical sensor; and
second and third lenses spaced inter-axially from one another and located on opposite sides of the first lens and directing respective second and third images having substantially equal amounts of purposeful width distortion adapted for stereoscopic evaluation onto respective second and third regions of the sensor.
13. The apparatus of claim 12, further comprising a processing unit performing stereoscopic analysis on the second and third images.
14. The apparatus of claim 13, further comprising a data buffer associated with the processing unit.
15. The apparatus of claim 12, further comprising a display screen for displaying the first image to a driver of a host vehicle.
16. The apparatus of claim 12, wherein at least one of the lenses comprises at least two mirrors optically aligned with one another.
17. The apparatus of claim 12, further comprising a fourth lens inter-axially spaced from the first lens, the second lens, and the third lens to direct a fourth image onto the sensor, the fourth image exhibiting purposeful height distortion.
18. The apparatus of claim 17, further comprising a processing unit performing analysis on the fourth image to detect lateral movement of a host vehicle.
19. The apparatus of claim 18, wherein the processing unit detects lateral movement of the vehicle by analyzing lane markings appearing in the fourth image.
US15/248,294 2009-02-02 2016-08-26 Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle Abandoned US20170006276A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US15/248,294 US20170006276A1 (en) 2009-02-02 2016-08-26 Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
DE102009000550.1 2009-02-02
DE102009000550.1A DE102009000550B4 (en) 2009-02-02 2009-02-02 Wide-angle imaging system for providing an image of the surroundings of a vehicle, in particular of a motor vehicle
US12/695,554 US9469249B2 (en) 2009-02-02 2010-01-28 Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle
US15/248,294 US20170006276A1 (en) 2009-02-02 2016-08-26 Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US12/695,554 Division US9469249B2 (en) 2009-02-02 2010-01-28 Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle

Publications (1)

Publication Number Publication Date
US20170006276A1 true US20170006276A1 (en) 2017-01-05

Family

ID=42317242

Family Applications (2)

Application Number Title Priority Date Filing Date
US12/695,554 Expired - Fee Related US9469249B2 (en) 2009-02-02 2010-01-28 Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle
US15/248,294 Abandoned US20170006276A1 (en) 2009-02-02 2016-08-26 Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US12/695,554 Expired - Fee Related US9469249B2 (en) 2009-02-02 2010-01-28 Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle

Country Status (3)

Country Link
US (2) US9469249B2 (en)
CN (1) CN101881923B (en)
DE (1) DE102009000550B4 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20230027882A1 (en) * 2021-07-22 2023-01-26 Argo AI, LLC Motorized Mounting Device for Positioning an Optical Element Within a Field-of-View of an Optical Sensor and Method of Use

Families Citing this family (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102009000550B4 (en) * 2009-02-02 2018-10-04 Ford Global Technologies, Llc Wide-angle imaging system for providing an image of the surroundings of a vehicle, in particular of a motor vehicle
JP2010263272A (en) * 2009-04-30 2010-11-18 Koito Mfg Co Ltd Monitoring camera for vehicle and monitoring camera system for vehicle
JP6019998B2 (en) * 2012-02-17 2016-11-02 ソニー株式会社 Imaging apparatus, imaging control program, and imaging method
WO2013126715A2 (en) * 2012-02-22 2013-08-29 Magna Electronics, Inc. Vehicle camera system with image manipulation
US9242602B2 (en) * 2012-08-27 2016-01-26 Fotonation Limited Rearview imaging systems for vehicle
KR101428296B1 (en) 2012-12-18 2014-08-08 현대오트론 주식회사 Camera module for vehicle and monitoring system provided with the same
DE102013012216B4 (en) * 2013-07-23 2016-12-29 Audi Ag Optical sensor arrangement for a motor vehicle and motor vehicle
JP6200481B2 (en) * 2015-11-25 2017-09-20 株式会社Subaru Outside environment recognition device
JP6310899B2 (en) 2015-11-25 2018-04-11 株式会社Subaru Outside environment recognition device
US10023120B2 (en) * 2016-03-30 2018-07-17 Delphi Technologies, Inc. Multi-purpose camera device for use on a vehicle
CN107021016A (en) * 2016-11-08 2017-08-08 王向阳 Motor vehicle light guide blind-area-free panoramic display device
CN106791617B (en) * 2016-12-02 2019-11-05 湖南农业大学 Farm machinery wireless video control loop and its video switching method
WO2018103795A1 (en) * 2016-12-06 2018-06-14 Conti Temic Microelectronic Gmbh Camera device and method for capturing a surrounding region of a vehicle in a situation-adapted manner
JP6837931B2 (en) * 2017-06-26 2021-03-03 アルパイン株式会社 Display control device, display control method and camera monitoring system
DE102017210845A1 (en) * 2017-06-27 2018-12-27 Conti Temic Microelectronic Gmbh Camera apparatus and method for environmental detection of an environmental area of a vehicle
US11341607B2 (en) * 2019-06-07 2022-05-24 Texas Instruments Incorporated Enhanced rendering of surround view images
CN110667477A (en) * 2019-11-11 2020-01-10 西南交通大学 Vehicle night video processing system and method

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060029256A1 (en) * 2004-08-09 2006-02-09 Takashi Miyoshi Method of generating image and device
US20070090311A1 (en) * 2005-10-21 2007-04-26 C.R.F. Societa Consortile Per Azioni Orbassano (Torino), Italy Optical sensor device to be installed on board a motor-vehicle for aid in driving and/or for automatic activation of systems provided on the motor-vehicle

Family Cites Families (34)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4457584A (en) * 1982-03-08 1984-07-03 Pryor Eugene F Stereoscopic viewer with variable fields of vision
DE3441745C2 (en) * 1984-11-15 1986-11-13 Jos. Schneider Optische Werke Kreuznach GmbH & Co KG, 6550 Bad Kreuznach Room image viewing device
US4989078A (en) * 1988-08-15 1991-01-29 Eastman Kodak Company Still video camera for recording stereo images on a video disk
US5142357A (en) * 1990-10-11 1992-08-25 Stereographics Corp. Stereoscopic video camera with image sensors having variable effective position
US5063441A (en) * 1990-10-11 1991-11-05 Stereographics Corporation Stereoscopic video cameras with image sensors having variable effective position
JP3167752B2 (en) * 1991-10-22 2001-05-21 富士重工業株式会社 Vehicle distance detection device
US5756988A (en) * 1995-06-27 1998-05-26 Furuta; Yoshihisa Device for checking lateral views at front/rear ends of vehicle
WO1999030508A1 (en) * 1997-12-05 1999-06-17 Mcgill University Stereoscopic gaze controller
US6201642B1 (en) 1999-07-27 2001-03-13 Donnelly Corporation Vehicular vision system with a wide angle lens including a diffractive element
US6717610B1 (en) * 1998-11-25 2004-04-06 Donnelly Corporation Wide angle image capture system for vehicle
JP2000295604A (en) * 1999-04-06 2000-10-20 Yazaki Corp Rear and side monitoring device for vehicle
US6757109B2 (en) * 1999-07-27 2004-06-29 Donnelly Corporation Plastic lens system for vehicle imaging system
US6963661B1 (en) * 1999-09-09 2005-11-08 Kabushiki Kaisha Toshiba Obstacle detection system and method therefor
US7852462B2 (en) * 2000-05-08 2010-12-14 Automotive Technologies International, Inc. Vehicular component control methods based on blind spot monitoring
KR20020033816A (en) * 2000-07-19 2002-05-07 마츠시타 덴끼 산교 가부시키가이샤 Monitoring system
JP3759429B2 (en) * 2001-05-23 2006-03-22 株式会社東芝 Obstacle detection apparatus and method
US7697027B2 (en) * 2001-07-31 2010-04-13 Donnelly Corporation Vehicular video system
DE10149115A1 (en) * 2001-10-05 2003-04-17 Bosch Gmbh Robert Object detection device for motor vehicle driver assistance systems checks data measured by sensor systems for freedom from conflict and outputs fault signal on detecting a conflict
US7253833B2 (en) * 2001-11-16 2007-08-07 Autonetworks Technologies, Ltd. Vehicle periphery visual recognition system, camera and vehicle periphery monitoring apparatus and vehicle periphery monitoring system
JP2003186085A (en) * 2001-12-14 2003-07-03 Auto Network Gijutsu Kenkyusho:Kk Vehicle circumference monitor device
JP4052650B2 (en) * 2004-01-23 2008-02-27 株式会社東芝 Obstacle detection device, method and program
US7131740B2 (en) * 2004-03-17 2006-11-07 Olympus Corporation Optical system and optical apparatus provided with the same
JP2006050263A (en) * 2004-08-04 2006-02-16 Olympus Corp Image generation method and device
DE102004053416A1 (en) 2004-11-05 2006-05-11 Robert Bosch Gmbh Stereoscopic distance measurement system to determine distance of object from motor vehicle has splitter mirror element to deflect first part of virtual beam bundle from camera which is then overlapped by second part of beam bundle
JP2006159933A (en) * 2004-12-02 2006-06-22 Denso Corp Vehicle periphery display assistance system
DE102005043411A1 (en) 2005-09-13 2007-03-15 Robert Bosch Gmbh Low-cost stereo optics system
ITTO20060214A1 (en) * 2006-03-22 2007-09-23 Kria S R L VEHICLE DETECTION SYSTEM
JP2008180808A (en) * 2007-01-23 2008-08-07 Funai Electric Co Ltd Stereoscopic imaging apparatus
US8358332B2 (en) * 2007-07-23 2013-01-22 Disney Enterprises, Inc. Generation of three-dimensional movies with improved depth control
KR100882011B1 (en) * 2007-07-29 2009-02-04 주식회사 나노포토닉스 Methods of obtaining panoramic images using rotationally symmetric wide-angle lenses and devices thereof
DE102007044536A1 (en) * 2007-09-18 2009-03-19 Bayerische Motoren Werke Aktiengesellschaft Device for monitoring the environment of a motor vehicle
DE102007044535B4 (en) * 2007-09-18 2022-07-14 Bayerische Motoren Werke Aktiengesellschaft Method for driver information in a motor vehicle
DE102009000550B4 (en) * 2009-02-02 2018-10-04 Ford Global Technologies, Llc Wide-angle imaging system for providing an image of the surroundings of a vehicle, in particular of a motor vehicle
DE102010041490A1 (en) * 2010-09-27 2012-03-29 Carl Zeiss Microimaging Gmbh Optical instrument and method for optical monitoring

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060029256A1 (en) * 2004-08-09 2006-02-09 Takashi Miyoshi Method of generating image and device
US20070090311A1 (en) * 2005-10-21 2007-04-26 C.R.F. Societa Consortile Per Azioni Orbassano (Torino), Italy Optical sensor device to be installed on board a motor-vehicle for aid in driving and/or for automatic activation of systems provided on the motor-vehicle

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20230027882A1 (en) * 2021-07-22 2023-01-26 Argo AI, LLC Motorized Mounting Device for Positioning an Optical Element Within a Field-of-View of an Optical Sensor and Method of Use
US11800085B2 (en) * 2021-07-22 2023-10-24 Argo AI, LLC Motorized mounting device for positioning an optical element within a field-of-view of an optical sensor and method of use
US20230344982A1 (en) * 2021-07-22 2023-10-26 Argo Al, LLC Motorized Mounting Device for Positioning an Optical Element Within a Field-of-View of an Optical Sensor and Method of Use

Also Published As

Publication number Publication date
US20100194889A1 (en) 2010-08-05
CN101881923A (en) 2010-11-10
US9469249B2 (en) 2016-10-18
DE102009000550B4 (en) 2018-10-04
CN101881923B (en) 2015-09-09
DE102009000550A1 (en) 2010-08-12

Similar Documents

Publication Publication Date Title
US20170006276A1 (en) Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle
US7385680B2 (en) Camera module
US10899277B2 (en) Vehicular vision system with reduced distortion display
US6611202B2 (en) Vehicle camera display system
US7538864B2 (en) Vehicle wheel alignment system scanned beam imaging sensor
US8130270B2 (en) Vehicle-mounted image capturing apparatus
US20160379066A1 (en) Method and Camera System for Distance Determination of Objects from a Vehicle
US10183621B2 (en) Vehicular image processing apparatus and vehicular image processing system
US9426364B2 (en) Image processing apparatus and image processing method
JP4425495B2 (en) Outside monitoring device
US8009868B2 (en) Method of processing images photographed by plural cameras and apparatus for the same
JP2015057341A (en) Transportation vehicle visible system for displaying statutory visibility by main mirror and wide-angle mirror
US10869002B2 (en) Vehicle camera device for capturing the surroundings of a motor vehicle and driver assistance device for detecting objects with such a vehicle camera device
US10635914B2 (en) Optical test device for a vehicle camera and testing method
JP3765862B2 (en) Vehicle environment recognition device
US9649990B2 (en) Device and method for optically recording the underbody of a vehicle
US20050057651A1 (en) Apparatus for visually confirming vehicle periphery
JP2008037118A (en) Display for vehicle
CN113596389A (en) Method for aligning a camera of a vehicle camera surveillance system with a vehicle
US20230283758A1 (en) Multi-aperture zoom digital cameras and methods of using same
US8384779B2 (en) Display device for vehicle
US20220089103A1 (en) Vehicular camera test system using true and simulated targets to determine camera defocus
KR20210036386A (en) Camera monitoring system
CN102566010B (en) Imaging system and fisheye lens system
GB2541101A (en) Method and camera system for determining the distance of objects in relation to a vehicle

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION