EP3707679A1 - Method for displaying an image of the area surrounding a vehicle on a touch screen installed in the vehicle - Google Patents
Method for displaying an image of the area surrounding a vehicle on a touch screen installed in the vehicleInfo
- Publication number
- EP3707679A1 EP3707679A1 EP18795667.7A EP18795667A EP3707679A1 EP 3707679 A1 EP3707679 A1 EP 3707679A1 EP 18795667 A EP18795667 A EP 18795667A EP 3707679 A1 EP3707679 A1 EP 3707679A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- vehicle
- virtual camera
- image
- touch screen
- posture
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 26
- 238000004364 calculation method Methods 0.000 claims description 27
- 230000003287 optical effect Effects 0.000 claims description 11
- 239000000203 mixture Substances 0.000 claims description 2
- 230000036544 posture Effects 0.000 description 49
- 230000006870 function Effects 0.000 description 5
- 239000011159 matrix material Substances 0.000 description 3
- 239000002131 composite material Substances 0.000 description 2
- FGUUSXIOTUKUDN-IBGZPJMESA-N C1(=CC=CC=C1)N1C2=C(NC([C@H](C1)NC=1OC(=NN=1)C1=CC=CC=C1)=O)C=CC=C2 Chemical compound C1(=CC=CC=C1)N1C2=C(NC([C@H](C1)NC=1OC(=NN=1)C1=CC=CC=C1)=O)C=CC=C2 FGUUSXIOTUKUDN-IBGZPJMESA-N 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 239000003550 marker Substances 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
- 238000000844 transformation Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/10—Geometric effects
- G06T15/20—Perspective computation
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/10—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used
- B60R2300/102—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used using 360 degree surveillance camera system
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/60—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by monitoring and displaying vehicle exterior scenes from a transformed perspective
- B60R2300/602—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by monitoring and displaying vehicle exterior scenes from a transformed perspective with an adjustable viewpoint
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
- G06T2200/24—Indexing scheme for image data processing or generation, in general involving graphical user interfaces [GUIs]
Definitions
- the present invention relates generally to the field of driving aids for motor vehicles.
- It relates in particular to a method of displaying an image of the environment of a vehicle on a touch screen fitted to the vehicle.
- each of the four image sensors being placed on one of the four sides of the vehicle, each of the four image sensors being able to acquire a real image of the environment of the vehicle,
- a computing unit adapted to implement a display method as mentioned above.
- this composite image seems to be acquired from a single virtual camera filming the motor vehicle from a given location located remote from the vehicle.
- the location of this virtual camera can generally be chosen by the driver by displaying a representation of the motor vehicle and a representation of the virtual camera and inviting the driver to move the virtual camera around the vehicle by swiping the touch screen with his fingers.
- the help device then composes the image of the vehicle environment from the selected location of the virtual camera.
- the driver may need to change the location of the virtual camera while he is performing a maneuver.
- the selection of the location of the virtual camera using this device is not practical, and may require too much attention from the driver, which impacts his safety.
- the present invention proposes a method of displaying an image of the environment of a vehicle on a touch screen fitted to the vehicle, comprising steps of:
- the invention provides a method for the user to easily choose a new posture of the virtual camera.
- the choice is indeed simplified by the display of the graphic element to show the user which postures can be adopted by the virtual camera.
- the user selects the new posture of the virtual camera by exerting a simple press on the graphic element the touch screen.
- the selection of a new posture does not require much attention on the part of the user who can then focus on the road maneuvers he is performing.
- a step is provided for constituting a representation of the vehicle as seen from the virtual camera in the new posture determined in step c), and in which, in step e), the virtual image is obtained by superimposing the representation of the vehicle on the environment,
- the graphic element comprises a circle situated around the vehicle image and said posture parameter is an orientation angle of the virtual camera around the vehicle,
- the graphic element makes it possible to select the angle of orientation of the virtual camera around the vehicle with a step of 1 degree
- the graphic element comprises a segment situated vertically with respect to the image of the vehicle and said posture parameter is a vertical distance between the vehicle and the virtual camera,
- the graphic element comprises a segment situated horizontally with respect to the image of the vehicle and the posture parameter is a horizontal distance between the vehicle and the virtual camera,
- Steps c) to f) are repeated after each point support exerted by the user on the touch screen,
- the invention also proposes a help system; driving a vehicle as defined in the introduction of which the computing unit is programmed to implement a method as defined above.
- FIG. 1 is a diagrammatic section of a vehicle equipped with a driving assistance system according to the invention
- FIG. 2a illustrates a virtual camera of the vehicle of FIG. 1
- FIG. 2b illustrates a virtual image of the vehicle of FIG. 1 as seen by the virtual camera
- FIG. 3a illustrates parameters of the virtual camera
- FIG. 3b is a schematic view from above of the vehicle of FIG. 1 and of the virtual camera, on which are represented coordinates of the virtual camera,
- FIG. 3c is a schematic side view of the vehicle of FIG. 1 and of the virtual camera, on which are represented other coordinates of the virtual camera,
- FIG. 4a is a schematic perspective view of the vehicle of FIG. 1, on which is represented a first graphic element associated with a first posture parameter of the virtual camera,
- FIG. 4b is a schematic perspective view of the vehicle of FIG. 1, on which is represented a second graphic element associated with a second posture parameter of the virtual camera,
- FIG. 4c is a schematic perspective view of the vehicle of FIG. 1, on which is represented a third graphic element associated with a third posture parameter of the virtual camera,
- FIG. 5 represents a diagram illustrating the implementation steps of a display method according to the invention.
- FIG. 6 represents a perspective virtual image obtained according to the display method illustrated in FIG. 5.
- FIG. 1 diagrammatically shows an automobile vehicle 1 equipped with a driving assistance system 3.
- this driver assistance system 3 comprises image sensors 5, a touch screen 7 and a calculation unit 9.
- the image sensors 5 are able to acquire a plurality of real images of an environment outside the vehicle 1.
- the driver assistance system 3 comprises four image sensors 5.
- Each of the image sensors 5 is placed on one of the four sides of the vehicle 1 (only two of the sensors 5 are visible in FIG. ).
- a first image sensor (not shown) is placed at the front of the vehicle 1, for example below the vehicle logo 1.
- the first image sensor captures an area located at the front of the vehicle 1.
- a second image sensor (not shown) is placed at the rear of the vehicle, for example above the license plate.
- the second image sensor captures an area at the rear of the vehicle 1.
- a third image sensor 5 is placed on a first lateral side, for example the right side of the vehicle, for example under the right mirror.
- the third image sensor captures a first zone located laterally with respect to the vehicle 1, here to the right of the vehicle 1.
- a fourth image sensor 5 is placed on a second lateral side, for example the left side of the vehicle, for example under the left rearview mirror.
- the fourth image sensor 5 captures a second zone situated laterally with respect to the vehicle 1, here to the left of the vehicle 1.
- These image sensors 5 are here cameras. These cameras can be analog or digital. Here, digital cameras are preferably used because they make it possible to capture images of higher resolution than analog cameras, which subsequently makes it possible to obtain better results by applying the method of the invention.
- the cameras have a large opening angle, for example close to or greater than 180 °. These cameras are equipped with lenses known as "fish-eye" in English.
- a 'this stage can be defined with reference to figure 3b a vehicle reference frame (0, Xv, Yv, Zv) attached to the vehicle 1.
- the marker is considered an orthonormal coordinate system whose origin O is at the center of front side of the vehicle 1, the longitudinal axis (OXv) of which is oriented toward the rear of the vehicle 1, the transverse axis (OYv) of which is oriented towards the right of the vehicle and whose vertical axis (OZv) is oriented to the top.
- center C of the vehicle 1 is also possible to define the center C of the vehicle 1 as the point situated halfway between the vehicle 1 in the longitudinal direction, in the transverse direction and in the vertical direction of the vehicle 1.
- the calculation unit 9 of the vehicle 1 is able to receive the actual images acquired by the image sensors 5. It is furthermore programmed to compose a virtual image 13 of the environment of the vehicle 1 from the actual images acquired by the image sensors 5. Such a virtual image 13 is illustrated in FIG. 2b.
- the computing unit 9 is programmed to compose a virtual image 13 of the environment of the vehicle 1 as seen by a virtual camera 1 1 located at a distance from the vehicle 1.
- the virtual camera 11 is visible for example on Figure 2a.
- the virtual camera 1 1 is located above the vehicle 1. It will be noted that the virtual camera 11 does not exist in the real world: it is only a concept making it easier to understand the invention, which consists in composing a virtual image 13 corresponding to an image real as it would be captured by an actual camera if this actual camera was placed in the position of the virtual camera 11.
- FIG. 2b illustrates the virtual image 13 as seen by the virtual camera 11.
- the virtual image 13 comprises a representation 15 of the vehicle 1, viewed from above, as well as the reconstituted environment 17 of the vehicle 1 seen under the same angle.
- the optical axis Ao of the virtual camera 11 is chosen to pass through a fixed point of the vehicle 1, regardless of the position of the virtual camera 11.
- the fixed point is for example the center C of the vehicle 1.
- Figure 3a shows some parameters of the virtual camera 11.
- the virtual camera 11 thus has a focal length f and a focal plane PF.
- the focal plane PF has a width Cx and a height Cy.
- the virtual camera further has a horizontal resolution Resh and a vertical resolution Resv.
- the virtual camera 1 1 has a posture, that is to say a position and an orientation, which is marked by coordinates expressed in the vehicle reference (0, Xv, Yv, Zv). Coordinates of the virtual camera 1 1 comprise a first coordinate X Cam defined along the axis (OXv), a second coordinate Ycam defined along the axis (OYv), and a third coordinate -icam defined along the axis (OZv).
- a fourth coordinate G is represented by Figure 3b.
- the fourth coordinate ⁇ corresponds to an orientation angle ⁇ between the longitudinal axis (OXv) and the projected optical axis A 0 in the plane
- a fifth coordinate ⁇ is represented by FIG. 3c.
- the fifth coordinate ⁇ corresponds to an angle of inclination ⁇ formed by the projected optical axis A 0 in the plane and the axis
- the calculation unit 9 is able to calculate the coordinates
- a first, a second and a third posture parameter can be defined here, which the user can easily vary by pressing "graphic elements" displayed on the touch screen 7 in order to modify the posture of the virtual camera 1 1 in relation to the vehicle 1.
- the first posture parameter is the orientation angle ⁇ of the virtual camera 11.
- a first graphical element 19, shown in FIG. 4a, is associated with this first posture parameter.
- This first graphic element 19 comprises a first circle C1 of radius R1 centered on the center C.
- the first circle C1 is located in the plane (XvYv).
- the value of the first radius R1 is determined so that the first circle C1 encircles the representation 15 of the vehicle 1.
- the first radius R1 is for example equal to 2.5 m.
- the first graphic element 19 is stored in a memory unit, from which it can be retrieved by the computing unit 9 which is able to transmit it to the touch screen 7.
- the second posture parameter is the third coordinate Zcam which is reminded that it represents a vertical distance between the vehicle 1 and the virtual camera 11.
- This third Zcam coordinate has a range of values that is preset and is stored in the memory unit.
- the value range is for example a function of the resolution offered by the image sensors 5 and by the screen Touch 7. Indeed, if the virtual camera 11 is placed too high, the objects present on the virtual image 13 will be too small to be recognizable or even visible by the user. Such a virtual image 13 could therefore not help the driver to maneuver the vehicle 1.
- the third coordinate Zcam is for example between 1, 7 m and 3 m and can vary in steps of 5 cm.
- the third coordinate Zcam is associated with the angle of inclination ⁇ of the virtual camera 1 1, since the optical axis A 0 is defined as passing through the center C of the vehicle 1.
- a second graphic element 21, illustrated by FIG. 4b, is associated with this second posture parameter.
- This second graphic element 21 is here formed by a vertical segment.
- the second graphical element 21 is stored in the memory unit, since which it can be recovered by the computing unit 9 which is able to transmit it to the touch screen 7.
- the third posture parameter is the second radius R2 of the second circle C2 on which the virtual camera 1 1 is located.
- the second radius R2 represents a horizontal distance between the virtual camera 11 and the vehicle 1.
- the second radius R2 is within a predefined range of values and stored in the memory unit.
- the second radius R2 is for example between 3.5 m and 6 m, it can be varied in steps of 5 cm.
- a third graphic element 23 is associated with this third posture parameter.
- This third graphic element 23 is here formed by a horizontal segment. By pressing this third graphic element 23, the user will be able to vary the second radius R2.
- the third graphic element 23 is stored in the memory unit, from which it can be recovered by the computing unit 9 which is able to transmit it to the touch screen 7.
- the first, second and third posture parameters 19, 21, 23 allow the calculation unit 9 to calculate coordinates (Xcam, Ycam, Zcam, ⁇ , ⁇ ) of the virtual camera 11 in the vehicle coordinate system (0, Xv , Yv, Zv).
- the virtual camera 11 is able to be moved by the user around the representation of the vehicle 1, which causes a change in his
- the computing unit 9 is able to recover or generate a representation of the vehicle 1 as seen by the virtual camera 11.
- a three-dimensional model of the vehicle could be stored in a memory unit of the driver assistance system 3.
- a plurality of representations of the vehicle 1 is stored in a memory unit of the driver assistance system 3.
- the plurality of representations 15 is for example made from a simulation tool of the vehicle 1. For example 360 images are captured in perspective around the vehicle 1, in steps of 1 degree and remaining at a first value of the second radius R2 . Then, this operation is repeated by moving to a second value of the second radius R2.
- the second radius R2 is for example varied in millimeters.
- images of the vehicle 1 are captured by varying the third coordinate Zcam, and thus the angle of inclination ⁇ of the virtual camera 11.
- Each captured image is named with its orientation angle value ⁇ , its second radius value R2 and its angle of inclination ⁇ value, so the calculation unit 9 can easily retrieve the appropriate representation 15 for each posture of the virtual camera January 1.
- the calculation unit 9 is furthermore capable of superimposing the appropriate representation of the vehicle 1 on the reconstituted image of the environment 17 to form the virtual image 13 as seen by the virtual camera 11.
- the calculation unit 9 is also able to transmit the virtual image 13 to the touch screen 7.
- the touch screen 7 is placed in the passenger compartment of the vehicle 1, for example on the dashboard of the latter.
- the touch screen 7 is able to display the virtual image 13 as well as the graphic elements 19, 21, 23 transmitted by the calculation unit 9.
- the touch screen 7 is able to display images in colour.
- the touch screen 7 is furthermore able to acquire the position of a selected point P by the occasional support of a user on said touch screen 7.
- a screen mark (0e, Ye, Ze) is associated with the touch screen 7.
- the origin 0e of the screen mark (Oe, Ye, Ze) is for example located on a corner of the touch screen 7, for example the lower left corner.
- the axis (OeYe) extends along a horizontal axis of the touch screen 7.
- the axis (OeZe) extends along a vertical axis of the touch screen 7.
- the point P on which the user pressed has two-dimensional coordinates (yp, ZP). These two-dimensional coordinates (yp, ZP) are expressed in pixels.
- the touch screen 7 is able to determine the two-dimensional coordinates (yp, ZP) of the selected point P (for example by means of a computer associated with the touch screen 7) and to transmit them to the calculation unit 9.
- the operation a touch screen 7 is well known to those skilled in the art and will not be described in detail.
- FIG. 5 a method for displaying an image of the environment of a vehicle 1 on a touch screen 7 fitted to the vehicle 1 can be described in greater detail.
- This display method comprises a step a) of display on the touch screen 7 of one of the representations of the vehicle 1 and of at least one graphic element 19, 21, 23 among the first, second and third graphic elements 19, 21, 23 previously described.
- the vehicle 1 can be represented as seen from an initial position PosO of coordinates (Xcamo, Ycamo, Zcamo, ⁇ , ⁇ ) predefined of the initial virtual camera 1 1 in the vehicle reference (0, Xv, Yv , Zv).
- the touch screen 7 displays for example the first graphic element 19 associated with the first posture parameter (the orientation angle ⁇ ).
- the choice by the user of the posture of the virtual camera 1 1, and therefore the virtual view of the environment is facilitated. Indeed, the user quickly understands at which locations he can place the virtual camera 1 1. The choice of the location of the virtual camera 1 1 does not require great attention since a single point support with the finger is enough, and user can focus on the maneuvers he is doing.
- the touch screen 7 may further display instructions to the user. These instructions can explain to him how to select the location of the virtual camera 11. According to the invention, the selection is made by a one-off support of the user on the touch screen 7, more precisely on one of the graphical elements displayed. .
- the touch screen 7 acquires the position of a first point P1 of the first graphic element 19 selected by a user press exerted on the touch screen 7.
- the position of the first selected point P1 is expressed by a set of coordinates in pixels in the screen frame (0,
- the coordinates of the first selected point P1 are the coordinates of the first selected point P1
- the calculation unit 9 calculates the value of the first posture parameter, here the orientation angle ⁇ of the virtual camera 11 around the vehicle 1, as a function of the coordinates (yp). i, ZPI) of the first selected point P1 in order to deduce a new posture Pos1 of the virtual camera 1 1.
- the orientation angle ⁇ is determined by means of the formula:
- (Xc, Yc, Zc) are the coordinates of the center C of the vehicle 1 defined in the vehicle mark and that R1 is the radius of the first circle C1 (first graphical element 19).
- the coordinates (Xc, Yc, Zc), the value of R1 as well as other variables used in the calculations which will be described later, such as and R2, are stored in the memory unit.
- the coordinates (XVPI, yvpi, ZVPI) are defined as the three-dimensional coordinates of the first selected point P1 in the vehicle frame (0, Xv, Yv, Zv).
- the only unknown variable of the equation of the angle of orientation ⁇ is therefore the coordinate XVP-I.
- the coordinate is determined by the formula:
- Xcamo and Ycamo are the initial coordinates, and therefore known, of the camera in the vehicle reference.
- Zc is also known.
- R3 is determined using the formula:
- the coordinates (xpn, ypn,) are the coordinates of the first selected point P1 projected onto a first focal plane PF1 of the virtual camera 1 1, and expressed in the vehicle coordinate system (0, Xv, Yv, Zv) (the first focal plane PF1 being associated with the first posture of the virtual camera 1 1).
- the calculation unit 9 can use it to calculate the other coordinates of the first posture of the virtual camera 11
- a step d) the image sensors 5 acquire real images of the environment of the vehicle 1.
- the step d) can take place simultaneously with the preceding steps a), b), c), or before those -this.
- the actual images are transmitted to the computing unit 9 which composes a virtual image of the environment of the vehicle 1 as seen from the virtual camera 1 1 during a step e).
- the calculation unit 9 selects the real images to be used according to the posture of the virtual camera 1 1. For example, when the virtual camera is placed above the vehicle 1 with an optical axis parallel to the axis (OZv) vehicle mark the virtual image 13 displayed is that represented by FIG.
- the computing unit 9 can for example compose the virtual image 13 from the real images of only one, two or three image sensors 5.
- This step of composing the image does not form the heart of the invention and is already well known to those skilled in the art, it will not be described in more detail here.
- the display method further provides a step of reading a representation of the vehicle 1 as seen from the new posture Pos 1 of the virtual camera 11 deduced in step c).
- the representation 15 of the vehicle 1 is for example selected from the plurality of representations 15 stored in the memory unit.
- the virtual image 13 of the environment 17 and the representation 15 of the vehicle 1 are displayed on the touch screen 7 during a step f).
- the user can easily choose the location of the virtual camera 11, and accesses easily understandable information enabling him to better visualize the environment 17 of the vehicle 1.
- the calculation unit 9 controls the display of the second graphic element 21 in order to allow the user to adjust the height of the virtual camera 11, before displaying the third graphic element 23 in order to adjust the horizontal distance between the virtual camera 11 and the vehicle 1.
- the calculation unit 9 displays the second graphic element 21 associated with the second posture parameter, that is to say the vertical segment during step a).
- the vertical segment is for example displayed along a vertical axis passing through the center C.
- step b) the computing unit 9 acquires the position of a second selected point P2 by pressing on the touch screen at the vertical segment.
- This second selected point P2 makes it possible to calculate, during step c), on the one hand, a second value of the coordinate Zc a m 2 of the second posture Pos 2 of the virtual camera 1 1 and, on the other hand, a second value of its angle inclination 92.
- Coordinates are the coordinates of the second selected point P2 projected on a second focal plane PF2 of the virtual camera 11, and expressed in the vehicle coordinate system (the second focal plane
- yp2 and ZP2 are the coordinates of the second selected point P2 and therefore known.
- the orientation angle ⁇ around the vehicle 11 has the first value ⁇ 1 previously calculated.
- the virtual camera 11 is therefore positioned in
- Steps d) and e) are then repeated so as to display an image of the environment 17 and a representation of the vehicle 1, seen from an altitude corresponding to the wish of the user.
- calculation unit 9 controls the display of the third graphic element 23 on the touch screen 7 during a new step a).
- the horizontal segment of the third graphic element 23 is for example arranged along an axis passing through the center C and at an angle equal to the orientation angle ⁇ + ⁇ / 2 with respect to the longitudinal axis (OXv) .
- step b) the computing unit 9 acquires the position of the third selected point P3 by pressing the horizontal segment. Coordinates of the third selected point P3 on the touch screen 7 are transmitted to the calculation unit 9.
- This third selected point P3 makes it possible to calculate, during step c), a new horizontal distance 2f to be applied between the vehicle 1 and the virtual camera 1 1.
- the calculation unit 9 uses the values of orientation angle ⁇ 1 , and angle of inclination previously calculated.
- X vp is the coordinate of the point of intersection of the circle of radius R1 and the horizontal segment of the third graphic element 23.
- the coordinate x vp is obtained with the formula:
- Xcamo and Ycamo are the initial coordinates of the virtual camera 1 1 in the vehicle mark.
- Zc is also known.
- R3 is determined using the formula:
- the coordinates are the coordinates of the third selected point P3 projected on a third focal plane PF3 of the virtual camera 11, and expressed in the vehicle reference (0, Xv, Yv, Zv) (the third focal plane PF3 being associated with the third posture of the virtual camera 11).
- the Thales theorem is applied as follows:
- the orientation angle ⁇ around the vehicle 11 has the first value ⁇ ⁇ previously calculated
- the inclination angle ⁇ has the value 92 previously calculated.
- the coordinates of the virtual camera 1 1 are:
- Steps d) and e) are then repeated so as to display an image of the environment 17 and a representation of the vehicle 1, seen from a distance to the vehicle 1 corresponding to the wish of the user.
- the virtual image 13 comprising the environment 17 of the vehicle 1 as well as the representation 15 of the vehicle 1 is thus composed and displayed after each deduction of the new posture of the virtual camera 11.
- the virtual image 13 could be composed and displayed only after the driver has selected the three selected points P1, P2, P3.
- the virtual image 13 could be composed and displayed only after the driver has selected the three selected points P1, P2, P3.
- the calculation unit 9 could control the simultaneous display of the first graphic element 19, second graphic element 21 and third graphic element 23 previously described, during the display step a).
- the user could then choose with precision the new posture of the virtual camera 11 by choosing a point on any one of the first, second and third graphic elements 19, 21, 23.
- the acquisition of the first, second and third selected points P1, P2, P3 could be performed in any order. Therefore, the user could select in any order the height of the virtual camera 1 1, the rotation angle ⁇ of the virtual camera 11 and the horizontal distance R2 between the virtual camera 11 and the vehicle 1.
- the first graphic element 19, the second graphic element 21, and the third graphic element 23 each have a different color.
- the computing unit 9 can implement an automatic method for determining the posture of the virtual camera 11, when the conditions require it.
- the computing unit 9 is programmed to automatically modify the posture of the virtual camera 11 when an obstacle is detected near the vehicle 1.
- the calculation unit 9 determines the trajectory of the vehicle 1 during a prediction step.
- the calculation unit 9 determines this trajectory as a function of at least the angle of the steering wheel and the gear ratio engaged (forward or reverse).
- the computing unit 9 searches for an obstacle located near the vehicle 1. More precisely, the computing unit 9 determines if an obstacle is on the previously predicted trajectory.
- the computing unit 9 uses remote sensors (radar and / or sonar and / or lidar) located around the vehicle 1.
- the vehicle 1 is equipped with six remote sensors at the front and six remote sensors at the rear. Remote sensors not only detect the presence of an obstacle, but also assess the distance between the obstacle and the vehicle 1.
- the virtual camera is automatically positioned along the longitudinal axis (OXv) of the vehicle 1 towards the rear of the vehicle 1 , and directed towards the front of the vehicle 1.
- the virtual camera 1 1 is then placed at an orientation angle ⁇ equal to ⁇ radians.
- the computing unit 9 could choose the orientation angle ⁇ from a plurality of rotation angles ⁇ stored in the memory unit.
- the computing unit 9 determines the height of the obstacle.
- the angle of inclination ⁇ of the virtual camera 11 is then calculated so that the optical axis of the virtual camera 11 passes through the center of the obstacle.
- the horizontal distance R2 between the vehicle 1 and the virtual camera 1 1 is calculated as a function of the distance detected between the obstacle and the vehicle 1.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Computing Systems (AREA)
- Geometry (AREA)
- Computer Graphics (AREA)
- General Physics & Mathematics (AREA)
- Closed-Circuit Television Systems (AREA)
- Instrument Panels (AREA)
Abstract
Description
Claims
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
FR1760503A FR3073310B1 (en) | 2017-11-08 | 2017-11-08 | PROCESS FOR DISPLAYING AN IMAGE OF THE ENVIRONMENT OF A VEHICLE ON A TOUCH SCREEN EQUIPPING THE VEHICLE |
PCT/EP2018/080190 WO2019091920A1 (en) | 2017-11-08 | 2018-11-05 | Method for displaying an image of the area surrounding a vehicle on a touch screen installed in the vehicle |
Publications (1)
Publication Number | Publication Date |
---|---|
EP3707679A1 true EP3707679A1 (en) | 2020-09-16 |
Family
ID=61027915
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP18795667.7A Pending EP3707679A1 (en) | 2017-11-08 | 2018-11-05 | Method for displaying an image of the area surrounding a vehicle on a touch screen installed in the vehicle |
Country Status (3)
Country | Link |
---|---|
EP (1) | EP3707679A1 (en) |
FR (1) | FR3073310B1 (en) |
WO (1) | WO2019091920A1 (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
FR3098620B1 (en) * | 2019-07-12 | 2021-06-11 | Psa Automobiles Sa | Method of generating a visual representation of the driving environment of a vehicle |
CN114419949B (en) * | 2022-01-13 | 2022-12-06 | 武汉未来幻影科技有限公司 | Automobile rearview mirror image reconstruction method and rearview mirror |
-
2017
- 2017-11-08 FR FR1760503A patent/FR3073310B1/en active Active
-
2018
- 2018-11-05 EP EP18795667.7A patent/EP3707679A1/en active Pending
- 2018-11-05 WO PCT/EP2018/080190 patent/WO2019091920A1/en unknown
Also Published As
Publication number | Publication date |
---|---|
WO2019091920A1 (en) | 2019-05-16 |
FR3073310B1 (en) | 2021-02-26 |
FR3073310A1 (en) | 2019-05-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2133237B1 (en) | Process of displaying a parking aid | |
FR2871737A1 (en) | METHOD FOR ATTACHING A TRAILER TO A MOTOR VEHICLE | |
EP3601005B1 (en) | Method and device for assisting a driver of a self-driving vehicle, in particular when a manual control is resumed | |
FR3098172A1 (en) | Method of determining an avoidance trajectory of a motor vehicle | |
EP3707679A1 (en) | Method for displaying an image of the area surrounding a vehicle on a touch screen installed in the vehicle | |
EP2094531B1 (en) | Electronic rear-view device | |
EP3245117B1 (en) | Method and device to assist with the reversal manoeuvre of a motor vehicle | |
EP2750961B1 (en) | Device for estimating a future path of a vehicle and associating with parts that it comprises aspects that differ according to their positions in relation to an obstacle, for a drive-assist system | |
WO2010037795A1 (en) | Method of control and associated driving aid system | |
FR3016573B1 (en) | VEHICLE MIRROR INSTALLATION | |
EP2801075B1 (en) | Image processing method for an on-board camera installed on a vehicle and corresponding processing device | |
FR2976244A1 (en) | METHOD FOR ASSISTING THE PARKING MANEUVER OF A VEHICLE AND SYSTEM FOR IMPLEMENTING IT | |
EP3682320B1 (en) | Method for displaying the surroundings of a motor vehicle on a screen, and motor vehicle in which such a method is implemented | |
EP3377386B1 (en) | Driver assistance method for a motor vehicle | |
EP2648963B1 (en) | Driving aid device allowing the driver of a vehicle to check a parking area he has just left | |
EP3418121B1 (en) | Device for displaying a motion of a vehicle door and associated method | |
WO2024121172A1 (en) | Method for detecting a centre line of a traffic lane and for controlling a motor vehicle | |
EP4430569A1 (en) | Method for the dynamic estimation of the pitch of a motor vehicle by means of an image capture sensor | |
FR3115008A1 (en) | Method for assisting the driving in reverse of a motor vehicle to the rear of which a trailer is hitched | |
FR2956364A1 (en) | Driver assistance device for handling e.g. car around building site, has processing unit generating image data defining image to be displayed in vehicle, where image data represents part of vehicle based on relative position of head | |
FR3132487A1 (en) | Method for detecting a center line of a traffic lane | |
FR3093968A1 (en) | Vehicle rear view system comprising a plurality of cameras and a computer | |
FR3070085A1 (en) | METHOD FOR CALIBRATING A CAMERA OF A MOTOR VEHICLE | |
FR2929196A1 (en) | Maneuver assistant for detecting and preventing collision to park e.g. automobile, has monitor or screen for displaying augmented reality image formed by superposition of image of ego-vehicle on selected anterior image |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: UNKNOWN |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE |
|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE |
|
17P | Request for examination filed |
Effective date: 20200428 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
AX | Request for extension of the european patent |
Extension state: BA ME |
|
DAV | Request for validation of the european patent (deleted) | ||
DAX | Request for extension of the european patent (deleted) | ||
RAP3 | Party data changed (applicant data changed or rights of an application transferred) |
Owner name: RENAULT S.A.S |
|
RAP3 | Party data changed (applicant data changed or rights of an application transferred) |
Owner name: RENAULT S.A.S |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: EXAMINATION IS IN PROGRESS |
|
17Q | First examination report despatched |
Effective date: 20220929 |
|
P01 | Opt-out of the competence of the unified patent court (upc) registered |
Effective date: 20230608 |
|
RAP1 | Party data changed (applicant data changed or rights of an application transferred) |
Owner name: AMPERE SAS |