GB2580400A - A control system, system and method for providing assistance to an occupant of a vehicle - Google Patents

A control system, system and method for providing assistance to an occupant of a vehicle Download PDF

Info

Publication number
GB2580400A
GB2580400A GB1900332.6A GB201900332A GB2580400A GB 2580400 A GB2580400 A GB 2580400A GB 201900332 A GB201900332 A GB 201900332A GB 2580400 A GB2580400 A GB 2580400A
Authority
GB
United Kingdom
Prior art keywords
vehicle
image
trajectory
composite image
indicator
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
GB1900332.6A
Other versions
GB2580400B (en
GB201900332D0 (en
Inventor
Aitidis Ilias
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jaguar Land Rover Ltd
Original Assignee
Jaguar Land Rover Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jaguar Land Rover Ltd filed Critical Jaguar Land Rover Ltd
Priority to GB1900332.6A priority Critical patent/GB2580400B/en
Publication of GB201900332D0 publication Critical patent/GB201900332D0/en
Priority to US17/422,163 priority patent/US11919513B2/en
Priority to DE112020000391.4T priority patent/DE112020000391T5/en
Priority to PCT/EP2020/050119 priority patent/WO2020144129A1/en
Publication of GB2580400A publication Critical patent/GB2580400A/en
Application granted granted Critical
Publication of GB2580400B publication Critical patent/GB2580400B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/26Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
    • G01C21/34Route searching; Route guidance
    • G01C21/36Input/output arrangements for on-board computers
    • G01C21/3626Details of the output of route guidance instructions
    • G01C21/3635Guidance using 3D or perspective road maps
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60KARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
    • B60K35/00Arrangement of adaptations of instruments
    • B60K35/22
    • B60K35/28
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/26Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
    • G01C21/34Route searching; Route guidance
    • G01C21/36Input/output arrangements for on-board computers
    • G01C21/3626Details of the output of route guidance instructions
    • G01C21/3644Landmark guidance, e.g. using POIs or conspicuous other objects
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/26Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
    • G01C21/34Route searching; Route guidance
    • G01C21/36Input/output arrangements for on-board computers
    • G01C21/3626Details of the output of route guidance instructions
    • G01C21/3658Lane guidance
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/26Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
    • G01C21/34Route searching; Route guidance
    • G01C21/36Input/output arrangements for on-board computers
    • G01C21/3667Display of a road map
    • G01C21/367Details, e.g. road map scale, orientation, zooming, illumination, level of detail, scrolling of road map or positioning of current position marker
    • B60K2360/00
    • B60K2360/166
    • B60K2360/176
    • B60K2360/177
    • B60K2360/21
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W50/00Details of control systems for road vehicle drive control not related to the control of a particular sub-unit, e.g. process diagnostic or vehicle driver interfaces
    • B60W50/08Interaction between the driver and the control system
    • B60W50/14Means for informing the driver, warning the driver or prompting a driver intervention
    • B60W2050/146Display means
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2420/00Indexing codes relating to the type of sensors based on the principle of their operation
    • B60W2420/40Photo or light sensitive means, e.g. infrared sensors
    • B60W2420/403Image sensing, e.g. optical camera
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W2552/00Input parameters relating to infrastructure
    • B60W2552/53Road markings, e.g. lane marker or crosswalk
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W30/00Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units, or advanced driver assistance systems for ensuring comfort, stability and safety or drive control systems for propelling or retarding the vehicle
    • B60W30/08Active safety systems predicting or avoiding probable or impending collision or attempting to minimise its consequences
    • B60W30/095Predicting travel path or likelihood of collision
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60WCONJOINT CONTROL OF VEHICLE SUB-UNITS OF DIFFERENT TYPE OR DIFFERENT FUNCTION; CONTROL SYSTEMS SPECIALLY ADAPTED FOR HYBRID VEHICLES; ROAD VEHICLE DRIVE CONTROL SYSTEMS FOR PURPOSES NOT RELATED TO THE CONTROL OF A PARTICULAR SUB-UNIT
    • B60W30/00Purposes of road vehicle drive control systems not related to the control of a particular sub-unit, e.g. of systems using conjoint control of vehicle sub-units, or advanced driver assistance systems for ensuring comfort, stability and safety or drive control systems for propelling or retarding the vehicle
    • B60W30/10Path keeping
    • B60W30/12Lane keeping
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/26Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
    • G01C21/34Route searching; Route guidance
    • G01C21/36Input/output arrangements for on-board computers
    • G01C21/3626Details of the output of route guidance instructions
    • G01C21/365Guidance using head up displays or projectors, e.g. virtual vehicles or arrows projected on the windscreen or on the road itself

Abstract

A control system, system and method are provided for assisting a vehicle occupant. The method comprises determining a composite image sequence comprising a sequence of image frames 32, 34, each comprising a captured image of an environment external to the vehicle and a trajectory indicator 39a-e indicative of the determined vehicle trajectory through that environment. The trajectory indicator 39a-e is positioned within the composite image sequence in dependence on movement by the vehicle 50 between the time at which images associated with the corresponding image frames 32, 34 are captured such that the trajectory indicator 39a-e appears substantially stationary in a direction of travel (D) of the vehicle 50 with respect to one or more image objects 36, 38 within the composite image sequence. The trajectory indicator may be positioned or oriented in accordance with sensed topography. The invention avoids an unnatural sense of trajectory lines not moving with an environment.

Description

A CONTROL SYSTEM, SYSTEM AND METHOD FOR PROVIDING ASSISTANCE TO AN OCCUPANT OF A VEHICLE
TECHNICAL FIELD
The present disclosure relates to a control system, a system and a method for providing assistance to an occupant of a vehicle. Aspects of the invention relate to a control system, to a system, to a vehicle, to a method, to a non-transitory computer readable medium, and to computer software for providing assistance whilst performing manoeuvres within a vehicle.
BACKGROUND
It is known to provide driver assistance systems which provide a visual representation of an environment external to a vehicle. Some of these systems include an image sequence of the environment captured from one or more cameras mounted on or within the vehicle. In addition, some systems include some form of indication of a predicted path of the vehicle through the image sequence. This generally takes the form of one or more trajectory lines.
Conventionally, trajectory lines may illustrate the predicted path of one or more wheels of the vehicle within the environment.
A disadvantage of prior art systems is that the trajectory lines themselves appear fixed with respect to the vehicle in the direction of travel of the vehicle. This can appear unnatural and/or confusing to a user as the trajectory lines may appear to "float" over a surface as the vehicle moves within the environment. This can make it difficult to assess any correspondence between the location of the trajectory lines within the image sequence and any objects within the environment external to the vehicle present in the image sequence.
It is an aim of the present invention to address one or more of the disadvantages associated with the prior art.
SUMMARY OF THE INVENTION
Aspects and embodiments of the invention provide a control system, a system, a vehicle, a method, a non-transitory computer readable medium and computer software as claimed in the appended claims.
According to an aspect of the present invention there is provided a control system for providing assistance to an occupant of a vehicle, the control system comprising one or more controllers, configured to: determine a vehicle trajectory from one or more vehicle parameters; and determine a composite image sequence comprising a sequence of image frames, each image frame comprising a captured image and a trajectory indicator indicative of the determined vehicle trajectory; wherein the control system is configured to position the trajectory indicator within the composite image sequence in dependence on a distance travelled by the vehicle such that the trajectory indicator appears substantially stationary in a direction of travel of the vehicle with respect to one or more image objects within the composite image sequence, the one or more image objects being external to the vehicle.
According to an aspect of the present invention there is provided a control system for providing assistance to an occupant of a vehicle, the control system comprising one or more controllers, configured to: receive one or more vehicle parameters; receive image data from an imaging device comprising captured images of an environment external to the vehicle; determine a vehicle trajectory in dependence on the or each parameter; determine a composite image sequence comprising a sequence of image frames, each image frame comprising a captured image and a trajectory indicator indicative of the determined vehicle trajectory; and output a signal indicative of the composite image sequence to a display for displaying the composite image sequence to the occupant of the vehicle; wherein the control system is configured to move the trajectory indicator from a first position in a first image frame of the composite image sequence, to a second position in a second image frame of the composite image sequence in dependence on a distance travelled by the vehicle between the time at which an image associated with the first image frame is captured and the time at which an image associated with the second image frame is captured such that the trajectory indicator appears substantially stationary in a direction of travel of the vehicle with respect to one or more image objects within the composite image sequence, the one or more image objects being external to the vehicle.
Advantageously, the vehicle may be made to appear to a user of the control system viewing the composite image to travel over a stationary trajectory indicator. This provides the user with a more natural and less confusing representation of the vehicle's movement through the environment. The control system of the present invention provides a visual representation of an environment of a vehicle which enables a user to better assess any correspondence between the location of a trajectory indicator, and hence a future position of a vehicle, with respect to one or more objects within the environment external to the vehicle.
In embodiments, the one or more controllers collectively comprise: at least one electronic processor having an electrical input for receiving the one or more vehicle parameters and/or the image data; and at least one electronic memory device operatively coupled to the at least one electronic processor and having instructions stored therein; wherein the at least one electronic processor is configured to access the at least one memory device and execute the instructions stored therein so as to determine the composite image sequence. Optionally, the at least one electronic memory device is electrically coupled to the at least one electronic processor.
In embodiments, the first position and the second position are separated within the composite image sequence by a distance directly proportional to the distance travelled by the vehicle between the time at which the image associated with the first image frame is captured and the time at which the image associated with the second image frame is captured.
The trajectory indicator may comprise a plurality of indicator sections. In some embodiments the control system is configured to move each of the indicator sections from respective first positions within the first image frame of the composite image sequence, to respective second positions within the second image frame of the composite image sequence in dependence on a distance travelled by the vehicle between the time at which an image associated with the first image frame is captured and the time at which an image associated with the second image frame is captured such each indicator section appears substantially stationary in the direction of travel of the vehicle within the composite image sequence with respect to the one or more image objects.
Advantageously using a plurality of trajectory sections as opposed to a solid trajectory line provides a better representation of movement of the trajectory indicator with respect to the vehicle.
In some embodiments the trajectory indicator comprises one or more lines indicative of a predicted path to be traversed by one or more wheels of the vehicle. In such embodiments, the trajectory indicator may comprise a first trajectory line representative of a predicted path to be traversed by a first wheel of the vehicle and a second trajectory line representative of a predicted path to be traversed by a second wheel of the vehicle. Advantageously, an occupant of the vehicle may be provided with information relating to the predicted path of one or more wheels of the vehicle through the environment represented by the composite image sequence. This may assist the occupant by providing information relating to the terrain or other image objects which may be encountered by the wheels of the vehicle were the vehicle to continue travelling along the predicted path.
In some embodiments the one or more image objects comprise a traversable surface within the composite image sequence. In such embodiments, the control system may be configured to move the trajectory indicator from the first position to the second position such that, to the user, the vehicle appears to move over the trajectory indicator as the vehicle traverses the surface along the determined vehicle trajectory. Advantageously, the control system may provide a trajectory indicator which is fixed in position with respect to the surface over which the vehicle is travelling. In this way, the occupant may be able to better assess the path of the vehicle over the surface.
In some embodiments the control system may be configured to analyse the received image data to determine a topography of the surface. In such embodiments, the control system may be configured to determine a position and/or orientation of the trajectory indicator within one or more image frames of the composite image sequence in dependence on the determined topography. In some embodiments the control system may be configured to: receive sensor data from one or more sensors indicative of a topography of the surface; and determine the composite image sequence in dependence on the received sensor data.
The control system may comprise an electrical input for receiving the sensor data from the one or more sensors.
In embodiments, the control system may be configured to process the received sensor data to determine a surface profile of the surface; and determine a position and/or orientation of the trajectory indicator within one or more image frames of the composite image sequence in dependence on the determined surface profile such that the trajectory indicator appears to lie on the surface within the composite image sequence.
In embodiments, the control system may be configured to: process the received image data to determine a speed and/or direction of motion of the one or more image objects within the composite image sequence; and move the trajectory indicator from the first position to the second position in dependence on the determined speed and/or direction of motion of the one or more image objects within the composite image sequence. By moving the trajectory indicator in dependence on the speed and/or direction of motion of the one or more image objects, the trajectory indicator may be made to appear to be in a fixed position with respect to the one or more image objects.
The one or more vehicle parameters may comprise a steering angle of the vehicle. The steering angle may comprise an angle of a steering wheel of the vehicle or the angle of one or more steerable wheels of the vehicle.
A control system as claimed in any preceding claim, wherein the one or more vehicle parameters comprise a velocity of the vehicle. In such embodiments, the control system may be configured to move the trajectory indicator within the composite image sequence at a speed and in a direction proportional to the velocity of the vehicle.
In some embodiments the control system may be configured to receive image data from a plurality of imaging devices. In such embodiments, the control system may be configured to process the image data received from the plurality of imaging devices to determine the composite image sequence. The control system may be configured to process the received image data by applying one or more image transformations and/or stitching image data from two or more different imaging devices together to form the composite image sequence, as will be known in the art. The first and second image frames of the composite image sequence may each comprise image data from two or more separate imaging devices.
According to a further aspect of the invention there is provided a system comprising the control system of any preceding aspect of the invention and an imaging device configured to capture one or more images of an environment external to the vehicle.
In embodiments, the system may comprise a display. The display may be a vehicle display configured to receive a signal indicative of the composite image sequence for displaying the composite image sequence to an occupant of the vehicle.
According to another aspect of the present invention there is provided a vehicle comprising a control system or a system according to any preceding aspect of the invention.
According to an aspect of the invention there is provided a method for providing assistance to an occupant of a vehicle, the method comprising: determining a vehicle trajectory from one or more vehicle parameters; and determining a composite image sequence comprising a sequence of image frames, each image frame comprising a captured image and a trajectory indicator indicative of the determined vehicle trajectory; wherein the method comprises positioning the trajectory indicator within the composite image sequence in dependence on a distance travelled by the vehicle such that the trajectory indicator appears substantially stationary in a direction of travel of the vehicle with respect to one or more image objects within the composite image sequence, the one or more image objects being external to the vehicle.
According to an aspect of the invention there is provided a method for providing assistance to an occupant of a vehicle, the method comprising: receiving one or more vehicle parameters; receiving image data comprising captured images of an environment external to the vehicle; determining a vehicle trajectory in dependence on the or each parameter; determining a composite image sequence comprising a sequence of image frames, each image frame comprising a captured image and a trajectory indicator indicative of the determined vehicle trajectory; and outputting a signal indicative of the composite image sequence to a display for displaying the composite image sequence to the occupant of the vehicle; wherein the method comprises moving the trajectory indicator from a first position in a first image frame of the composite image sequence, to a second position in a second image frame of the composite image sequence in dependence on a distance travelled by the vehicle between the time at which an image associated with the first image frame is captured and a time at which an image associated with the second image frame is captured such that to the trajectory indicator appears substantially stationary in a direction of travel of the vehicle with respect to one or more image objects within the composite image sequence, the one or more image objects being external to the vehicle.
In embodiments, the first position and the second position are separated within the composite image sequence by a distance directly proportional to the distance travelled by the vehicle between the time at which the image associated with the first image frame is captured and the time at which the image associated with the second image frame is captured.
Advantageously, the vehicle may be made to appear to a user of the control system viewing the composite image to travel over a stationary trajectory indicator. This provides the user with a more natural and less confusing representation of the vehicle's movement through the environment when compared with prior art systems. The method of the present invention provides a visual representation of an environment of a vehicle which enables a user to better assess any correspondence between the location of a trajectory indicator, and hence a future position of a vehicle, with respect to one or more objects within the environment external to the vehicle.
The trajectory indicator may comprise a plurality of indicator sections. In some embodiments the method may comprise moving each of the indicator sections from respective first positions within the first image frame of the composite image sequence, to respective second positions within the second image frame of the composite image sequence in dependence on a distance travelled by the vehicle between the time at which an image associated with the first image frame is captured and the time at which an image associated with the second image frame is captured such each indicator section appears substantially stationary in the direction of travel of the vehicle within the composite image sequence with respect to the one or more image objects.
In some embodiments the trajectory indicator comprises one or more lines indicative of a predicted path to be traversed by one or more wheels of the vehicle. In such embodiments, the trajectory indicator may comprise a first trajectory line representative of a predicted path to be traversed by a first wheel of the vehicle and a second trajectory line representative of a predicted path to be traversed by a second wheel of the vehicle. Advantageously, an occupant of the vehicle may be provided with information relating to the predicted path of one or more wheels of the vehicle through the environment represented by the composite image sequence. This may assist the occupant by providing information relating to the terrain or other image objects which may be encountered by the wheels of the vehicle were the vehicle to continue travelling along the predicted path.
In some embodiments the one or more image objects comprise a traversable surface within the composite image sequence. In such embodiments, the method may comprise moving the trajectory indicator from the first position to the second position such that, to the user, the vehicle appears to move over the trajectory indicator as the vehicle traverses the surface along the determined vehicle trajectory. Advantageously, the method may provide a trajectory indicator which is fixed in position with respect to the surface over which the vehicle is travelling. In this way, the occupant may be able to better assess the path of the vehicle over the surface.
In some embodiments the method may comprise analysing the received image data to determine a topography of the surface. The method may comprise determining a position and/or orientation of the trajectory indicator within one or more image frames of the composite image sequence in dependence on the determined topography. In some embodiments the method may comprise receiving sensor data from one or more sensors indicative of a topography of the surface; and determining the composite image sequence in dependence on the received sensor data. In some embodiments the method comprises processing the received sensor data to determine a surface profile of the surface; and determining a position and/or orientation of the trajectory indicator within one or more image frames of the composite image sequence in dependence on the determined surface profile such that the trajectory indicator appears to lie on the surface within the composite image sequence.
In embodiments, the method may comprise processing the received image data to determine a speed and/or direction of motion of the one or more image objects within the composite image sequence; and moving the trajectory indicator from the first position to the second position in dependence on the determined speed and/or direction of motion of the one or more image objects within the composite image sequence. By moving the trajectory indicator in dependence on the speed and/or direction of motion of the one or more image objects, the trajectory indicator may be made to appear to be in a fixed position with respect to the one or more image objects.
The one or more vehicle parameters may comprise a steering angle of the vehicle. The steering angle may comprise an angle of a steering wheel of the vehicle or the angle of one or more steerable wheels of the vehicle.
In some embodiments the method may comprise receiving image data from a plurality of imaging devices. In such embodiments, the method may comprise processing the image data received from the plurality of imaging devices to determine the composite image sequence. The processing may comprise processing the received image data by applying one or more image transformations and/or stitching image data from two or more different imaging devices together to form the composite image sequence, as will be known in the art. The first and second image frames of the composite image sequence may each comprise image data from two or more separate imaging devices.
According to an aspect of the invention there is provided a computer readable medium having instructions stored therein which, when executed by one or more electronic processors, cause the one or more electronic processors to perform the method of a preceding aspect of the invention. Optionally, the computer readable medium comprises a non-transitory computer readable medium.
According to a further aspect of the invention there is provided computer software which, when executed by one or more processors, causes performance of the method of a preceding aspect of the invention.
Any controller or controllers described herein may suitably comprise a control unit or computational device having one or more electronic processors. Thus the system may comprise a single control unit or electronic controller or alternatively different functions of the controller may be embodied in, or hosted in, different control units or controllers. As used herein the term "controller" or "control unit" will be understood to include both a single control unit or controller and a plurality of control units or controllers collectively operating to provide any stated control functionality. To configure a controller, a suitable set of instructions may be provided which, when executed, cause said control unit or computational device to implement the control techniques specified herein. The set of instructions may suitably be embedded in said one or more electronic processors. Alternatively, the set of instructions may be provided as software saved on one or more memory associated with said controller to be executed on said computational device. A first controller may be implemented in software run on one or more processors. One or more other controllers may be implemented in software run on one or more processors, optionally the same one or more processors as the first controller. Other suitable arrangements may also be used.
Within the scope of this application it is expressly intended that the various aspects, embodiments, examples and alternatives set out in the preceding paragraphs, in the claims and/or in the following description and drawings, and in particular the individual features thereof, may be taken independently or in any combination. That is, all embodiments and/or features of any embodiment can be combined in any way and/or combination, unless such features are incompatible. The applicant reserves the right to change any originally filed claim or file any new claim accordingly, including the right to amend any originally filed claim to depend from and/or incorporate any feature of any other claim although not originally claimed in that manner.
BRIEF DESCRIPTION OF THE DRAWINGS
One or more embodiments of the invention will now be described, by way of example only, with reference to the accompanying drawings, in which: Figure 1 shows a schematic representation of an embodiment of a control system in accordance with the present invention; Figure 2 shows a schematic representation of an embodiment of a vehicle in accordance with the present invention; Figure 3 is a flowchart illustrating an embodiment of a method in accordance with the present invention; Figure 4 illustrates an example method for use in embodiments of the present invention; Figure 5 illustrates an example method for use in embodiments of the present invention; Figure 6 illustrates different types of trajectory indicator which may be used in embodiments of the present invention; Figures 7A & 73 illustrate an example method for use in embodiments of the present invention; Figure 7C illustrates the operational use of embodiments of the invention, showing a graphical representation of a composite image sequence formed in embodiments of the present invention; Figure 8A is a graphical representation illustrating a composite image sequence
generated using prior art systems and methods;
Figure 83 illustrates the operational use of embodiments of the invention, showing a graphical representation of a composite image sequence formed in embodiments of the present invention; and Figure 9 illustrates the operational use of embodiments of the invention, showing a graphical representation of a composite image sequence formed in embodiments of the present invention.
DETAILED DESCRIPTION
A control system 10, system 30, vehicle 50 and method 100 in accordance with embodiments of the present invention are described herein with reference to the accompanying Figures.
With reference to Figure 1, a system 30 in accordance with the invention comprises a control system 10. The control system 10 is operatively coupled to a vehicle system 16, an imaging device in the form of a camera 20, one or more sensors 28 and a display 24 as shown in Figure 1. The control system 10 includes a processor 12, memory device 26, electrical inputs 14, 18, 27, and an electrical output 22.
The camera 20 is configured to capture images of an environment external to the vehicle. As will be described herein, the control system 10 is configured to receive image data representative of the images captured by the camera 20, and use this data to determine a composite image sequence.
The vehicle system 16 may be any system capable of outputting a signal indicative of one or more vehicle parameters, specifically relating to the motion of the vehicle. The vehicle system 16 may comprise a steering system of the vehicle which may be capable of outputting a signal indicative of a steering angle of the vehicle. The steering angle may be an angular position of a steering wheel of the vehicle. Additionally or alternatively, the steering angle may comprise an angular position of one or more steerable wheels of the vehicle. The steering angle may relate to a predicted radius of curvature of a vehicle path due to the angular position of one or more steerable wheels of the vehicle. The vehicle system 16 may comprise a braking system of the vehicle, such as an anti-lock braking system (ABS) which may be configured to output a signal indicative of a wheel speed of the vehicle, and hence a speed of the vehicle. The vehicle system 16 may comprise a power unit management system which may be configured to output a signal indicative of an engine and/or motor speed of the vehicle, for example. In use, the control system 10 may be configured to determine the vehicle trajectory in dependence on any one or more such types of vehicle parameter.
In embodiments, the vehicle system 16 is an imaging system, and may comprise the camera 20, for example. In such embodiments, the vehicle system 16 comprises an image processing unit configured to analyse movement of one or more objects within image data captured by the imaging system. Such analysis may be used to infer a speed of the vehicle relative to those objects, and hence a speed of the vehicle within the environment.
The one or more sensors 28 may comprise sensors capable of detecting obstacles, surfaces and other such objects within the environment of the vehicle, and/or information relating to the position / orientation of the vehicle within that environment. For example, the one or more sensors may comprise an imaging device such as a camera, RADAR, LIDAR, ultrasonic sensors, etc. The one or more sensors may comprise the camera 20 configured to capture images of the environment external to the vehicle. The data received from the one or more sensors 28 may be used to map the environment external to the vehicle. For example, in embodiments the control system 10 is configured to utilise data received from the sensor(s) to determine a topography of a surface over which the vehicle is currently travelling or will travel were it to continue along the determined trajectory. This topography may be used to determine the composite image sequence as described herein. The sensor(s) 28 may additionally include an inertial measurement unit (IMU) for determining an orientation of the vehicle along one or more axes, and/or sensors relating to a GNSS module (e.g. a GPS module) within the vehicle suitable for determining a position of the vehicle within a mapped environment.
In use, the control system 10 receives, at electrical input 14, one or more vehicle parameters from the vehicle system 16. The one or more vehicle parameters may comprise a speed of the vehicle, velocity of the vehicle, steering angle, and/or orientation of the vehicle (e.g. roll, pitch and/or yaw angle), for example. The control system 10 is further configured to receive image data, at electrical input 18, from the camera 20. The image data comprises captured images of an environment external to the vehicle. The processor 12 determines a vehicle trajectory using the received vehicle parameters, and uses the determined vehicle trajectory along with the received image data to determine a composite image sequence. Specifically, the processor 12 is configured to form a composite image sequence comprising a sequence of image frames, each comprising a captured image and a trajectory indicator indicative of the determined vehicle trajectory. A control signal indicative of the composite image sequence is output via electrical output 22 to the display 24 for displaying the composite image sequence to an occupant of the vehicle.
As will be described in further detail herein, in generating the composite image sequence the control system 10 is configured to move the trajectory indicator within the composite image sequence such that the trajectory indicator appears substantially stationary in a direction of travel of the vehicle. Specifically, the control system 10 is configured to move the trajectory indicator within the composite image sequence such that the trajectory indicator appears substantially stationary with respect to one or more image objects within the composite image sequence, the image object(s) being external to the vehicle. In some embodiments the control system 10 is configured to receive, at electrical input 27, sensor data from the sensor(s) 28. The sensor data may be indicative of a topography of a surface over which the vehicle may travel.
In embodiments, in generating the composite image sequence the control system 10 is configured to determine a position and/or orientation of the trajectory indicator image frames of the composite image sequence in dependence on a topography of a surface over which the vehicle is predicted to travel. Sensor data from the sensor(s) 28 is used to determine a topography of a traversable surface within the environment of the vehicle. This is achieved by mapping, e.g. using a point cloud map, contour map, sparse map, etc. an environment using one or more sensor(s). The mapping of the environment is used to determine a three dimensional representation of the environment, and specifically a traversable surface within the environment of the vehicle. Overlaying or otherwise aligning this three dimensional representation with the images obtained by camera 20, and positioning/orientating the trajectory indicator with respect to the three dimensional representation, ensures that, to a user, the trajectory indicator(s) lie on the surface (and are fixed with respect thereto). For example, in determining the composite image sequence, the control system 10 may be configured to position and/or orientate the trajectory indicator such that it aligns to the orientation of the surface at the desired position of the trajectory indicator within an image frame(s) of the composite image sequence. In this way, the trajectory indicator appears, to a user of the system, to lie on top of a traversable surface within the environment.
As will be appreciated, any process step (or method step) described herein may be performed by running computer software, e.g. computer software 29, on one or more processors 12 as shown in Figure 1. Any such software 29 may be stored in in a location accessible by the processor 12, e.g. at memory 26.
Figure 2 illustrates an embodiment of a vehicle 50 in accordance with an embodiment of the invention. As shown, the vehicle 50 comprises a system 30 which includes control system 10, imaging devices in the form of cameras 20a, 20b, the vehicle system 16, sensor 28 and the display 24.
An embodiment of a method 100 in accordance with the invention will now be described with reference to Figure 3.
The method comprises receiving 102 one or more vehicle parameters. As described herein, the one or more parameters may be received from a vehicle system and relate to the motion of the vehicle. At 104, the method comprises receiving image data from an imaging device. Typically, this comprises receiving image data from a camera mounted on or within the vehicle configured to capture images of an environment external to the vehicle. It will, however, be appreciated that the method may comprise receiving image data from a plurality of imaging devices. At 106, the method comprises determining a vehicle trajectory (as discussed in more detail below). The vehicle trajectory is determined in dependence on the received one or more vehicle parameters. At 108, the method comprises determining a composite image sequence. The composite image sequence comprises a sequence of image frames, each image frame comprising a captured image (received from the imaging device) and a trajectory indicator indicative of the determined vehicle trajectory. At 110, a control signal indicative of the composite image sequence is output to a display for displaying the composite image sequence to an occupant of the vehicle.
In determining the composite image sequence, the method 100 comprises moving the trajectory indicator from a first position in a first image frame of the composite image sequence, to a second position in a second image frame of the composite image sequence in dependence on a distance travelled by the vehicle between the time at which an image associated with the first image frame is captured and a time at which an image associated with the second image frame is captured. In this way, the trajectory indicator is made to appear substantially stationary in a direction of travel of the vehicle with respect to one or more image objects within the composite image sequence, the one or more image objects being external to the vehicle. Determination of the composite image sequence is discussed in detail below with reference to Figures 4-9.
Figure 4 illustrates a method for determining a vehicle trajectory. Specifically Figure 4 shows how a Bicycle Model is used to model the trajectory of a vehicle 50.
In the model shown in Figure 4, the vehicle 50 comprises four steerable wheels 52, 54, 56, 58, two front wheels 52, 54 associated with a front axle 53 of the vehicle 50, and two rear wheels 56, 58 associated with a rear axle 57 of the vehicle 50. In this example, the front wheels 52, 54 are configured to be steered at the same angle, m front with respect to longitudinal axis of the vehicle, y. Accordingly, front wheels 52, 54 can be modelled as a single wheel positioned at the centre of the front axle 53 at an angle m rfront with respect to longitudinal axis of the vehicle, y. Similarly, the rear wheels 56, 58 are configured to be steered at the same angle, (rear with respect to longitudinal axis of the vehicle, y.
Accordingly, rear wheels 56, 58 can be modelled as a single wheel positioned at the centre of the rear axle 57 at an angle co rear with respect to longitudinal axis of the vehicle, y.
It will be appreciated that the vehicle 50 may comprise other steering configurations each requiring to be modelled in a similar but different way. For example, where a vehicle comprises only two steerable wheels, e.g. front wheels 52, 54, with the rear wheels 56, 58 being rotationally fixed with respect to the longitudinal axis of the vehicle y, the front wheels 52, 54 can be modelled as described above. The rear wheels 56, 58 can be modelled as a single wheel positioned at the centre of the rear axle 57 at an angle of 00 with respect to longitudinal axis of the vehicle, y. Where each of the wheels 52, 54, 56, 58 are steerable by different angles with respect to one another, each must be modelled individually.
The model shown in Figure 4 is used to determine a centre of rotation C for the vehicle 50 in dependence on the steering angle co, front of the front wheels 52, 54 and the steering angle rear of the rear wheels 56, 58. Defining the frame of reference with respect to the longitudinal axis y and the lateral axis x of the vehicle 50, with the origin at the centre of the rear axle 57, the centre of rotation C of the vehicle 50 is determined as the point of intersection of the straight lines Y * front, Yrear passing through the centres of the front and rear axles 53, 57, normal to the steering directions of the relevant axles 53, 57. The equations of these lines are given by the following equations: 'front = co tan (, front) X W [Equation 1] rear -tan(q) * x [Equation 2] where W is the wheelbase of the vehicle 50. Solving Equations 1 and 2 for x and Y gives the centre of rotation C, where x = R (radius) and V = 0 (offset from lateral axis).
Once the centre of rotation C has been found, the trajectory of the vehicle 50 is defined as a circle about the centre of rotation C at a radius R. This is shown in Figure 5 by a trajectory indicator in the form of a pair of parallel trajectory lines 40, 42. A first trajectory line 40 represents a predicted path to be traversed by a first point on the vehicle 50 and a second trajectory line 42 represents a predicted path to be traversed by a second point on the vehicle 50. Typically, and as shown in Figure 5, the first point on the vehicle 50 will be the point on the vehicle which is furthest in distance from the identified centre of rotation C. This will typically be the far-side front corner of the vehicle 50, where "far-side' refers to the side of the vehicle 50 furthest from the centre of rotation C. The second point on the vehicle 50 may correspond to the nearside rear wheel of the vehicle 50, where "nearside" refers to the side of the vehicle 50 closest to the centre of rotation C. In embodiments, the trajectory lines 40, 42 represent a predicted path traversed by other vehicle components. For example, the first trajectory line 40 represents a predicted path to be traversed by a first front wheel of the vehicle 50, and the second trajectory line 42 represents a predicted path to be traversed by a second front wheel of the vehicle 50.
The linear speed of the vehicle 50 may be used to determine an angular velocity, vvehide, of the vehicle 50 when moving along the determined trajectory, using the following: Vlinear Vvehicie =D [Equation 3] tvehicie where Vino& is the linear speed of the vehicle (which may be determined from the one or more vehicle parameters as described herein), and Rvehloo is the perpendicular distance between the longitudinal axis of the vehicle 50 and the centre of rotation C. This equation may be used to determine the angular velocity of any point on the vehicle, for instance the first point and second point for defining first and second trajectory lines 40, 42.
For calculating the angular velocity, vout, of the point on the vehicle 50 defining the first trajectory line 40, the following equation may be used: vout Vlinear Rout [Equation 4] where R"t is radius of curvature of the first trajectory line 40 about the centre of rotation C. For calculating the angular velocity, vin, of the point on the vehicle 50 defining the second trajectory line 42, the following equation may be used: VUnear [Equation 5] Rin where Rin is the radius of curvature of the second trajectory line 42 about the centre of rotation C. The extent to which the trajectory lines 40, 42 are required to be moved within the composite image sequence is different for each of the lines 40, 42 in the case where a steering angle is applied to the vehicle 50. When the vehicle is travelling in a straight line (i.e. forwards or backwards along the longitudinal axis of the vehicle), Row = (the centre of rotation C sits on the longitudinal axis of the vehicle), meaning that V"t = Vin and the distance travelled by a first point on the vehicle 50 (e.g. the farside front corner) along the first trajectory line 40 is the same as the distance travelled by a second point on the vehicle 50 (e.g. the near side rear wheel) along the second trajectory line 42. However, when any steering angle is applied, Vag is less than Vin. Similarly, the distance travelled by a first point on the vehicle 50 (e.g. the farside front corner) along the first trajectory line 40 is different to the distance travelled by a second point on the vehicle 50 (e.g. the near side rear wheel) along the second trajectory line 42.
The angular velocity of the relevant points of the vehicle 50 which define the position of trajectory lines 40, 42 may be used to determine the distance by which the lines 40, 42 must be moved within the composite image sequence between image frames in order for the trajectory indicator to appear fixed with respect to image objects external to the vehicle 50. Specifically, and as described throughout the specification, the trajectory indicator is moved between image frames within the composite image sequence in dependence on a distance travelled by the vehicle 50 between the time at which images corresponding to the relevant image frames are captured. This may be achieved by determining the distance travelled by each of the points on the vehicle defining the trajectory lines 40, 42 in the time between capturing a first image associated with a first image frame and a second image associated with a second image frame, and moving the corresponding line 40, 42 (or a section or segment of the line 40, 42 -see Figure 6 and the accompanying description, below) within the composite image sequence, in particular from a first position within the first and a second position in the second image frame, by a distance proportional to the determined distance travelled by the corresponding point on the vehicle 50. This process may be repeated over any number of image frames to create the composite image sequence. The same effect may also be achieved by moving the trajectory lines 40, 42 (or line section / segments within the composite image sequence at a velocity proportional to the determined angular velocity of the corresponding point on the vehicle 50 about the centre of rotation C Specifically, the trajectory lines 40, 42 or line sections / segments thereof are moved within the composite image sequence in a direction or with a velocity which is proportional, but in an opposite sense to the direction of motion or velocity of the vehicle 50. In this way, the present invention results in a trajectory indicator which appears substantially stationary in the direction of motion of the vehicle with respect to one or more image objects external to the vehicle. The effect to the user is that the vehicle 50 appears to travel over a stationary trajectory indicator. This is believed to provide to a user a more natural and less confusing representation of the vehicle's movement through the environment.
Figure 6 illustrates different types of trajectory indicator 39 which may be used in embodiments of the present invention.
Figure 6(a) illustrates a trajectory indicator 39 which comprises a plurality of distinct indicator sections 39a, 39b, 39c, 39d. In embodiments using the trajectory indicator 39 as shown in Figure 6(a), each of the indicator sections 39a, 39b, 39c, 39d may be moved within the composite image sequence in dependence on the movement of the vehicle such that the trajectory indicator 39 appears substantially stationary in a direction of travel of the vehicle with respect to the one or more image objects within the composite image sequence, as is described herein. For example, each of the indicator sections 39a, 39b, 39c, 39d may be moved from respective first positions within a first image frame of the composite image sequence, to respective second positions within the second image frame of the composite image sequence in dependence on a distance travelled by the vehicle 50 between the time at which an image associated with the first image frame is captured and the time at which an image associated with the second image frame is captured such each indicator section 39a, 39b, 39c, 39d appears substantially stationary in the direction of travel of the vehicle 50 within the composite image sequence with respect to the one or more image objects.
Figure 6(b) illustrates a trajectory indicator 39 which comprises only a single indicator section 39a. The section includes a hatched pattern. In embodiments using the trajectory indicator 39 shown in Figure 6(b), the position of the single indicator section 39a may be changed in dependence on the movement of the vehicle such that the trajectory indicator 39 appears substantially stationary in a direction of travel of the vehicle with respect to the one or more image objects within the composite image sequence, as is described herein. The provision of the hatched pattern provides a visual indication to the user of the system of the relative position of the trajectory indicator 39 within the composite image sequence which would not be provided were the trajectory indicator 39 to comprise a single monochrome indicator section, for example.
Figure 6(c) illustrates a trajectory indicator 39 which comprises a plurality of indicator sections 39a, 39b, 39c, 39d, 39e in the form of a pattern of shaded and unshaded regions along a trajectory line. In embodiments using the trajectory indicator 39 as shown in Figure 6(c), each of the indicator sections 39a, 39b, 39c, 39d, 39e, in this case the shaded regions, may be moved within the composite image sequence, specifically along the trajectory line in dependence on the movement of the vehicle such that the trajectory indicator 39 appears substantially stationary in a direction of travel of the vehicle with respect to the one or more image objects within the composite image sequence, as is described herein. For example, each of the indicator sections 39a, 39b, 39c, 39d, 39e may be moved along the trajectory line from respective first positions within a first image frame of the composite image sequence, to respective second positions within the second image frame of the composite image sequence in dependence on a distance travelled by the vehicle 50 between the time at which an image associated with the first image frame is captured and the time at which an image associated with the second image frame is captured such each indicator section 39a, 39b, 39c, 39d, 39e appears substantially stationary in the direction of travel of the vehicle 50 within the composite image sequence.
Figure 6(d) illustrates a trajectory indicator 39 which comprises a plurality of indicator sections 39a, 39b, 39c, 39d, 39e, 39f, 39g, 39h in the form of a series of notches along a trajectory line. In embodiments using the trajectory indicator 39 as shown in Figure 6(d), each of the indicator sections 39a, 39b, 39c, 39d, 39e, 39f, 39g, 39h, in this case the notches, may be moved within the composite image sequence, specifically along the trajectory line in dependence on the movement of the vehicle such that the trajectory indicator 39 appears substantially stationary in a direction of travel of the vehicle with respect to the one or more image objects within the composite image sequence, as is described herein. For example, each of the indicator sections 39a, 39b, 39c, 39d, 39e, 39f, 39g, 39h may be moved along the trajectory line from respective first positions within a first image frame of the composite image sequence, to respective second positions within the second image frame of the composite image sequence in dependence on a distance travelled by the vehicle 50 between the time at which an image associated with the first image frame is captured and the time at which an image associated with the second image frame is captured such each indicator section 39a, 39b, 39c, 39d, 39e, 39f, 39g, 39h appears substantially stationary in the direction of travel of the vehicle 50 within the composite image sequence with respect to the one or more image objects.
A method of controlling the position of the trajectory indicator between image frames of the composite image sequence to convey movement of the trajectory indicator within the composite image sequence is illustrated in Figures 7A and 7B. These Figures show the position of a camera 20, and virtual display 60 with respect to the vehicle 50, and projection outlines 62, 64. The projections outlines 62, 64 link the position of a trajectory indicator section 39a to the position of the camera 20. Specifically, projection outline 62 links the position of a first end of the trajectory indicator section 39a to the camera 20, and projection outline 64 links the position of a second end of the trajectory indicator section 39a to the camera 20. The position at which the projection outlines 62, 64 intersect the virtual display 60 illustrates how the position of a representation of the trajectory indicator section 39a within an image/image sequence, e.g. the composite image sequence, can convey depth to a user viewing the display, as is described herein.
In Figure 7A, a trajectory indicator in the form of two parallel trajectory lines 40, 42 is shown in a first position with respect to a vehicle 50 at time to. Likewise, in Figure 7B the trajectory lines 40, 42 are shown in a second position with respect to a vehicle 50 at time ti. Between to and ti the vehicle has moved in direction of motion D. As described herein, in order to convey to a user that the trajectory indicator remains fixed with respect to the vehicle's environment rather than the vehicle 50, the trajectory indicator is moved within the composite image sequence to appear to move towards the vehicle 50 by a distance or at a velocity dependent on the distance travelled by the vehicle 50 between to and As shown, the projection outlines 62, 64 intersect the virtual display 60 at respective first positions at time to, as shown in Figure 7A. With the indicator section 39a in a second position as shown in Figure 7B (at time t1), the points of intersection of the projection outlines 62, 64 with the virtual display 60 are changed. Specifically, the points of intersection move towards an edge of the virtual display 60, and provide a greater area of intersection with the virtual display 60. Accordingly, in order to convey depth within the composite image sequence, or in order to convey movement of the trajectory indicator towards the vehicle, the trajectory indicator must be moved within the composite image toward an edge of the image frame, and increased in size.
Although shown in Figures 7A and 7B as a 1-dimensional virtual display 60, the same technique can be applied equally for a virtual display in 2 (or more) dimensions, as will be appreciated. Figure 7C illustrates such an embodiment.
Specifically, in Figure 70, a representation of a composite image sequence comprising a first image frame 32 and a second image frame 34 is shown. The first image frame 32 represents a representation of an image obtained from a camera mounted on or within, or integral to a vehicle 50 at time to. The first image frame 32 includes a trajectory indicator in the form of trajectory lines 40, 42. A single trajectory indicator section 39a is shown highlighted.
Similarly, the second image frame 34 represents a representation of an image obtained from the camera at time t1, subsequent to to. The second image frame 34 also includes the trajectory lines 40, 42, and indicator section 39a. Between to and ti, the vehicle 50 has moved in a longitudinal direction. Accordingly, the indicator section 39a has been moved between a first position in the first image frame 32 to a second position in a second image frame 34 in order to convey movement of the indicator section 39a with respect to the vehicle 50. Specifically, the indicator section 39a is moved towards a corner of the composite image sequence, or closer to the position of the vehicle 50 within the image sequence and towards an edge of within the image sequence. By moving the indicator section 39a by a distance within the composite image sequence in dependence on, and preferably proportional to, a distance travelled by the vehicle between time to and ti, the arrangement of the present invention can convey movement of the trajectory indicator with respect to the vehicle 50 within the composite image sequence. Or more preferably, moving the trajectory indicator within the composite image sequence such that it appears substantially stationary with respect to image objects external to the vehicle -i.e. such that the vehicle 50 appears, to a user, to drive over the trajectory indicator.
As shown in Figure 70, as the trajectory indicator sections, e.g. section 39a, are moved between image frames to convey movement with respect to the vehicle, new indicator sections 39b are introduced at the end of the trajectory lines 40, 42 furthest from the vehicle 50. In the illustrated embodiment, the new sections 39b are introduced to maintain a set length of the trajectory lines 40, 42. It will be appreciated, however, that new trajectory indicators 39b may be introduced to maintain the number of sections in each trajectory line 40, 42. In embodiments, the new indicator sections 39b are "faded" in, e.g. by increasing the opacity of the new sections 39b.
Figures 8A & 8B show a graphical representation of a composite image sequence formed in embodiments of the present invention in comparison to composite image sequences formed using prior art systems and methods. Where like components and features are present, the same reference numerals are used in both Figures 8A and 8B to illustrate the differences.
The graphical representations shown in Figures 8A & 8B are provided from a perspective above the vehicle 50. It will be appreciated that such a representation may be provided by suitable processing of image data obtained from one or more cameras mounted on or within the vehicle 50.
Figure 8A is an example graphical representation of a composite image sequence formed using prior art systems and methods. The representation shows a first image frame 32 comprising an image captured at time to, and a second image frame 34 comprising an image captured at time ti. Within each image frame 32, 34 there is provided a graphical representation of a portion of the vehicle 50, a trajectory indicator in the form of two parallel trajectory lines 40, 42 each having a plurality of indicator sections (only 39a labelled here for conciseness). For simplicity, the trajectory lines 40, 42 are shown straight ahead of the vehicle 50 in the direction of travel D. It will, however, be appreciated that these lines 40, 42 could have any curvature in dependence on a determined trajectory of the vehicle (determined as described hereinabove, for example). Within each image frame 32, 34 there is shown the position of two image objects 36, 38 within the environment external to the vehicle 50. A first image object in the form of a section of a traversable surface (surface section) is shown by box 36. A second image object in the form of an obstacle 38 is also shown. The surface section 36 and obstacle 38 are positioned within the environment external to the vehicle and are thus captured within the images corresponding to the first and second image frames 32, 34.
Between the time at which an image associated with the first image frame 32 is captured (to), and a time at which an image associated with the second image frame 34 is captured (ti), the vehicle has moved a given distance in the direction of travel D. Accordingly, within the second image frame 34 of the composite image sequence, the surface section 36 and the obstacle 38 appear comparatively closer to the front of the vehicle 50 when compared with their position in the first image frame 32. However, in such systems, the trajectory lines 40, 42 are fixed with respect to the location of the vehicle within the composite image sequence. Accordingly, and as shown in Figure 8A, the trajectory lines 40, 42 and hence indicator sections 39a, etc. appear to move with the vehicle with respect to the surface section 36 and obstacle 38. This appears to an occupant using the system that the trajectory lines 40, 42 are 'floating' over the surface which can appear unnatural and/or confusing to the occupant as it is difficult to assess the exact correspondence of any given point on a trajectory indicator of this type and image objects external to the vehicle 50.
To overcome such issues, the present invention provides a means to 'fix' the position of the trajectory indicator with respect to one or more image objects external to the vehicle, rather than to the vehicle 50 itself. As discussed herein, this is achieved in the present invention by moving the trajectory indicator (or individual sections / segments thereof) between image frames of the composite image sequence in dependence on movement of the vehicle 50 such that the trajectory indicator appears substantially stationary in the direction of travel D of the vehicle 50 with respect to image objects 36, 38 external to the vehicle 50.
As with Figure 8A, Figure 8B shows a first image frame 32 comprising an image captured at time to, and a second image frame 34 comprising an image captured at time ti. Similarly, within each image frame 32, 34 there is provided a graphical representation of a portion of a vehicle 50, a trajectory indicator in the form of two parallel trajectory lines 40, 42 each having a plurality of indicator sections (only 39a labelled here for conciseness). Surface section 36 and obstacle 38 are also shown and similarly are positioned within the environment external to the vehicle and are thus captured within the images corresponding to the first and second image frames 32, 34.
As shown, between to and tl, the vehicle has moved a given distance in the direction of travel D. Accordingly, within the second image frame 34 of the composite image sequence, the surface section 36 and the obstacle 38 appear comparatively closer to the front of the vehicle 50 when compared with their position in the first image frame 32. The present invention differs from the prior art system shown in Figure 8A in that the position of the trajectory lines 40, 42, and specifically the position of the indicator sections is changed between the first and the second image frames 32, 34. Specifically, and focusing in particular on indicator section 39a (although the following applies equally to all indicator sections), indicator section 39a is moved from a first position as shown in the first image frame 32 where the indicator section 39a is mostly in front of the representation of the vehicle 50, to a second position as shown in the second image frame 34 where the indicator section 39a is presented alongside the representation of the vehicle 50. The distance between the first position of the indicator section 39a and the second position of the indicator section 39a is determined in dependence on a distance travelled by the vehicle 50 between to and tl. In embodiments, the distance between the first and second positions of the indicator section 39a may be directly proportional to the distance travelled by the vehicle 50 between to and tl. In embodiments, the present invention may include performing an analysis of the images captured by the associated imaging device to determine movement of one or more image objects (e.g. surface section 36, obstacle 38). The distance between the first and second positions of the indicator section 39a may then be determined in dependence on a distance moved by the one or more image objects between image frames of the composite image sequence. In this way, the trajectory lines 40, 42 and hence indicator sections 39a, etc. appear to be fixed in position with respect to the surface section 36 and obstacle 38. This appears to an occupant using the system that the vehicle 50 is travelling over the trajectory lines 40, 42 thereby providing a better representation of the correspondence of the position of any given point on a trajectory indicator and image objects external to the vehicle 50.
Figure 9 illustrates an equivalent scenario as shown in Figure 8B. However, the composite image sequence is shown from a different perspective, in this case looking out from the vehicle rather than overhead as in Figure 8B. In addition, the trajectory lines 40, 42 are shown as curved a predicted vehicle path where a steering angle has been applied to the vehicle 50, and the second image frame 34 includes an additional obstacle 38' not visible in the first image frame 32. However, the invention may be implemented in the same way as that shown in the preceding Figures and discussed herein.
Again, between to and tl, the vehicle has moved a given distance in the direction of travel D. Accordingly, within the second image frame 34 of the composite image sequence, the surface section 36 and the obstacle 38 appear comparatively closer to the front of the vehicle 50 when compared with their position in the first image frame 32. Indicator section 39a is moved from a first position as shown in the first image frame 32 where the indicator section 39a is in front of the representation of the vehicle 50, to a second position as shown in the second image frame 34 where the indicator section 39a is presented next to and in part underneath the representation of the vehicle 50. The distance between the first position of the indicator section 39a and the second position of the indicator section 39a is determined in dependence on a distance travelled by the vehicle 50 between to and t1. In this way, the trajectory lines 40, 42 and hence indicator sections 39a, etc. appear to be fixed in position with respect to the surface section 36 and obstacle 38. This appears to an occupant using the system that the vehicle 50 is travelling over the trajectory lines 40, 42 thereby providing a better representation of the correspondence of the position of any given point on a trajectory indicator and image objects 36, 38, 38' external to the vehicle 50.
It will be appreciated that in the illustrated embodiments the movement of the trajectory indicator between image frames has been exaggerated to illustrate the operational use of the invention. It is expected that the trajectory indicator will move a relatively minimal distance between consecutive image frames as the vehicle moves within the environment. The composite image sequence preferably comprises image frames relating to image data captured at a rate high enough to convey substantially continuous movement of the vehicle within the environment. For example, the image data is captured at a rate such that a composite image sequence can be formed having an image frame rate of 50 frames per second, or higher.
It will be appreciated that embodiments of the present invention can be realised in the form of hardware, software or a combination of hardware and software. Any such software may be stored in the form of volatile or non-volatile storage such as, for example, a storage device like a ROM, whether erasable or rewritable or not, or in the form of memory such as, for example, RAM, memory chips, device or integrated circuits or on an optically or magnetically readable medium such as, for example, a CD, DVD, magnetic disk or magnetic tape. It will be appreciated that the storage devices and storage media are embodiments of machine-readable storage that are suitable for storing a program or programs that, when executed, implement embodiments of the present invention. Accordingly, embodiments provide a computer program comprising code for implementing a system or method as claimed, and a machine-readable storage storing such a program (e.g. a non-transitory computer readable medium). Still further, embodiments of the present invention may be conveyed electronically via any medium such as a communication signal carried over a wired or wireless connection and embodiments suitably encompass the same.
It will be appreciated that various changes and modifications can be made to the present invention without departing from the scope of the present application.

Claims (20)

  1. CLAIMS1 A control system for providing assistance to an occupant of a vehicle, the control system comprising one or more controllers, configured to: receive one or more vehicle parameters; receive image data from an imaging device comprising captured images of an environment external to the vehicle; determine a vehicle trajectory in dependence on the or each parameter; determine a composite image sequence comprising a sequence of image frames, each image frame comprising a captured image and a trajectory indicator indicative of the determined vehicle trajectory; and output a signal indicative of the composite image sequence to a display for displaying the composite image sequence to the occupant of the vehicle; wherein the control system is configured to move the trajectory indicator from a first position in a first image frame of the composite image sequence, to a second position in a second image frame of the composite image sequence in dependence on a distance travelled by the vehicle between the time at which an image associated with the first image frame is captured and the time at which an image associated with the second image frame is captured such that the trajectory indicator appears substantially stationary in a direction of travel of the vehicle with respect to one or more image objects within the composite image sequence, the one or more image objects being external to the vehicle.
  2. 2. A control system of claim 1, wherein the one or more controllers collectively comprise: at least one electronic processor having one or more electrical inputs for receiving the one or more vehicle parameters and/or the image data; and at least one electronic memory device operatively coupled to the at least one electronic processor and having instructions stored therein; wherein the at least one electronic processor is configured to access the at least one memory device and execute the instructions stored therein so as to determine the composite image sequence.
  3. 3. A control system as claimed in claim 1 or claim 2, wherein the first position and the second position are separated within the composite image sequence by a distance directly proportional to the distance travelled by the vehicle between the time at 4. 5. 6. 7. 8. 9. 10.which the image associated with the first image frame is captured and the time at which the image associated with the second image frame is captured.
  4. A control system as claimed in any preceding claim, wherein the trajectory indicator comprises a plurality of indicator sections.
  5. A control system as claimed in claim 4, configured to move each of the indicator sections from respective first positions within the first image frame of the composite image sequence, to respective second positions within the second image frame of the composite image sequence in dependence on a distance travelled by the vehicle between the time at which an image associated with the first image frame is captured and the time at which an image associated with the second image frame is captured such each indicator section appears substantially stationary in the direction of travel of the vehicle within the composite image sequence with respect to the one or more image objects.
  6. A control system as claimed in claim 4 or claim 5, wherein the trajectory indicator comprises one or more lines indicative of a predicted path to be traversed by one or more wheels of the vehicle.
  7. A control system as claimed in claim 6, wherein the trajectory indicator comprises: a first trajectory line representative of a predicted path to be traversed by a first wheel of the vehicle; and a second trajectory line representative of a predicted path to be traversed by a second wheel of the vehicle.
  8. A control system as claimed in any preceding claim, wherein the one or more image objects comprise a traversable surface within the composite image sequence.
  9. A control system as claimed in claim 8, configured to move the trajectory indicator from the first position to the second position such that, to the user, the vehicle appears to move over the trajectory indicator as the vehicle traverses the surface along the determined vehicle trajectory.
  10. A control system as claimed in claim 8 or claim 9, configured to: receive sensor data from one or more sensors indicative of a topography of the surface; and 11. 12. 13. 14. 15. 16: 17:determine a position and/or orientation of the trajectory indicator within one or more image frames of the composite image sequence in dependence on the received sensor data.
  11. A control system as claimed in claim 10, configured to: process the received sensor data to determine a surface profile of the surface; and determine the position and/or orientation of the trajectory indicator within one or more image frames of the composite image sequence in dependence on the determined surface profile such that the trajectory indicator appears to lie on the surface within the composite image sequence.
  12. A control system of any preceding claim, configured to: process the received image data to determine a speed and/or direction of motion of the one or more image objects within the composite image sequence; and move the trajectory indicator from the first position to the second position in dependence on the determined speed and/or direction of motion of the one or more image objects within the composite image sequence.
  13. A control system as claimed in any preceding claim wherein the one or more vehicle parameters comprise a steering angle of the vehicle.
  14. A control system as claimed in claim 13, wherein the steering angle comprises an angle of a steering wheel of the vehicle or the angle of one or more steerable wheels of the vehicle.
  15. A control system as claimed in any preceding claim, wherein the one or more vehicle parameters comprise a velocity of the vehicle within the composite image sequence at a speed and in a direction proportional to the velocity of the vehicle.
  16. A system comprising the control system of any one of claims 1 to 15, and an imaging device configured to capture one or more images of an environment external to the vehicle.
  17. A vehicle comprising the control system of any one of claims 1 to 15, or a system according to claim 16.
  18. 18: A method for providing assistance to an occupant of a vehicle, the method comprising: receiving one or more vehicle parameters; receiving image data comprising captured images of an environment external to the vehicle; determining a vehicle trajectory in dependence on the or each parameter; determining a composite image sequence comprising a sequence of image frames, each image frame comprising a captured image and a trajectory indicator indicative of the determined vehicle trajectory; and outputting a signal indicative of the composite image sequence to a display for displaying the composite image sequence to the occupant of the vehicle; wherein the method comprises moving the trajectory indicator from a first position in a first image frame of the composite image sequence, to a second position in a second image frame of the composite image sequence in dependence on a distance travelled by the vehicle between the time at which an image associated with the first image frame is captured and a time at which an image associated with the second image frame is captured such that to the trajectory indicator appears substantially stationary in a direction of travel of the vehicle with respect to one or more image objects within the composite image sequence, the one or more image objects being external to the vehicle.
  19. 19: A non-transitory computer readable medium having instructions stored therein which, when executed by one or more electronic processors, cause the one or more electronic processors to perform the method according to claim 18.
  20. 20. Computer software which, when executed by one or more processors, causes performance of the method of claim 18.
GB1900332.6A 2019-01-10 2019-01-10 A control system, system and method for providing assistance to an occupant of a vehicle Active GB2580400B (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
GB1900332.6A GB2580400B (en) 2019-01-10 2019-01-10 A control system, system and method for providing assistance to an occupant of a vehicle
US17/422,163 US11919513B2 (en) 2019-01-10 2020-01-06 Control system, system and method for providing assistance to an occupant of a vehicle
DE112020000391.4T DE112020000391T5 (en) 2019-01-10 2020-01-06 A control system, system and method for assisting a vehicle occupant
PCT/EP2020/050119 WO2020144129A1 (en) 2019-01-10 2020-01-06 A control system, system and method for providing assistance to an occupant of a vehicle

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
GB1900332.6A GB2580400B (en) 2019-01-10 2019-01-10 A control system, system and method for providing assistance to an occupant of a vehicle

Publications (3)

Publication Number Publication Date
GB201900332D0 GB201900332D0 (en) 2019-02-27
GB2580400A true GB2580400A (en) 2020-07-22
GB2580400B GB2580400B (en) 2021-11-03

Family

ID=65528044

Family Applications (1)

Application Number Title Priority Date Filing Date
GB1900332.6A Active GB2580400B (en) 2019-01-10 2019-01-10 A control system, system and method for providing assistance to an occupant of a vehicle

Country Status (1)

Country Link
GB (1) GB2580400B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210253127A1 (en) * 2020-02-19 2021-08-19 Beijing Baidu Netcom Science And Technology Co., Ltd. Bend driving control method for autonomous vehicle, device, and medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102005035412A1 (en) * 2005-07-28 2007-02-01 Robert Bosch Gmbh Driver assistance device for motor vehicle, has display representing road surface and lane pathway that are narrower than detected actual pathway, which are represented only when measure of detected tiredness of driver is exceeded
US20080195315A1 (en) * 2004-09-28 2008-08-14 National University Corporation Kumamoto University Movable-Body Navigation Information Display Method and Movable-Body Navigation Information Display Unit
US20180244153A1 (en) * 2014-12-17 2018-08-30 Nippon Seiki Co., Ltd. Vehicle information projecting system and vehicle information projecting method

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5161760B2 (en) * 2008-12-26 2013-03-13 株式会社東芝 In-vehicle display system and display method
JP6273976B2 (en) * 2014-03-31 2018-02-07 株式会社デンソー Display control device for vehicle
CN109477967A (en) * 2016-05-11 2019-03-15 威瑞股份公司 Has the head-up-display system of variable focal plane
JP6695049B2 (en) * 2017-05-10 2020-05-20 パナソニックIpマネジメント株式会社 Display device and display control method
JP2019095851A (en) * 2017-11-17 2019-06-20 アイシン・エィ・ダブリュ株式会社 Vehicle drive auxiliary system, vehicle drive auxiliary method and vehicle drive auxiliary program

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080195315A1 (en) * 2004-09-28 2008-08-14 National University Corporation Kumamoto University Movable-Body Navigation Information Display Method and Movable-Body Navigation Information Display Unit
DE102005035412A1 (en) * 2005-07-28 2007-02-01 Robert Bosch Gmbh Driver assistance device for motor vehicle, has display representing road surface and lane pathway that are narrower than detected actual pathway, which are represented only when measure of detected tiredness of driver is exceeded
US20180244153A1 (en) * 2014-12-17 2018-08-30 Nippon Seiki Co., Ltd. Vehicle information projecting system and vehicle information projecting method

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210253127A1 (en) * 2020-02-19 2021-08-19 Beijing Baidu Netcom Science And Technology Co., Ltd. Bend driving control method for autonomous vehicle, device, and medium
US11767029B2 (en) * 2020-02-19 2023-09-26 Beijing Baidu Netcom Science And Technology Co., Ltd. Bend driving control method for autonomous vehicle, device, and medium

Also Published As

Publication number Publication date
GB2580400B (en) 2021-11-03
GB201900332D0 (en) 2019-02-27

Similar Documents

Publication Publication Date Title
JP7155204B2 (en) System and method for calibrating the neutral position of a steering wheel
CN108692699B (en) Vehicle and method for collision avoidance assistance
US10046803B2 (en) Vehicle control system
US10354151B2 (en) Method of detecting obstacle around vehicle
US9734719B2 (en) Method and apparatus for guiding a vehicle in the surroundings of an object
US9902425B2 (en) System for guiding trailer along target route during reversing maneuver
EP3623761B1 (en) Localization method and apparatus of displaying virtual object in augmented reality
EP2209091B1 (en) System and method for object motion detection based on multiple 3D warping and vehicle equipped with such system
CN111060094A (en) Vehicle positioning method and device
KR102508843B1 (en) Method and device for the estimation of car egomotion from surround view images
JP2018536572A (en) Control system for steering means of motor-driven vehicles in a situation where collision with an obstacle is imminent
CN106043306A (en) Vehicle control device
CN111123950A (en) Driving control method and device and vehicle
CN112433531A (en) Trajectory tracking method and device for automatic driving vehicle and computer equipment
EP4049916A1 (en) Vehicle control method, vehicle control system, and vehicle
GB2580400A (en) A control system, system and method for providing assistance to an occupant of a vehicle
US11919513B2 (en) Control system, system and method for providing assistance to an occupant of a vehicle
CN114644014A (en) Intelligent driving method based on lane line and related equipment
CN114291071A (en) Method and system for judging vehicle stability control active intervention opportunity, readable storage medium and vehicle
GB2580401A (en) A control system, system and method for providing assistance to an occupant of a vehicle
CN112859127A (en) Navigation device and operation method thereof
Mancuso Study and implementation of lane detection and lane keeping for autonomous driving vehicles
CN117456715A (en) Creating a travel zone from positioning data and vehicle dimensions