US20170293837A1 - Multi-Modal Driving Danger Prediction System for Automobiles - Google Patents
Multi-Modal Driving Danger Prediction System for Automobiles Download PDFInfo
- Publication number
- US20170293837A1 US20170293837A1 US15/478,823 US201715478823A US2017293837A1 US 20170293837 A1 US20170293837 A1 US 20170293837A1 US 201715478823 A US201715478823 A US 201715478823A US 2017293837 A1 US2017293837 A1 US 2017293837A1
- Authority
- US
- United States
- Prior art keywords
- data
- extractor
- vehicle
- training
- modal sensors
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/64—Three-dimensional objects
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60Q—ARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
- B60Q9/00—Arrangement or adaptation of signal devices not provided for in one of main groups B60Q1/00 - B60Q7/00, e.g. haptic signalling
- B60Q9/008—Arrangement or adaptation of signal devices not provided for in one of main groups B60Q1/00 - B60Q7/00, e.g. haptic signalling for anti-collision purposes
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2413—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
- G06V10/443—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by matching or filtering
- G06V10/449—Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters
- G06V10/451—Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters with interaction between the filter responses, e.g. cortical complex cells
- G06V10/454—Integrating the filters into a hierarchical structure, e.g. convolutional neural networks [CNN]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/58—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
Definitions
- the present invention relates to neural networks and, more particularly, to a multi-modal driving danger prediction system for automobiles using deep convolutional neural networks.
- a computer-implemented method for training a deep neural network to recognize traffic scenes (TSs) from multi-modal sensors and knowledge data includes receiving data from the multi-modal sensors and the knowledge data and extracting feature maps from the multi-modal sensors and the knowledge data by using a traffic participant (TS) extractor to generate a first set of data, using a static objects extractor to generate a second set of data, and using an additional information extractor.
- the computer-implemented method further includes training the deep neural network, with training data to recognize the TSs from a viewpoint of a vehicle.
- a system for training a deep neural network to recognize traffic scenes (TS s) from multi-modal sensors and knowledge data includes a memory and a processor in communication with the memory, wherein the processor is configured to receive data from the multi-modal sensors and the knowledge data and extract feature maps from the multi-modal sensors and the knowledge data by using a traffic participant (TS) extractor to generate a first set of data, using a static objects extractor to generate a second set of data, and using an additional information extractor.
- the computer-implemented system further includes training the deep neural network, with training data to recognize the TSs from a viewpoint of a vehicle.
- a non-transitory computer-readable storage medium comprising a computer-readable program for training a deep neural network to recognize traffic scenes (TSs) from multi-modal sensors and knowledge data
- the computer-readable program when executed on a computer causes the computer to perform the steps of receiving data from the multi-modal sensors and the knowledge data and extracting feature maps from the multi-modal sensors and the knowledge data by using a traffic participant (TS) extractor to generate a first set of data, using a static objects extractor to generate a second set of data, and using an additional information extractor.
- the computer-implemented method further includes training the deep neural network, with training data to recognize the TSs from a viewpoint of a vehicle.
- FIG. 1 is a block/flow diagram of a method for predicting dangerous situations ahead of time, in accordance with embodiments of the present invention
- FIG. 2 is a block/flow diagram of an architecture of a deep convolutional neural network, in accordance with embodiments of the present invention.
- FIG. 3 is a block/flow diagram of a user interface (UI) of a vehicle illustrating a dangerous situation on the road ahead, in accordance with embodiments of the present invention
- FIG. 4 is a block/flow diagram of an interior of an automobile depicting a navigation device mounted on a windshield or incorporated within a dashboard and/or console area, in accordance with embodiments of the present invention.
- FIG. 5 is a block/flow diagram of the navigation device of FIG. 4 including a video camera for predicting one or more dangerous driving situations while a user drives the car, in accordance with embodiments of the present invention.
- the exemplary embodiments of the present invention introduce driver assistance systems which support the driver of a motor vehicle in driving the vehicle and/or warn him/her of special dangerous situations, for example, an imminent collision, or automatically initiate measures for avoiding the collision or for mitigating the consequences of the collision.
- Such determinations are based on a position finding system, in particular a radar or light detection and ranging (lidar) system, locating objects in the surroundings of the vehicle, such as other vehicles.
- the distances and relative velocities or speeds of the objects located in the surroundings of the vehicle are measured accurately by employing such position finding system to allow for a prediction of a special situation ahead.
- the present invention provides the advantage that it makes detection of such special situations possible, so that appropriate countermeasures may be initiated or at least a warning may be output to the driver.
- the present invention is based on the principle that, while locating a single stationary object on the road is insufficient for inferring a potentially special (dangerous) situation, such situations may very well be detected when a set of a plurality of objects is considered together, and a characteristic signature is recognizable from the distribution of the position finding data such as, but not limited to, distances, relative velocities, relative speeds, and/or relative accelerations, which indicate a certain class of special situations.
- the exemplary embodiments of the present invention relate to a driving situation predicting device having a video capturing unit for capturing real-time video of a driver driving a car and for making predictions on whether a special driving situation is imminent. If a special driving situation is imminent, the driver is notified with a warning.
- traffic scenes are described by traffic participants (TP) (cars, trucks, bikes, pedestrians, etc.), geometry of the road (intersection, lanes, pedestrian crossing, signals, signs, etc.), information about surroundings (schools, train stations, bus stops, temperature, weather, traffic, etc.) and a status of the driving car (speed, driver status, car occupancy, radio, phone usage, etc.).
- TP traffic participants
- TP traffic participants
- the TSs can be captured by a plurality of sensors and video devices, such as video cameras.
- the TSs can be captured continuously and in real-time.
- the TSs can be traffic environments pertaining to: peripheral road information related to a road located around a vehicle on which a dangerous drive predicting apparatus has been mounted; signal mechanism information related to a signal mechanism; peripheral vehicle information related to other vehicles which are driven around a vehicle; and parking lot information for indicating whether or not a parking lot is located around the vehicles and the like.
- the above-described peripheral road information corresponds to information for indicating whether a road along which the vehicle is driven is a single lane, a crossing, or a priority-sided road.
- the above-described signal mechanism information corresponds to information for indicating whether or not a signal mechanism is present, information for indicating whether or not the vehicle approaches a signal mechanism, and information related to signal light turning situations of the signal mechanism.
- the above-described peripheral vehicle information corresponds to information for indicating whether or not a preceding drive vehicle is present, information for indicating whether or not a succeeding drive vehicle is present, and information for indicating whether or not an intersecting vehicle is present.
- the judgement as to whether or not the vehicle is being driven on a single lane may be made by judging, for instance, information held in a car navigation system to be mounted on the vehicle, or by detecting a lane line by operating a camera to be mounted on the vehicle so as to judge a curved degree of the lane line.
- the judgement as to whether or not the vehicle is being driven on a crossing may be made by judging, for instance, information held in a car navigation system to be mounted on the vehicle, or by detecting a signal mechanism, or a stop line by operating the camera to be mounted on the vehicle so as to measure a distance.
- a driving danger prediction system is realized by continuously matching the current TS to a codebook of TSs or predetermined TSs that have been identified as leading to special situations, such as, but not limited to dangerous situations for a driver of a vehicle 12 . When a match occurs, a warning can be transmitted to the driver of the vehicle 12 ( FIG. 3 ).
- a traffic environment detecting unit (navigation units presented herein) can detect a shape of a road around the vehicle, into which the vehicle tries to enter; a condition whether or not a signal mechanism is present; a signal light turning situation; a condition whether or not a preceding drive vehicle is present; and the like.
- a driving characteristic determining unit (incorporated with the navigation units herein) can judge a traffic environment based upon the detected information, namely judges whether or not the traffic environment corresponds to such a previously determined traffic environment to which a dangerous driving situation has been assigned to.
- embodiments of the present invention describe how to fit an end-to-end convolutional neural network learning approach (convnet) to the challenge of matching TSs.
- FIG. 1 depicts the flowchart of the system.
- Data arrives into the system from at least two sources.
- One source is sensor frame data 100 obtained dynamically at regular synchronized sampling intervals from various sensors throughout the car (e.g., radar, controller area network (CAN), video, global positioning system (GPS), etc.), and the second source of data 110 is from non-sampled sources such as the GPS street map or externally generated information such as weather, traffic info, etc.
- the TSs can be evaluated for captured content, such as cars, trucks, bikes, pedestrians, intersection, lanes, pedestrian crossing, signals, signs, schools, train stations, bus stops, etc. All this information is processed to create a number of feature maps.
- Feature maps 300 , 310 , 320 which are two-dimensional pixel maps, are suitable for direct input into a convnet, and are extracted from input data to be entered into a training module 610 .
- the TP extractor 200 uses sensors to extract position, size, type and velocities/speeds of traffic participants. Specialized automotive radar units are capable of extracting positions and velocities of several TPs simultaneously. Light detection and ranging (LIDAR) units are more precise and can sense the size of TPs more accurately. Video can be employed along with advanced processing techniques such as structure from motion (SFM) to obtain position, size and even type of TPs. When size and type cannot be obtained from one of the sensors, standard default values are utilized.
- SFM structure from motion
- the radar system can detect the presence of an object (e.g., car) in a detection zone, obtain its distance and relative velocity.
- the camera system alone can also be used to detect the presence of the object (e.g., car) in its detection zone, obtain its distance, relative velocity and size information.
- the radar system can be used to detect the presence of the object, its distance and relative velocity and the vision system can be used to confirm the above items and also to provide additional information about the size of the object.
- the feature maps 300 are built from the TP extractor 200 .
- the first feature map encodes position, size and type of each TP with oriented filled rectangles.
- the orientation is defined by a velocity vector and the type is encoded by filling a rectangle with a particular value.
- the second and third feature maps are velocity feature maps. The same approach as above is employed, except that normalized relative velocity amplitudes (forward and lateral, respectively) are used as fill values for the rectangles.
- the GPS sensor is employed to obtain an absolute position of the car in world coordinates and query a portion (e.g., a box) of a street map around those coordinates. Such queries are made either to a GPS map unit in the car or to a cloud server via, e.g., wireless connection.
- detailed lane information may not be available (e.g., multi-lane roadways may be returned as single lines). In that case, video or dedicated lane-following units may provide detailed lane configurations. Otherwise, standard lane widths are used as an approximation.
- Each individual road lane is then drawn to the feature map using a value reflecting a type of road it is (e.g., highway, rural, etc.) and a direction of a lane (e.g., same as driving vehicle, opposite).
- Other feature maps may be generated from building, signage, tree, fences and other static objects. Bridge or tunnel segments may also be encoded to create separate feature maps.
- the third type of feature map extractor 220 creates or generates feature maps including a single value 320 .
- Such feature maps allow certain global parameters of a TS to be combined with a dynamic TP and static feature maps.
- One such parameter is a speed of a driving car.
- Such parameter is not included in the TP feature maps as all speeds are relative to the driving car and hence need to be inputted separately.
- Other global parameters such as weather (e.g., rainy, snowy, foggy, dark, etc.) can be inputted into those feature maps.
- vehicle speed information is related to a speed of the vehicle
- acceleration information is related to acceleration of the vehicle
- steering angle information is related to a steering angle thereof
- vehicle position information is related to a position of the vehicle
- vehicle-to-vehicle distance information is related to a vehicle-to-vehicle distance.
- the drive information is such an information containing vehicle information.
- FIG. 2 shows an example of a possible convnet with, e.g., five (5) feature maps of, e.g., size 20 ⁇ 100 pixels.
- An output layer of the convnet has as many outputs as a number of traffic scenes (TS) to be recognized.
- TS traffic scenes
- a deep convolutional network provides for a set of initial feature maps.
- These feature maps can be, e.g., a position feature map 701 , a speed feature map 703 (in a direction x), a speed feature map 705 (in a direction y), a road feature map 707 , and a global knowledge feature map 709 .
- a convolution operation takes place to generate a second layer of feature maps 711 .
- a sub-sampling operation then takes place to generate a third layer of feature maps 713 .
- Another convolution operation takes place to generate a fourth layer of feature maps 715 .
- Another subsampling operation takes place to generate a fifth layer of feature maps 717 .
- the data is then reshaped 719 and fully connected 721 .
- the fully connected layer 721 uses a softmax activation function 723 in the output layer.
- a training set of labeled examples 500 is created.
- a sufficient number of training examples 500 is employed to train the deep convnet, so that it can generalize adequately.
- the exemplary embodiments of the present invention also describe an approach to generate large amounts of labeled sensor frame data 130 from recorded naturalistic driving datasets.
- a set of constraints on sensor data variables is defined for all TSs that are to be detected.
- a set of constraints for a particular TS would be: the driving vehicle's speed should be above 100 km/h, there should be a car in front at less than 30 m and that car's relative speed should be negative. That kind of description is a natural way to define a particular TS, yet can also be easily transcribed into computer code, resulting in a list of coded rules 120 for each TS.
- the variables of corresponding sensors are fed to the rule verification module 400 . When all rules describing a TS are verified, the data frame 500 is labeled as belonging to that TS, otherwise a negative label is given to the data frame.
- a convnet model 600 can be trained by employing machine learning techniques.
- negative examples need to be introduced into the training set.
- the number of negative examples is a parameter of the system.
- a method and system for training the deep convnet to recognize complex traffic scenes from multi-modal sensors and knowledge data is further provided.
- the methods and systems further use a common representation to combine sensors of different modalities.
- the methods and systems use a two dimensional top-view representation of the traffic scene (TS) suitable for direct input to the deep convnet via a set of two dimensional pixel maps where the origin is set by a position of the driving car.
- TS traffic scene
- Radar sensor data is represented in three pixel maps by: (i) drawing oriented filled rectangles at positions and sizes corresponding to detected radar objects, at orientations based on a direction of their velocity, and filling pixels with a particular value, (ii) for the first pixel map, a value represents a type of object, (iii) for a second pixel map, a value represents an amplitude of a forward relative velocity, and (iv) for a third pixel map, a value represents an amplitude of a lateral relative velocity.
- the methods and systems further introduce a GPS sensor for providing an instantaneous absolute location of the driving car on a street map that is represented by a set of pixel maps as follows: (a) a first pixel map is for lane-level detail around the driving car and is obtained by querying a map subsystem, (i) for systems that don't have lane-level detail, lane information can be obtained by video processing, or by using default lane width values for each type of roadway, and (ii) lanes are drawn onto a pixel map with values corresponding to a type of roadway. The lane position (from rightmost to leftmost) and whether the lane direction is the same or opposite direction as the driving car, and (b) additional pixel maps can be employed for intersections, signage and buildings. Similarly, the information can be drawn as shapes with a particular value.
- extra sensor data is represented in extra pixel maps by setting the entire pixel map to a certain value.
- training data is labeled semi-automatically by (i) for each label, defining a set of constraints on the sensory variables and encode them into a set of computer rules and (ii) sensor data frames for which all rules are verified are assigned a corresponding label.
- FIG. 3 is a block/flow diagram of a user interface (UI) of a vehicle illustrating a dangerous situation on the road ahead, in accordance with embodiments of the present invention.
- UI user interface
- System 10 includes a first vehicle 12 (vehicle A).
- vehicle 12 is driving down a road 14 .
- vehicle 12 encounters a second vehicle 16 (vehicle B), a third vehicle 18 (vehicle C), and a fourth vehicle 19 (vehicle D).
- the second and third vehicles 16 , 18 are driving within lane 15
- the fourth vehicle 19 is driving in lane 13 .
- User interface control panel 60 of the vehicle 12 indicates a special situation (e.g., a dangerous situation) to the driver of the vehicle 12 .
- a traffic scene TS
- the driver of vehicle 16 may have made a sudden move to pass vehicle 18 , even though vehicle 19 is basically adjacent to vehicle 18 . This could lead to a dangerous situation, such as an accident caused by the second vehicle 16 .
- the driver of vehicle 12 can view this special situation on a map 70 positioned somewhere within the vehicle 12 .
- the user interface control panel 60 can display a plurality of different data/information to the driver of the vehicle 12 .
- a speed 90 of vehicle A can be displayed relative to a speed 92 of vehicle B and a speed 94 of the vehicle C.
- the user interface control panel 60 can further display a plurality of indicators 80 .
- the indicators 80 can be warning indicators, notifications indicators, alarm indicators, etc.
- the indicators 80 can be light emitting diodes (LEDs).
- the indicators 80 can display text.
- One skilled in the art can contemplate a plurality of different types of indicators.
- the vehicle 12 can receive a plurality of inputs.
- the first input 20 can include sensor inputs, such as vehicle sensors 22 , speed/velocity sensors 24 , and/or position sensors 26 .
- other sensors can also be included for detecting/monitoring a plurality of other variables/parameters.
- the vehicle 12 can receive second inputs 30 , including, but not limited to, global data 32 , weather information 34 , traffic information 36 , and/or global positioning system (GPS) data.
- GPS global positioning system
- Data/information can also be collected from one or more video cameras mounted on or within the vehicle 12 or from one or more video cameras in a vicinity of the vehicle 12 (e.g., within the immediate environment).
- the vehicle 12 can further receive data/information from a plurality of networks.
- the vehicle 12 can receive data from a first network 40 (e.g., Internet) and a second network 50 (e.g., a deep convolutional neural network).
- a first network 40 e.g., Internet
- a second network 50 e.g., a deep convolutional neural network
- One skilled in the art can contemplate a plurality of other networks for communicating with the vehicle 12 .
- FIG. 4 is a block/flow diagram of an interior of an automobile depicting a navigation device mounted on a dashboard or incorporated within a dashboard and/or console area, in accordance with embodiments of the present invention.
- the car interior 1000 depicts a navigation device 1010 mounted around or in proximity to a windshield area of the car.
- the car interior 1000 also depicts a navigation device 1020 incorporated within one or more console areas of the car. It is contemplated that the navigation devices 1010 , 1020 can be incorporated or positioned on any interior portion of the car interior 1000 .
- FIG. 5 is a block/flow diagram of the navigation device of FIG. 4 including a video camera for predicting driving situations while a user drives the car, in accordance with embodiments of the present invention.
- the navigation devices 1010 , 1020 can include a housing 1002 having a display screen 1004 .
- the display screen 1004 can be, e.g., a light emitting diode (LED) display or an organic light emitting diode (OLED) display.
- the navigation devices 1010 , 1020 can further include a plurality of control buttons 1006 .
- the navigation devices 1010 , 1020 can further include a plurality of indicators 1008 .
- a video camera 1101 and a global positioning system (GPS) 1102 can be incorporated with the navigations devices 1010 , 1020 .
- GPS global positioning system
- the navigation devices 1010 , 1020 can be portable and compatible with a number of different vehicles.
- Portable devices brought into a vehicle by, for example, a guest, who may not be the primary user of the vehicle may also wirelessly communicate with the car system.
- Car systems may wirelessly communicate with the portable device by way of, e.g., a Bluetooth communication link.
- Conventional protocols for establishing a communication link via Bluetooth may first pair the system with the portable device by using a pin code to authenticate the portable device requesting the wireless connection. Once the pin code is authenticated and a media access control (MAC) address is saved by both the car system and the portable device, pairing is complete and the car system and the portable device can automatically establish wireless communication when the two are within communication range (i.e., the portable device is brought into the car).
- MAC media access control
- the navigation devices 1010 , 1020 can be attachable/detachable to different parts of the car.
- the navigation devices 1010 , 1020 can be attached to various portions of the center console or dashboard.
- the navigation devices 1010 , 1020 can operate with a car power system.
- the car power supply system provides i) power to crank start an engine through a starter motor and ii) continuous supply to vehicle systems either from a battery or from an alternator driven by the engine. This keeps the vehicle functioning through various controls and helps provide electrical command signals to respond to driver inputs and to satisfy passenger needs during usage of the vehicle.
- the power supply is of low voltage DC and is provided by a combination of a rechargeable lead acid battery and an alternator with an in-built rectifier and voltage regulator. The nominal voltage of the supply is normally 12 volts DC in cars and 24 volts on commercial vehicles.
- the positive terminals of the battery and of the alternator are joined/connected to form a power supply positive terminal bus for the vehicle and power can be drawn through the bus from both sources or from either one of the sources.
- the power is distributed to various loads picking off power from the bus.
- the negative terminal of the supply system is connected to the body/chassis of the vehicle and the return lines from the loads placed at different locations on the vehicle are terminated at grounding studs welded/fitted at suitable locations on the vehicle body or chassis.
- the return currents from the different loads are returned to the negative terminals of the battery and the alternator through the body/chassis of the vehicle.
- the navigation devices 1010 , 1020 can receive multiple inputs from the car power system described to monitor various components of the car power system.
- the navigation devices 1010 , 1020 can operate with any other components of the car.
- the navigation devices 1010 , 1020 can receive input from any of the car components.
- any car cameras or sensors or GPS units can provide their data/information to the navigation devices 1010 , 1020 .
- braking system components, engine components, engine cooling system components, engine oil system components, fuel supply components, suspension and steering component systems, and/or transmission system components can be configured to electrically communicate with the navigation devices 1010 , 1020 for maintaining optimal performance during detection of a special situation.
- the navigation devices 1010 , 1020 can operate with a controller employed to control the activation of a countermeasure system.
- Each countermeasure may have an individual actuator associated therewith.
- the controller may direct the individual countermeasure actuator to activate the countermeasure.
- Various types of countermeasure systems will be evident to those skilled in the art. Examples of a countermeasure within countermeasure system include occupant belt pre-tensioning, bumper height changing including nose dipping, braking, the pre-arming of internal airbags, the deployment of exterior or internal airbags, pedal control, steering column position, head restraint and knee bolster control.
- the controller is programmed to activate the appropriate countermeasure in response to the inputs from the various sensors or from the navigation devices 1010 , 1020 .
- the navigation devices 1010 , 1020 can operate with braking system conditions sensors.
- Braking system conditions may be a sensor or sensors such as a brake pedal position sensor or a brake pressure monitor.
- the brake system conditions may be used to determine occupant driving characteristics and therefore provide an improved collision prediction and as a result provide an improved countermeasure deployment decision.
- Exemplary embodiments of the present invention can be incorporated within a vehicle, such as an automobile, motorcycle, etc.
- the exemplary embodiments of the present invention can be represented or displayed on a user interface (UI) within a vehicle.
- the UI can provide data/information to the driver of the vehicle in real-time and on a continuous or periodic basis.
- the data/information can be related to a plurality of different variables/parameters detected and/or monitored, either within the immediate vicinity or environment of the vehicle or from external sources.
- traffic scenes (TS s) are recognized and compared to predetermined or predefined TSs in order to predict dangerous situations ahead of time.
- the data/information is represented as feature maps, that is, two-dimensional pixel maps to be fed directly into a deep convolutional neural network including trained data/information.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Health & Medical Sciences (AREA)
- Multimedia (AREA)
- General Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computing Systems (AREA)
- Data Mining & Analysis (AREA)
- Software Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Molecular Biology (AREA)
- Biomedical Technology (AREA)
- General Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- Medical Informatics (AREA)
- Databases & Information Systems (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Biodiversity & Conservation Biology (AREA)
- Mechanical Engineering (AREA)
- Human Computer Interaction (AREA)
- Traffic Control Systems (AREA)
Abstract
Description
- This application claims priority to Provisional Application No. 62/318,852, filed on Apr. 6, 2016, incorporated herein by reference in its entirety. Moreover, the present disclosure is related to the subject matter of commonly assigned, co-pending U.S. application Ser. No. TBD (Atty. Doc. No. 15085B).
- The present invention relates to neural networks and, more particularly, to a multi-modal driving danger prediction system for automobiles using deep convolutional neural networks.
- In this age of increasing driving distractions, drivers are ever more likely to fail to recognize potential danger in complex driving situations. Collision avoidance systems are becoming common in cars. These systems apply the brakes when the driver fails to notice obstacles at close range. These systems operate in a short time range and are mostly reactive. However, these systems do not predict danger on a longer time horizon, nor focus on the driver's attentiveness.
- A computer-implemented method for training a deep neural network to recognize traffic scenes (TSs) from multi-modal sensors and knowledge data is presented. The method includes receiving data from the multi-modal sensors and the knowledge data and extracting feature maps from the multi-modal sensors and the knowledge data by using a traffic participant (TS) extractor to generate a first set of data, using a static objects extractor to generate a second set of data, and using an additional information extractor. The computer-implemented method further includes training the deep neural network, with training data to recognize the TSs from a viewpoint of a vehicle.
- A system for training a deep neural network to recognize traffic scenes (TS s) from multi-modal sensors and knowledge data is presented. The system includes a memory and a processor in communication with the memory, wherein the processor is configured to receive data from the multi-modal sensors and the knowledge data and extract feature maps from the multi-modal sensors and the knowledge data by using a traffic participant (TS) extractor to generate a first set of data, using a static objects extractor to generate a second set of data, and using an additional information extractor. The computer-implemented system further includes training the deep neural network, with training data to recognize the TSs from a viewpoint of a vehicle.
- A non-transitory computer-readable storage medium comprising a computer-readable program for training a deep neural network to recognize traffic scenes (TSs) from multi-modal sensors and knowledge data is presented, wherein the computer-readable program when executed on a computer causes the computer to perform the steps of receiving data from the multi-modal sensors and the knowledge data and extracting feature maps from the multi-modal sensors and the knowledge data by using a traffic participant (TS) extractor to generate a first set of data, using a static objects extractor to generate a second set of data, and using an additional information extractor. The computer-implemented method further includes training the deep neural network, with training data to recognize the TSs from a viewpoint of a vehicle.
- These and other features and advantages will become apparent from the following detailed description of illustrative embodiments thereof, which is to be read in connection with the accompanying drawings.
- The disclosure will provide details in the following description of preferred embodiments with reference to the following figures wherein:
-
FIG. 1 is a block/flow diagram of a method for predicting dangerous situations ahead of time, in accordance with embodiments of the present invention; -
FIG. 2 is a block/flow diagram of an architecture of a deep convolutional neural network, in accordance with embodiments of the present invention; and -
FIG. 3 is a block/flow diagram of a user interface (UI) of a vehicle illustrating a dangerous situation on the road ahead, in accordance with embodiments of the present invention; -
FIG. 4 is a block/flow diagram of an interior of an automobile depicting a navigation device mounted on a windshield or incorporated within a dashboard and/or console area, in accordance with embodiments of the present invention; and -
FIG. 5 is a block/flow diagram of the navigation device ofFIG. 4 including a video camera for predicting one or more dangerous driving situations while a user drives the car, in accordance with embodiments of the present invention. - The availability of on-board electronics and in-vehicle information systems has demanded the development of more intelligent vehicles. One such important intelligence is the possibility to evaluate the driving danger level to prevent potential driving risks.
- The exemplary embodiments of the present invention introduce driver assistance systems which support the driver of a motor vehicle in driving the vehicle and/or warn him/her of special dangerous situations, for example, an imminent collision, or automatically initiate measures for avoiding the collision or for mitigating the consequences of the collision. Such determinations are based on a position finding system, in particular a radar or light detection and ranging (lidar) system, locating objects in the surroundings of the vehicle, such as other vehicles. The distances and relative velocities or speeds of the objects located in the surroundings of the vehicle are measured accurately by employing such position finding system to allow for a prediction of a special situation ahead.
- The present invention provides the advantage that it makes detection of such special situations possible, so that appropriate countermeasures may be initiated or at least a warning may be output to the driver. The present invention is based on the principle that, while locating a single stationary object on the road is insufficient for inferring a potentially special (dangerous) situation, such situations may very well be detected when a set of a plurality of objects is considered together, and a characteristic signature is recognizable from the distribution of the position finding data such as, but not limited to, distances, relative velocities, relative speeds, and/or relative accelerations, which indicate a certain class of special situations.
- The exemplary embodiments of the present invention relate to a driving situation predicting device having a video capturing unit for capturing real-time video of a driver driving a car and for making predictions on whether a special driving situation is imminent. If a special driving situation is imminent, the driver is notified with a warning.
- Referring to
FIG. 1 , traffic scenes (TSs) are described by traffic participants (TP) (cars, trucks, bikes, pedestrians, etc.), geometry of the road (intersection, lanes, pedestrian crossing, signals, signs, etc.), information about surroundings (schools, train stations, bus stops, temperature, weather, traffic, etc.) and a status of the driving car (speed, driver status, car occupancy, radio, phone usage, etc.). The TSs can be captured by a plurality of sensors and video devices, such as video cameras. The TSs can be captured continuously and in real-time. - The TSs can be traffic environments pertaining to: peripheral road information related to a road located around a vehicle on which a dangerous drive predicting apparatus has been mounted; signal mechanism information related to a signal mechanism; peripheral vehicle information related to other vehicles which are driven around a vehicle; and parking lot information for indicating whether or not a parking lot is located around the vehicles and the like. The above-described peripheral road information corresponds to information for indicating whether a road along which the vehicle is driven is a single lane, a crossing, or a priority-sided road. The above-described signal mechanism information corresponds to information for indicating whether or not a signal mechanism is present, information for indicating whether or not the vehicle approaches a signal mechanism, and information related to signal light turning situations of the signal mechanism. The above-described peripheral vehicle information corresponds to information for indicating whether or not a preceding drive vehicle is present, information for indicating whether or not a succeeding drive vehicle is present, and information for indicating whether or not an intersecting vehicle is present.
- The judgement as to whether or not the vehicle is being driven on a single lane may be made by judging, for instance, information held in a car navigation system to be mounted on the vehicle, or by detecting a lane line by operating a camera to be mounted on the vehicle so as to judge a curved degree of the lane line.
- Also, the judgement as to whether or not the vehicle is being driven on a crossing may be made by judging, for instance, information held in a car navigation system to be mounted on the vehicle, or by detecting a signal mechanism, or a stop line by operating the camera to be mounted on the vehicle so as to measure a distance.
- A driving danger prediction system is realized by continuously matching the current TS to a codebook of TSs or predetermined TSs that have been identified as leading to special situations, such as, but not limited to dangerous situations for a driver of a
vehicle 12. When a match occurs, a warning can be transmitted to the driver of the vehicle 12 (FIG. 3 ). A traffic environment detecting unit (navigation units presented herein) can detect a shape of a road around the vehicle, into which the vehicle tries to enter; a condition whether or not a signal mechanism is present; a signal light turning situation; a condition whether or not a preceding drive vehicle is present; and the like. A driving characteristic determining unit (incorporated with the navigation units herein) can judge a traffic environment based upon the detected information, namely judges whether or not the traffic environment corresponds to such a previously determined traffic environment to which a dangerous driving situation has been assigned to. - Moreover, embodiments of the present invention describe how to fit an end-to-end convolutional neural network learning approach (convnet) to the challenge of matching TSs.
-
FIG. 1 depicts the flowchart of the system. Data arrives into the system from at least two sources. One source issensor frame data 100 obtained dynamically at regular synchronized sampling intervals from various sensors throughout the car (e.g., radar, controller area network (CAN), video, global positioning system (GPS), etc.), and the second source ofdata 110 is from non-sampled sources such as the GPS street map or externally generated information such as weather, traffic info, etc. The TSs can be evaluated for captured content, such as cars, trucks, bikes, pedestrians, intersection, lanes, pedestrian crossing, signals, signs, schools, train stations, bus stops, etc. All this information is processed to create a number of feature maps. -
Feature maps training module 610. - At least three types of feature map extractors can be employed. The
TP extractor 200, thestatic objects extractor 210, and theadditional info extractor 220. The TPfeature map extractor 200 uses sensors to extract position, size, type and velocities/speeds of traffic participants. Specialized automotive radar units are capable of extracting positions and velocities of several TPs simultaneously. Light detection and ranging (LIDAR) units are more precise and can sense the size of TPs more accurately. Video can be employed along with advanced processing techniques such as structure from motion (SFM) to obtain position, size and even type of TPs. When size and type cannot be obtained from one of the sensors, standard default values are utilized. - The radar system can detect the presence of an object (e.g., car) in a detection zone, obtain its distance and relative velocity. The camera system alone can also be used to detect the presence of the object (e.g., car) in its detection zone, obtain its distance, relative velocity and size information. Alternately, the radar system can be used to detect the presence of the object, its distance and relative velocity and the vision system can be used to confirm the above items and also to provide additional information about the size of the object. For pre-crash sensing applications, it is permissible to have both radar and vision based systems to ensure good performance under all weather conditions and also to provide redundancy for improved reliability.
- The feature maps 300 are built from the
TP extractor 200. The first feature map encodes position, size and type of each TP with oriented filled rectangles. The orientation is defined by a velocity vector and the type is encoded by filling a rectangle with a particular value. The second and third feature maps are velocity feature maps. The same approach as above is employed, except that normalized relative velocity amplitudes (forward and lateral, respectively) are used as fill values for the rectangles. - From the
static object extractor 210, several feature maps may be built or generated depending on how rich the available map information is. First is the lane-accurate street feature map. The GPS sensor is employed to obtain an absolute position of the car in world coordinates and query a portion (e.g., a box) of a street map around those coordinates. Such queries are made either to a GPS map unit in the car or to a cloud server via, e.g., wireless connection. Depending on accuracy of the street map, detailed lane information may not be available (e.g., multi-lane roadways may be returned as single lines). In that case, video or dedicated lane-following units may provide detailed lane configurations. Otherwise, standard lane widths are used as an approximation. Each individual road lane is then drawn to the feature map using a value reflecting a type of road it is (e.g., highway, rural, etc.) and a direction of a lane (e.g., same as driving vehicle, opposite). Other feature maps may be generated from building, signage, tree, fences and other static objects. Bridge or tunnel segments may also be encoded to create separate feature maps. - The third type of
feature map extractor 220 creates or generates feature maps including asingle value 320. Such feature maps allow certain global parameters of a TS to be combined with a dynamic TP and static feature maps. One such parameter is a speed of a driving car. Such parameter is not included in the TP feature maps as all speeds are relative to the driving car and hence need to be inputted separately. Other global parameters such as weather (e.g., rainy, snowy, foggy, dark, etc.) can be inputted into those feature maps. - Further, other parameters detected by a vehicle information detecting unit are vehicle speed information, acceleration information, steering angle information, vehicle position information, and vehicle-to-vehicle distance information, which appear as a result obtained when an accelerator, a brake pedal, and a steering wheel, which are manipulated by a vehicle driver, control a vehicle. The vehicle speed information is related to a speed of the vehicle; the acceleration information is related to acceleration of the vehicle; the steering angle information is related to a steering angle thereof; the vehicle position information is related to a position of the vehicle, and the vehicle-to-vehicle distance information is related to a vehicle-to-vehicle distance. The drive information is such an information containing vehicle information.
- The number and size of the feature maps dictates a complexity of the convnet model.
FIG. 2 shows an example of a possible convnet with, e.g., five (5) feature maps of, e.g.,size 20×100 pixels. An output layer of the convnet has as many outputs as a number of traffic scenes (TS) to be recognized. - In particular, a deep convolutional network (convnet) provides for a set of initial feature maps. These feature maps can be, e.g., a
position feature map 701, a speed feature map 703 (in a direction x), a speed feature map 705 (in a direction y), aroad feature map 707, and a globalknowledge feature map 709. A convolution operation takes place to generate a second layer of feature maps 711. A sub-sampling operation then takes place to generate a third layer of feature maps 713. Another convolution operation takes place to generate a fourth layer of feature maps 715. Another subsampling operation takes place to generate a fifth layer of feature maps 717. The data is then reshaped 719 and fully connected 721. The fully connectedlayer 721 uses asoftmax activation function 723 in the output layer. - Referring back to
FIG. 1 , as with all supervised learning approaches, a training set of labeled examples 500 is created. When employing a deep convnet, a sufficient number of training examples 500 is employed to train the deep convnet, so that it can generalize adequately. Hence, the exemplary embodiments of the present invention also describe an approach to generate large amounts of labeledsensor frame data 130 from recorded naturalistic driving datasets. - Initially, a set of constraints on sensor data variables is defined for all TSs that are to be detected. For example, a set of constraints for a particular TS would be: the driving vehicle's speed should be above 100 km/h, there should be a car in front at less than 30 m and that car's relative speed should be negative. That kind of description is a natural way to define a particular TS, yet can also be easily transcribed into computer code, resulting in a list of coded
rules 120 for each TS. For each data frame, the variables of corresponding sensors are fed to therule verification module 400. When all rules describing a TS are verified, thedata frame 500 is labeled as belonging to that TS, otherwise a negative label is given to the data frame. - Using the input feature maps along with their corresponding label, a
convnet model 600 can be trained by employing machine learning techniques. - Since, during normal driving, most of the time, no TS occur, negative examples need to be introduced into the training set. The number of negative examples is a parameter of the system.
- A method and system for training the deep convnet to recognize complex traffic scenes from multi-modal sensors and knowledge data is further provided. The methods and systems further use a common representation to combine sensors of different modalities. The methods and systems use a two dimensional top-view representation of the traffic scene (TS) suitable for direct input to the deep convnet via a set of two dimensional pixel maps where the origin is set by a position of the driving car.
- Radar sensor data is represented in three pixel maps by: (i) drawing oriented filled rectangles at positions and sizes corresponding to detected radar objects, at orientations based on a direction of their velocity, and filling pixels with a particular value, (ii) for the first pixel map, a value represents a type of object, (iii) for a second pixel map, a value represents an amplitude of a forward relative velocity, and (iv) for a third pixel map, a value represents an amplitude of a lateral relative velocity.
- The methods and systems further introduce a GPS sensor for providing an instantaneous absolute location of the driving car on a street map that is represented by a set of pixel maps as follows: (a) a first pixel map is for lane-level detail around the driving car and is obtained by querying a map subsystem, (i) for systems that don't have lane-level detail, lane information can be obtained by video processing, or by using default lane width values for each type of roadway, and (ii) lanes are drawn onto a pixel map with values corresponding to a type of roadway. The lane position (from rightmost to leftmost) and whether the lane direction is the same or opposite direction as the driving car, and (b) additional pixel maps can be employed for intersections, signage and buildings. Similarly, the information can be drawn as shapes with a particular value.
- Moreover, extra sensor data is represented in extra pixel maps by setting the entire pixel map to a certain value. Additionally, training data is labeled semi-automatically by (i) for each label, defining a set of constraints on the sensory variables and encode them into a set of computer rules and (ii) sensor data frames for which all rules are verified are assigned a corresponding label.
-
FIG. 3 is a block/flow diagram of a user interface (UI) of a vehicle illustrating a dangerous situation on the road ahead, in accordance with embodiments of the present invention. -
System 10 includes a first vehicle 12 (vehicle A). Thevehicle 12 is driving down aroad 14. On theroad 14,vehicle 12 encounters a second vehicle 16 (vehicle B), a third vehicle 18 (vehicle C), and a fourth vehicle 19 (vehicle D). The second andthird vehicles lane 15, whereas thefourth vehicle 19 is driving inlane 13. Userinterface control panel 60 of thevehicle 12 indicates a special situation (e.g., a dangerous situation) to the driver of thevehicle 12. For example, it is detected from a traffic scene (TS) that thesecond vehicle 16 is attempting to pass thethird vehicle 18 vialane 13 where thefourth vehicle 19 is also detected close by. The driver ofvehicle 16 may have made a sudden move to passvehicle 18, even thoughvehicle 19 is basically adjacent tovehicle 18. This could lead to a dangerous situation, such as an accident caused by thesecond vehicle 16. - The driver of
vehicle 12 can view this special situation on a map 70 positioned somewhere within thevehicle 12. The userinterface control panel 60 can display a plurality of different data/information to the driver of thevehicle 12. For example, a speed 90 of vehicle A can be displayed relative to aspeed 92 of vehicle B and aspeed 94 of the vehicle C. Of course, one skilled in the art can contemplate displaying a plurality of other information to the user (e.g., position information related to each vehicle A, B, C, D, etc.). The userinterface control panel 60 can further display a plurality ofindicators 80. Theindicators 80 can be warning indicators, notifications indicators, alarm indicators, etc. Theindicators 80 can be light emitting diodes (LEDs). Theindicators 80 can display text. One skilled in the art can contemplate a plurality of different types of indicators. - The
vehicle 12 can receive a plurality of inputs. For example, thefirst input 20 can include sensor inputs, such asvehicle sensors 22, speed/velocity sensors 24, and/orposition sensors 26. Of course, other sensors can also be included for detecting/monitoring a plurality of other variables/parameters. Moreover, thevehicle 12 can receive second inputs 30, including, but not limited to,global data 32, weather information 34, traffic information 36, and/or global positioning system (GPS) data. Data/information can also be collected from one or more video cameras mounted on or within thevehicle 12 or from one or more video cameras in a vicinity of the vehicle 12 (e.g., within the immediate environment). - The
vehicle 12 can further receive data/information from a plurality of networks. For example, thevehicle 12 can receive data from a first network 40 (e.g., Internet) and a second network 50 (e.g., a deep convolutional neural network). One skilled in the art can contemplate a plurality of other networks for communicating with thevehicle 12. -
FIG. 4 is a block/flow diagram of an interior of an automobile depicting a navigation device mounted on a dashboard or incorporated within a dashboard and/or console area, in accordance with embodiments of the present invention. - The
car interior 1000 depicts anavigation device 1010 mounted around or in proximity to a windshield area of the car. Thecar interior 1000 also depicts anavigation device 1020 incorporated within one or more console areas of the car. It is contemplated that thenavigation devices car interior 1000. -
FIG. 5 is a block/flow diagram of the navigation device ofFIG. 4 including a video camera for predicting driving situations while a user drives the car, in accordance with embodiments of the present invention. - The
navigation devices housing 1002 having adisplay screen 1004. Thedisplay screen 1004 can be, e.g., a light emitting diode (LED) display or an organic light emitting diode (OLED) display. Thenavigation devices control buttons 1006. Thenavigation devices indicators 1008. Additionally, avideo camera 1101 and a global positioning system (GPS) 1102 can be incorporated with thenavigations devices - The
navigation devices - The
navigation devices navigation devices - The
navigation devices - The positive terminals of the battery and of the alternator are joined/connected to form a power supply positive terminal bus for the vehicle and power can be drawn through the bus from both sources or from either one of the sources. The power is distributed to various loads picking off power from the bus. The negative terminal of the supply system is connected to the body/chassis of the vehicle and the return lines from the loads placed at different locations on the vehicle are terminated at grounding studs welded/fitted at suitable locations on the vehicle body or chassis. The return currents from the different loads are returned to the negative terminals of the battery and the alternator through the body/chassis of the vehicle. The
navigation devices - The
navigation devices navigation devices navigation devices navigation devices - The
navigation devices navigation devices - The
navigation devices - Exemplary embodiments of the present invention can be incorporated within a vehicle, such as an automobile, motorcycle, etc. The exemplary embodiments of the present invention can be represented or displayed on a user interface (UI) within a vehicle. The UI can provide data/information to the driver of the vehicle in real-time and on a continuous or periodic basis. The data/information can be related to a plurality of different variables/parameters detected and/or monitored, either within the immediate vicinity or environment of the vehicle or from external sources. In one aspect, traffic scenes (TS s) are recognized and compared to predetermined or predefined TSs in order to predict dangerous situations ahead of time. In another aspect, the data/information is represented as feature maps, that is, two-dimensional pixel maps to be fed directly into a deep convolutional neural network including trained data/information.
- The foregoing is to be understood as being in every respect illustrative and exemplary, but not restrictive, and the scope of the invention disclosed herein is not to be determined from the Detailed Description, but rather from the claims as interpreted according to the full breadth permitted by the patent laws. It is to be understood that the embodiments shown and described herein are only illustrative of the principles of the present invention and that those skilled in the art may implement various modifications without departing from the scope and spirit of the invention. Those skilled in the art could implement various other feature combinations without departing from the scope and spirit of the invention. Having thus described aspects of the invention, with the details and particularity required by the patent laws, what is claimed and desired protected by Letters Patent is set forth in the appended claims.
Claims (20)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/478,823 US20170293837A1 (en) | 2016-04-06 | 2017-04-04 | Multi-Modal Driving Danger Prediction System for Automobiles |
PCT/US2017/026362 WO2017177005A1 (en) | 2016-04-06 | 2017-04-06 | Multi-modal driving danger prediction system for automobiles |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201662318852P | 2016-04-06 | 2016-04-06 | |
US15/478,823 US20170293837A1 (en) | 2016-04-06 | 2017-04-04 | Multi-Modal Driving Danger Prediction System for Automobiles |
Publications (1)
Publication Number | Publication Date |
---|---|
US20170293837A1 true US20170293837A1 (en) | 2017-10-12 |
Family
ID=59998746
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/478,823 Abandoned US20170293837A1 (en) | 2016-04-06 | 2017-04-04 | Multi-Modal Driving Danger Prediction System for Automobiles |
US15/478,886 Active 2037-07-08 US10296796B2 (en) | 2016-04-06 | 2017-04-04 | Video capturing device for predicting special driving situations |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/478,886 Active 2037-07-08 US10296796B2 (en) | 2016-04-06 | 2017-04-04 | Video capturing device for predicting special driving situations |
Country Status (2)
Country | Link |
---|---|
US (2) | US20170293837A1 (en) |
WO (2) | WO2017177008A1 (en) |
Cited By (31)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE102017127600A1 (en) | 2017-11-22 | 2019-05-23 | Connaught Electronics Ltd. | Method for automatic parking for a motor vehicle with a driving support system |
DE102017128082A1 (en) | 2017-11-28 | 2019-05-29 | Connaught Electronics Ltd. | Meta-architecture design for a CNN network |
SE1851125A1 (en) * | 2018-09-21 | 2019-06-17 | Scania Cv Ab | Method and control arrangement for machine learning of a model-based vehicle application in a vehicle |
CN110008978A (en) * | 2017-12-07 | 2019-07-12 | 亿目朗欧洲股份有限公司 | Classification of risks training method, classification of risks method, auxiliary or Automated Vehicle Operation system |
US10395144B2 (en) * | 2017-07-24 | 2019-08-27 | GM Global Technology Operations LLC | Deeply integrated fusion architecture for automated driving systems |
CN110234018A (en) * | 2019-07-09 | 2019-09-13 | 腾讯科技(深圳)有限公司 | Multimedia content description generation method, training method, device, equipment and medium |
DE102018206208A1 (en) * | 2018-04-23 | 2019-10-24 | Robert Bosch Gmbh | Method, device, product and computer program for operating a technical system |
CN110696835A (en) * | 2019-10-11 | 2020-01-17 | 深圳职业技术学院 | Automatic early warning method and automatic early warning system for dangerous driving behaviors of vehicle |
US10599146B2 (en) | 2018-03-26 | 2020-03-24 | Ford Global Technologies, Llc | Action-conditioned vehicle control |
CN112166438A (en) * | 2018-03-13 | 2021-01-01 | 雷哥尼公司 | Deterministic labeled data generation and artificial intelligence training pathways |
US10943154B2 (en) * | 2019-01-22 | 2021-03-09 | Honda Motor Co., Ltd. | Systems for modeling uncertainty in multi-modal retrieval and methods thereof |
CN112558510A (en) * | 2020-10-20 | 2021-03-26 | 山东亦贝数据技术有限公司 | Intelligent networking automobile safety early warning system and early warning method |
CN112638737A (en) * | 2018-10-16 | 2021-04-09 | 丰田自动车工程及制造北美公司 | Vehicle speed predictor using neural networks based on V2X data augmentation for predictive optimal control of networked and automated vehicles |
CN112825136A (en) * | 2019-11-21 | 2021-05-21 | 辉达公司 | Deep neural network for detecting obstacles using RADAR sensors in autonomous machine applications |
CN112937441A (en) * | 2021-03-29 | 2021-06-11 | 陕西科技大学 | Automobile anti-collision system and control method |
US20210326699A1 (en) * | 2020-04-21 | 2021-10-21 | Inrix, Inc. | Travel speed prediction |
US11194331B2 (en) | 2018-10-30 | 2021-12-07 | The Regents Of The University Of Michigan | Unsupervised classification of encountering scenarios using connected vehicle datasets |
CN113792598A (en) * | 2021-08-10 | 2021-12-14 | 西安电子科技大学广州研究院 | Vehicle-mounted camera-based vehicle collision prediction system and method |
US11341398B2 (en) * | 2016-10-03 | 2022-05-24 | Hitachi, Ltd. | Recognition apparatus and learning system using neural networks |
CN114881078A (en) * | 2022-05-07 | 2022-08-09 | 安徽蔚来智驾科技有限公司 | Method and system for filtering data under predetermined scenarios |
US11410327B2 (en) * | 2018-03-02 | 2022-08-09 | Nippon Telegraph And Telephone Corporation | Location determination apparatus, location determination method and computer program |
US11558584B2 (en) * | 2019-07-11 | 2023-01-17 | Chris Pritchard | Systems and methods for providing real-time surveillance in automobiles |
US11676488B2 (en) | 2019-10-11 | 2023-06-13 | Aptiv Technologies Limited | Method and system for determining an attribute of an object at a pre-determined time point |
US20230322248A1 (en) * | 2022-04-06 | 2023-10-12 | Gm Global Technology Operation Llc | Collision warning system for a motor vehicle having an augmented reality head up display |
WO2023246389A1 (en) * | 2022-06-20 | 2023-12-28 | 支付宝(杭州)信息技术有限公司 | Fusion training based on knowledge representation learning for risk control model |
US11941509B2 (en) | 2020-02-27 | 2024-03-26 | Aptiv Technologies AG | Method and system for determining information on an expected trajectory of an object |
US12050285B2 (en) | 2019-11-21 | 2024-07-30 | Nvidia Corporation | Deep neural network for detecting obstacle instances using radar sensors in autonomous machine applications |
US12072443B2 (en) | 2019-11-15 | 2024-08-27 | Nvidia Corporation | Segmentation of lidar range images |
US12080078B2 (en) | 2019-11-15 | 2024-09-03 | Nvidia Corporation | Multi-view deep neural network for LiDAR perception |
US12111386B2 (en) | 2020-07-24 | 2024-10-08 | Aptiv Technologies AG | Methods and systems for predicting a trajectory of an object |
US20250013672A1 (en) * | 2023-07-03 | 2025-01-09 | Red Atlas Inc. | Systems and methods for region-based segmentation of a knowledge base developed using data collected from myriad sources |
Families Citing this family (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9373059B1 (en) * | 2014-05-05 | 2016-06-21 | Atomwise Inc. | Systems and methods for applying a convolutional network to spatial data |
US10061909B2 (en) * | 2016-12-29 | 2018-08-28 | Qualcomm Incorporated | Device authentication based on behavior classification using convolution neural network |
US10546237B2 (en) | 2017-03-30 | 2020-01-28 | Atomwise Inc. | Systems and methods for correcting error in a first classifier by evaluating classifier output in parallel |
US10595039B2 (en) | 2017-03-31 | 2020-03-17 | Nvidia Corporation | System and method for content and motion controlled action video generation |
CN109688351B (en) | 2017-10-13 | 2020-12-15 | 华为技术有限公司 | An image signal processing method, device and equipment |
CN107945534A (en) * | 2017-12-13 | 2018-04-20 | 浙江大学城市学院 | A kind of special bus method for predicting based on GMDH neutral nets |
US11206375B2 (en) | 2018-03-28 | 2021-12-21 | Gal Zuckerman | Analyzing past events by utilizing imagery data captured by a plurality of on-road vehicles |
CN108846519B (en) * | 2018-06-14 | 2021-09-03 | 大唐高鸿信息通信(义乌)有限公司 | Safe driving K nearest neighbor prediction method based on vehicle-mounted short-distance communication network |
CN109063751B (en) * | 2018-07-16 | 2021-09-17 | 江苏智通交通科技有限公司 | Traffic high-risk personnel identification method based on gradient lifting decision tree algorithm |
US11138418B2 (en) | 2018-08-06 | 2021-10-05 | Gal Zuckerman | Systems and methods for tracking persons by utilizing imagery data captured by on-road vehicles |
CN109344804A (en) * | 2018-10-30 | 2019-02-15 | 百度在线网络技术(北京)有限公司 | A kind of recognition methods of laser point cloud data, device, equipment and medium |
CN110126846B (en) * | 2019-05-24 | 2021-07-23 | 北京百度网讯科技有限公司 | Representation method, device, system and storage medium of driving scene |
CN110386144B (en) * | 2019-06-19 | 2020-09-08 | 长安大学 | GHMM/GGAP-RBF hybrid model for identifying driver braking intention and identification method |
CN110826520B (en) * | 2019-11-14 | 2023-05-26 | 燕山大学 | A Port Grab Detection Method Based on Improved YOLOv3-tiny Algorithm |
CN111016901A (en) * | 2019-12-30 | 2020-04-17 | 苏州安智汽车零部件有限公司 | Intelligent driving decision method and system based on deep learning |
CN111242015B (en) * | 2020-01-10 | 2023-05-02 | 同济大学 | A Method of Predicting Dangerous Driving Scenes Based on Motion Contour Semantic Map |
CN111239728B (en) * | 2020-02-26 | 2021-06-15 | 深圳雷研技术有限公司 | Passenger counting method and system based on millimeter wave radar |
CN112026669A (en) * | 2020-09-03 | 2020-12-04 | 杭州麦扑文化创意有限公司 | Map drawing program capable of quickly integrating travel information |
Family Cites Families (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0785280B2 (en) * | 1992-08-04 | 1995-09-13 | タカタ株式会社 | Collision prediction judgment system by neural network |
US9007197B2 (en) * | 2002-05-20 | 2015-04-14 | Intelligent Technologies International, Inc. | Vehicular anticipatory sensor system |
US7839292B2 (en) | 2007-04-11 | 2010-11-23 | Nec Laboratories America, Inc. | Real-time driving danger level prediction |
KR101030210B1 (en) * | 2009-12-03 | 2011-04-22 | 쓰리에이치비젼주식회사 | Vehicle obstacle recognition system and method |
US8700552B2 (en) * | 2011-11-28 | 2014-04-15 | Microsoft Corporation | Exploiting sparseness in training deep neural networks |
EP2615598B1 (en) * | 2012-01-11 | 2017-12-06 | Honda Research Institute Europe GmbH | Vehicle with computing means for monitoring and predicting traffic participant objects |
US9390370B2 (en) * | 2012-08-28 | 2016-07-12 | International Business Machines Corporation | Training deep neural network acoustic models using distributed hessian-free optimization |
DE102013214233B4 (en) | 2013-07-19 | 2015-02-05 | Honda Motor Co., Ltd. | Intelligent forward collision warning system |
US20160259842A1 (en) * | 2013-10-23 | 2016-09-08 | Ge Aviation Systems Limited | System and method for categorizing events |
US9568611B2 (en) * | 2014-08-20 | 2017-02-14 | Nec Corporation | Detecting objects obstructing a driver's view of a road |
US9592828B2 (en) | 2015-04-13 | 2017-03-14 | Nec Corporation | Long term driving danger prediction system |
US9286524B1 (en) * | 2015-04-15 | 2016-03-15 | Toyota Motor Engineering & Manufacturing North America, Inc. | Multi-task deep convolutional neural networks for efficient and robust traffic lane detection |
US9275347B1 (en) * | 2015-10-09 | 2016-03-01 | AlpacaDB, Inc. | Online content classifier which updates a classification score based on a count of labeled data classified by machine deep learning |
-
2017
- 2017-04-04 US US15/478,823 patent/US20170293837A1/en not_active Abandoned
- 2017-04-04 US US15/478,886 patent/US10296796B2/en active Active
- 2017-04-06 WO PCT/US2017/026365 patent/WO2017177008A1/en active Application Filing
- 2017-04-06 WO PCT/US2017/026362 patent/WO2017177005A1/en active Application Filing
Cited By (38)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11341398B2 (en) * | 2016-10-03 | 2022-05-24 | Hitachi, Ltd. | Recognition apparatus and learning system using neural networks |
US10395144B2 (en) * | 2017-07-24 | 2019-08-27 | GM Global Technology Operations LLC | Deeply integrated fusion architecture for automated driving systems |
DE102017127600A1 (en) | 2017-11-22 | 2019-05-23 | Connaught Electronics Ltd. | Method for automatic parking for a motor vehicle with a driving support system |
DE102017128082A1 (en) | 2017-11-28 | 2019-05-29 | Connaught Electronics Ltd. | Meta-architecture design for a CNN network |
CN110008978A (en) * | 2017-12-07 | 2019-07-12 | 亿目朗欧洲股份有限公司 | Classification of risks training method, classification of risks method, auxiliary or Automated Vehicle Operation system |
US11410327B2 (en) * | 2018-03-02 | 2022-08-09 | Nippon Telegraph And Telephone Corporation | Location determination apparatus, location determination method and computer program |
CN112166438A (en) * | 2018-03-13 | 2021-01-01 | 雷哥尼公司 | Deterministic labeled data generation and artificial intelligence training pathways |
US10599146B2 (en) | 2018-03-26 | 2020-03-24 | Ford Global Technologies, Llc | Action-conditioned vehicle control |
DE102018206208A1 (en) * | 2018-04-23 | 2019-10-24 | Robert Bosch Gmbh | Method, device, product and computer program for operating a technical system |
US11138467B2 (en) | 2018-04-23 | 2021-10-05 | Robert Bosch Gmbh | Method, device, product, and computer program for operating a technical system |
SE1851125A1 (en) * | 2018-09-21 | 2019-06-17 | Scania Cv Ab | Method and control arrangement for machine learning of a model-based vehicle application in a vehicle |
CN112638737A (en) * | 2018-10-16 | 2021-04-09 | 丰田自动车工程及制造北美公司 | Vehicle speed predictor using neural networks based on V2X data augmentation for predictive optimal control of networked and automated vehicles |
US11194331B2 (en) | 2018-10-30 | 2021-12-07 | The Regents Of The University Of Michigan | Unsupervised classification of encountering scenarios using connected vehicle datasets |
US10943154B2 (en) * | 2019-01-22 | 2021-03-09 | Honda Motor Co., Ltd. | Systems for modeling uncertainty in multi-modal retrieval and methods thereof |
CN110234018A (en) * | 2019-07-09 | 2019-09-13 | 腾讯科技(深圳)有限公司 | Multimedia content description generation method, training method, device, equipment and medium |
US11558584B2 (en) * | 2019-07-11 | 2023-01-17 | Chris Pritchard | Systems and methods for providing real-time surveillance in automobiles |
CN110696835A (en) * | 2019-10-11 | 2020-01-17 | 深圳职业技术学院 | Automatic early warning method and automatic early warning system for dangerous driving behaviors of vehicle |
US11676488B2 (en) | 2019-10-11 | 2023-06-13 | Aptiv Technologies Limited | Method and system for determining an attribute of an object at a pre-determined time point |
US12072443B2 (en) | 2019-11-15 | 2024-08-27 | Nvidia Corporation | Segmentation of lidar range images |
US12164059B2 (en) | 2019-11-15 | 2024-12-10 | Nvidia Corporation | Top-down object detection from LiDAR point clouds |
US12080078B2 (en) | 2019-11-15 | 2024-09-03 | Nvidia Corporation | Multi-view deep neural network for LiDAR perception |
CN112825136A (en) * | 2019-11-21 | 2021-05-21 | 辉达公司 | Deep neural network for detecting obstacles using RADAR sensors in autonomous machine applications |
US12399253B2 (en) | 2019-11-21 | 2025-08-26 | Nvidia Corporation | Deep neural network for detecting obstacle instances using radar sensors in autonomous machine applications |
US12050285B2 (en) | 2019-11-21 | 2024-07-30 | Nvidia Corporation | Deep neural network for detecting obstacle instances using radar sensors in autonomous machine applications |
US11941509B2 (en) | 2020-02-27 | 2024-03-26 | Aptiv Technologies AG | Method and system for determining information on an expected trajectory of an object |
US20210326699A1 (en) * | 2020-04-21 | 2021-10-21 | Inrix, Inc. | Travel speed prediction |
US12111386B2 (en) | 2020-07-24 | 2024-10-08 | Aptiv Technologies AG | Methods and systems for predicting a trajectory of an object |
CN112558510A (en) * | 2020-10-20 | 2021-03-26 | 山东亦贝数据技术有限公司 | Intelligent networking automobile safety early warning system and early warning method |
CN112937441A (en) * | 2021-03-29 | 2021-06-11 | 陕西科技大学 | Automobile anti-collision system and control method |
CN113792598A (en) * | 2021-08-10 | 2021-12-14 | 西安电子科技大学广州研究院 | Vehicle-mounted camera-based vehicle collision prediction system and method |
US12030512B2 (en) * | 2022-04-06 | 2024-07-09 | GM Global Technology Operations LLC | Collision warning system for a motor vehicle having an augmented reality head up display |
US20230322248A1 (en) * | 2022-04-06 | 2023-10-12 | Gm Global Technology Operation Llc | Collision warning system for a motor vehicle having an augmented reality head up display |
CN114881078A (en) * | 2022-05-07 | 2022-08-09 | 安徽蔚来智驾科技有限公司 | Method and system for filtering data under predetermined scenarios |
WO2023246389A1 (en) * | 2022-06-20 | 2023-12-28 | 支付宝(杭州)信息技术有限公司 | Fusion training based on knowledge representation learning for risk control model |
US20250013672A1 (en) * | 2023-07-03 | 2025-01-09 | Red Atlas Inc. | Systems and methods for region-based segmentation of a knowledge base developed using data collected from myriad sources |
US12293301B2 (en) | 2023-07-03 | 2025-05-06 | Red Atlas Inc. | Systems and methods for developing a knowledge base comprised of multi-modal data from myriad sources |
US12332922B2 (en) | 2023-07-03 | 2025-06-17 | Red Atlas, Inc. | Systems and methods for developing and organizing a knowledge base comprised of data collected from myriad sources |
US12339878B2 (en) * | 2023-07-03 | 2025-06-24 | Red Atlas Inc. | Systems and methods for region-based segmentation of a knowledge base developed using data collected from myriad sources |
Also Published As
Publication number | Publication date |
---|---|
US10296796B2 (en) | 2019-05-21 |
WO2017177005A1 (en) | 2017-10-12 |
US20170293815A1 (en) | 2017-10-12 |
WO2017177008A1 (en) | 2017-10-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10296796B2 (en) | Video capturing device for predicting special driving situations | |
US12236327B1 (en) | Automated system and method for modeling the behavior of vehicles and other agents | |
US12228926B1 (en) | System and method for predicting behaviors of detected objects through environment representation | |
US12233852B1 (en) | Handling sensor occlusions for autonomous vehicles | |
US11619940B2 (en) | Operating an autonomous vehicle according to road user reaction modeling with occlusions | |
US11055605B2 (en) | Detecting dangerous driving situations by parsing a scene graph of radar detections | |
US20190340522A1 (en) | Event prediction system, event prediction method, recording media, and moving body | |
US10773643B1 (en) | Maintaining road safety when there is a disabled autonomous vehicle | |
JP2020109681A (en) | Determining future traveling direction using wheel attitude | |
KR20190126258A (en) | Electronic device for vehicle and method for operating the same | |
US20190256087A1 (en) | Autonomous vehicle and operating method thereof | |
US20180004205A1 (en) | Control device of vehicle | |
CN107615353A (en) | Automobile Peripheral Information Display System | |
US11651692B2 (en) | Presenting relevant warnings to a vehicle operator | |
CN110803169B (en) | System and method for displaying information in a vehicle | |
US12030512B2 (en) | Collision warning system for a motor vehicle having an augmented reality head up display | |
KR20230159450A (en) | Information processing devices, information processing methods and programs | |
US20250218065A1 (en) | Information processing apparatus and information processing method | |
WO2025022997A1 (en) | Signal processing device, signal processing method, and signal processing system | |
WO2024232244A1 (en) | Information processing device, information processing method, and program | |
CN120319019A (en) | Vehicle information prompting method and vehicle | |
KR20230067799A (en) | Method and Apparatus for controlling virtual lane based on environmental conditions | |
CN119142248A (en) | Vehicle control method, apparatus, computer device, storage medium, and computer program product | |
CN114998885A (en) | Page data processing method and device, vehicle and storage medium | |
CN116278739A (en) | Risk reminding method and device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: NEC LABORATORIES AMERICA, INC., NEW JERSEY Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:COSATTO, ERIC;MELVIN, IAIN;GRAF, HANS PETER;REEL/FRAME:041847/0836 Effective date: 20170330 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |