WO2021058300A1 - Object detection circuitry and object detection method - Google Patents

Object detection circuitry and object detection method Download PDF

Info

Publication number
WO2021058300A1
WO2021058300A1 PCT/EP2020/075417 EP2020075417W WO2021058300A1 WO 2021058300 A1 WO2021058300 A1 WO 2021058300A1 EP 2020075417 W EP2020075417 W EP 2020075417W WO 2021058300 A1 WO2021058300 A1 WO 2021058300A1
Authority
WO
WIPO (PCT)
Prior art keywords
data
feature
sensor
predefined
object detection
Prior art date
Application number
PCT/EP2020/075417
Other languages
French (fr)
Inventor
Florian Becker
Original Assignee
Sony Corporation
Sony Europe B.V.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sony Corporation, Sony Europe B.V. filed Critical Sony Corporation
Priority to EP20768050.5A priority Critical patent/EP4035059A1/en
Priority to CN202080065788.8A priority patent/CN114424254A/en
Priority to US17/635,385 priority patent/US20220406044A1/en
Priority to JP2022517741A priority patent/JP7586173B2/en
Publication of WO2021058300A1 publication Critical patent/WO2021058300A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • G06T7/74Determining position or orientation of objects or cameras using feature-based methods involving reference images or patches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/7715Feature extraction, e.g. by transforming the feature space, e.g. multi-dimensional scaling [MDS]; Mappings, e.g. subspace methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/776Validation; Performance evaluation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/80Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
    • G06V10/806Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10032Satellite or aerial image; Remote sensing
    • G06T2207/10044Radar image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Definitions

  • the present disclosure generally pertains to an object detection circuitry and an object detection method.
  • Known ways of detecting an object rely on an output result of one or multiple sensors.
  • an object detection may be performed and after that, sensor data or the data indica tive of the detected object for each sensor may be fused in order to receive a final object detection.
  • the disclosure provides an object detection circuitry configured to: obtain first feature data which are based on first sensing data of a first sensor; compare the first feature data to a first predetermined feature model being representative of a predefined object, wherein the first predetermined feature model is specific for the first sensor, thereby generating first object probability data; obtain second feature data which are based on second sensing data of a second sensor; compare the second feature data to a second predetermined feature model being representa tive of the predefined object, wherein the second predetermined feature model is specific for the second sensor, thereby generating second object probability data; and combine the first and the second object probability data, thereby generating combined prob ability data for detecting the predefined object.
  • the disclosure provides an object detection method comprising: obtaining first feature data which are based on first sensing data of a first sensor; comparing the first feature data to a first predetermined feature model being representative of a predefined object, wherein the first predetermined feature model is specific for the first sensor, thereby generating first object probability data; obtaining second feature data which are based on second sensing data of a second sensor; comparing the second feature data to a second predetermined feature model being repre sentative of the predefined object, wherein the second predetermined feature model is specific for the second sensor, thereby generating second object probability data; and combining the first and the second object probability data, thereby generating combined probability data for detecting the predefined object.
  • Fig. 1 depicts a coordinate system including a signal distribution acquired by a stereo camera
  • Fig. 2 depicts a coordinate system including a signal distribution acquired by a radar sensor
  • Fig. 3 depicts a template (feature model) for a stereo camera
  • Fig. 4 depicts a template (feature model) for a radar sensor
  • Fig. 5 depicts a fused object pose probability distribution
  • Fig. 6 depicts a coordinate system including detected objects based on feature maps of the stereo camera and the radar sensor
  • Fig. 7 depicts a block diagram of a method according to the present disclosure
  • Fig. 8 schematically illustrates a method for generating a feature map according to the present disclo sure
  • Fig. 9 schematically illustrates a method for determining object probability data according to the pre sent disclosure.
  • Fig. 10 depicts a block diagram of a vehicle including an object detection circuitry according to the present disclosure.
  • known ways of detecting an object may lead to a high processing com plexity, and thereby to a high power consumption and an increase in costs.
  • a low-level sensor measurement e.g. a distance
  • planning purposes e.g. of a route
  • relevant information may be bundled in a representation of an environment, wherein the representation may fulfill requirements of com pactness, completeness and preciseness, which may, for example, be a list of objects including pa rameters indicating their position.
  • a sensor-specific appearance of the objects must be taken into account (e.g. an object may be represented differendy through a camera than through a radar).
  • a further requirement may be sensor fusion. Since different sensors may rely on different measure ment principles and output data of each sensor may be different, as well, there may be an increased effort to operate the sensors without increasing a system complexity.
  • a further requirement may be a fusion of all available measurements. It has been recognized that known devices may translate low-level measurements into high-level abstract representations for each sensor and then fuse the obtained information into a joint representation. However, infor mation, which is discarded in such an abstraction, may not be available during or after the fusion an ymore, such that a quality may be reduced. Therefore, it is desirable to at least reduce an amount of a quality loss.
  • a further requirement may be an overcoming of a view-dependent object appearance. It has been recognized that a predefined object may vary in a sensor appearance, such that an object orientation may depend on a sensor viewing direction (i.e. a viewing angle).
  • a self occlusion effect of the object such that a sensor may only observe a part of an object.
  • the observed part may appear differently depending of a viewing angle of a sensor. Therefore, it is generally desir able to overcome such an effect for reliably detecting the object.
  • a further requirement may be in-class object invariance. Different parts of the same object may be considered to belong to different objects. For example, an object may have different shapes and/or colors at different spots, such that a different shape may be recognized as a different object. How ever, it is generally desirable to provide a way to detect an object as one object although there are different shapes and/or colors at the object.
  • a further requirement may be multi-object detection. For example, it may be necessary to detect more than one object (e.g. the closest one), such as detecting all objects, which are within a detection range.
  • a further requirement may be sensor noise.
  • some embodiments pertain to an object detection circuitry configured to: obtain first feature data which are based on first sensing data of a first sensor; compare the first feature data to a first predetermined feature model being representative of a predefined object, wherein the first predeter mined feature model is specific for the first sensor, thereby generating first object probability data; obtain second feature data which based on second sensing data of a second sensor; compare the sec ond feature data to a second predetermined feature model being representative of the predefined object, wherein the second predetermined feature model is specific for the second sensor, thereby generating second object probability data; and combine the first and the second object probability data, thereby generating combined probability data for detecting the predefined object.
  • the object detection circuitry may include one or multiple processors, such as a CPU (central pro cessing unit), GPU (graphic processing units), one or multiple FPGAs (field programmable gate ar ray), a computer, a server, and the like, also combinations of multiple of the named elements, and the like.
  • processors such as a CPU (central pro cessing unit), GPU (graphic processing units), one or multiple FPGAs (field programmable gate ar ray), a computer, a server, and the like, also combinations of multiple of the named elements, and the like.
  • the object detection circuitry may further be included in or associated to an automotive environ ment, such that a detection of objects may be performed according to an operational state of, for example, a car, without limiting the present disclosure in that regard, since the object detection cir cuitry may be applied in any field, in which sensor data are utilized for an object detection, such as (surveillance) camera systems, and the like.
  • the object detection circuitry may be configured to obtain feature data, i.e. the feature data may be provided to the object detection circuitry by at least one sensor to which the object detection cir cuitry may be coupled, connected, associated, and the like.
  • the at least one sensor may be included in the object detection circuitry, such that the feature data may be acquired within the object detection circuitry.
  • feature data which are obtained by multiple sensors may be (structurally or logically) distinguished, i.e. feature data from a first sensor may be referred to as first feature data, feature data from a second sensor may be referred to as second feature data, and so on.
  • the first feature data may be based on first sensing data.
  • the first sensor may perform a measurement according to an internal configuration; for example it may detect a reflection of a radio wave (e.g. in the case of a radar sensor), such that, for example, an ana log or digital sensor output may be generated, to which it may be referred to as first sensing data.
  • a radio wave e.g. in the case of a radar sensor
  • the first feature data may include an evaluation result of the first sensing data, for example a signal intensity according to a distance in a predefined spatial resolution, such that a feature map, as it is generally known, can be generated.
  • the first sensor may be configured to generate the first feature data, whereas in some embodiments, the first feature data are generated in the object detection circuitry.
  • a requirement of abstraction as discussed above, may fulfilled.
  • a detection list of objects including the predefined object, including a position, an orienta tion, and a correction estimate may be provided, such that a further pose refinement and/ or a track ing of the object may be carried out.
  • the object detection circuitry may be further configured to compare the first feature data to a first predefined model.
  • the comparison may be performed by performing a mapping between the first feature data and model feature data (e.g. comparing the generated feature map with a predefined feature map), and detecting common and/ or comparable signals.
  • the feature model may include data indicative of a predefined feature map, which may be specific for the first sensor in a way that the first feature data may be comparable to the feature model, i.e. a radar feature model may be compared to feature data obtained based on first sensing data of a radar sensor, and the like, without limiting the present disclosure to a radar sensor.
  • a camera RGB, infrared, time-of-flight
  • a lidar sensor may con stitute the first sensor
  • the first sensing data may be fused sensing data of multiple sensors, such as a fused signal of an RGB and a time-of-flight camera, and the like.
  • the feature model may be representative of a predefined object.
  • a predefined object may be a car, a pedestrian, a street sign, and the like, which may be indicated by a specific signal signature, i.e. object specific sensing data may be generated in response to a detection of the predefined object.
  • the object specific sensing data may be predefined, acquired in a calibration process, a training pro cess, and the like, of the object detection circuitry, such that an assignment between the predefined object and the first feature data can be performed based on the comparison between the first feature data and the first predetermined feature model.
  • first object probability data may be generated, which may indicate a correspondence between the first feature data and the first predetermined feature model. In other words, it may be deter mined to which extent the first feature data may correspond to the predetermined object or how probable it is that the first feature data indicate the predetermined object.
  • a similar process may be performed to second feature data based on second sensing data of a sec ond sensor, such that a repetitive description thereof is omitted.
  • the first and the second sensor may be same or different sensors, e.g. the first sensor may be a radar, the second sensor may be an RGB camera, or the like.
  • the present disclosure is not limiting in that the second sensor is configured to work the same way as the first sensor.
  • the first sensor the first feature data may be generated based on the first sensing data
  • the second sensor only the second sensing data may be acquired and the second feature data may be generated in the object detection circuitry, or vice versa.
  • the object detection circuitry may be further configured to combine the first and the second object probability data.
  • the combination may be based on a multiplexing, multiplying, adding, and the like, of the first and the second object probability data, which may result in a probability value, but, in some embodi ments, the combination may be based on an abstract data structure, which may be generated by an artificial intelligence, a neural network, based on a supervised or unsupervised training, and the like.
  • the combination procedure may be followed by a normalization of the combined probability data, as it is generally known.
  • the predefined object may be detected based on the combined probability data, i.e. a position in a space of the object may be determined. Moreover, a posture of the predefined object may be deter mined, since the first and the second feature data may be acquired for a plurality of locations with a (sensor specific) spatial resolution.
  • the object detection circuitry is further configured to: detect a plurality of maxima of the combined probability data being indicative for at least one position parameter of the predefined object; and determine the at least one position parameter of the predefined object.
  • the combined probability data may represent a distribution of probabilities of features (or occur rences) of the predefined object in a space, such that a high probability (i.e. a probability above a predetermined threshold) may be a maximum and a probability below a predetermined threshold may be a minimum, such that probabilities above the predetermined threshold of the combined probability data may be taken into account for determining the at least one position parameter of the predefined object.
  • a high probability i.e. a probability above a predetermined threshold
  • a probability below a predetermined threshold may be a minimum
  • a maximum may indicate a possibility (or a hypothesis) of an object (e.g. a single or a plurality) at a predetermined location, such that a (symbolic) position of the maximum in a probability space may indicate a (real) spatial position and orientation of the object.
  • the detection of a plurality (i.e. at least two) of maxima may be based on known maxima detection processes, e.g. if a maximum is defined to have a Gaussian distribution (or any other distribution, such as Lorentzian, Dirac, Fermi, Poisson, etc.), a Gaussian distribution (or any other) may be detected, as it is generally known.
  • a Gaussian distribution or any other distribution, such as Lorentzian, Dirac, Fermi, Poisson, etc.
  • the plurality of maxima may be detected to be arranged in a spatial order (or pattern), such that a position parameter may be determined.
  • two maxima may be detected, which are aligned on a horizontal line of a (virtual) coordinate system, wherein the two maxima are two meters apart from each other. It may be a premise that the first sensor and the second sensor may be aligned on a horizontal axis, which may be parallel to the horizontal line. In such a case, it may be detected that a posture of the detected car is parallel to the horizontal axis of the sensors.
  • the sensors are implemented in a car including the object detection circuitry according to the present disclosure, it may be inferred that the detected car may be perpendicular to the car in cluding the object detection circuitry, such that, for example, a warning may be issued to a driver of the car including the object detection circuitry, or an emergency braking may be issued, and the like, without limiting the present disclosure in that regard.
  • a posture and/ or a positional angle of the detected car may be determined.
  • the present disclosure is not limited to the position parameter being a posture or an angle, since, in some embodiments, the at least one position parameter includes at least one of a position, a distance, an angle, and a posture.
  • a position may be determined and may be represented in a (virtual) coordinate system, which may also be shown to a user of the object detection circuitry on a display, and the like.
  • a distance may be concluded from the position, without limiting the present disclosure in that re gard, as the distance may be direcdy determined out of the combined probability data.
  • the posture and/or the angle may further include a direction of the predefined object. For example, if a car is the predefined object, it may be determined where a front and/ or where a back of the car is.
  • the object detection circuitry is further configured to determine at least one of a correctness and a precision of the detection of the predefined object.
  • a measurement may be deteriorated, e.g. by environmental influences (e.g. a tempera ture, a light condition, and the like), a deteriorated calibration of one (or both) of the sensors, due to a movement of the sensors (e.g. if implemented in a car) or of the predefined object (e.g. due to Doppler effect).
  • environmental influences e.g. a tempera ture, a light condition, and the like
  • a deteriorated calibration of one (or both) of the sensors due to a movement of the sensors (e.g. if implemented in a car) or of the predefined object (e.g. due to Doppler effect).
  • a precision of the detection may be determined based on an assumed error (function), which may take into account any of the conditions mentioned above, without limiting the present disclosure in that regard.
  • a correctness of the detection may be determined (e.g. based on the precision), which may indicate whether the predefined object corresponds to the detected object.
  • the object detection circuitry is further configured to track the detected pre defined object.
  • known object tracking technologies may be applied.
  • the object detection circuitry may perform the same processing for the tracking again as it is applied for the detection of the predefined object, such that a tracking may correspond to a repetitive detecting of the predefined object.
  • the object detection circuitry is further configured to: generate a first feature map based on the first sensing data; generate a second feature map based on the second sensing data; and transfer the first and the second feature map into a predefined coordinate system.
  • a feature map may be a map (e.g. based on the first and/ or the sensing data) depicting the obtained first and second feature data, either in one map or in two maps, for example.
  • the feature map in cludes, in some embodiments, the detected plurality of maxima, and may be represented by a coor dinate system from a bird’s eye perspective, without limiting the present disclosure in that regard.
  • a feature map may be generated for different layers of a space, such that data indicating a vertical direction may be acquired.
  • a feature represented in the feature map may be a detected signal (or one maximum or more of the plurality of maxima), which is indicative of a specific property of the predefined object.
  • a feature may be indicated by a detected signal stemming from a specific part of a car, e.g. a window, a back, and the like, which has a specific feature signature.
  • the predefined coordinate system may be a common coordinate system of the first and the second feature map or it may be different coordinate systems, which, however, may be convertible into each other, and may also be transferred into a common coordinate system.
  • the predefined coordinate system is not limited to a particular kind of coordinate systems as it may represent Euclidean coordinate, spherical coordinates, cylindrical coordinates, polar coordinates, and the like, and the origin of the predefined coordinate system may, for example be a particular location of one of the two sensors, a detected object, or any other location in a space.
  • the first and the second sensor include at least one of a radar sensor, a lidar sensor, a camera, or a time-of-flight sensor, as discussed herein.
  • At least one of the first predetermined feature model and the second prede termined feature model is based on a supervised training of an artificial intelligence.
  • the artificial intelligence may use machine learning based methods or explicit feature based methods, such as shape matching, for example by edge detection, histogram based methods, tem plate match based methods, color match based methods, or the like.
  • a ma chine learning algorithm may be used for performing object recognition, e.g. for comparing a detected predefined object with a recognized object to increase a correctness of a detection, which may be based on at least one of: Scale Invariant Feature Transfer (SIFT , Gray Level Co-occurrence Matrix (GLCM), Gabor Features, Tubeness, or the like.
  • SIFT Scale Invariant Feature Transfer
  • GLCM Gray Level Co-occurrence Matrix
  • the machine learning algorithm may be based on a classifier technique, wherein such a machine learning algorithm may be based on least one of: Random Forest; Support Vector Machine; Neural Net, Bayes Net, or the like. Further more, the machine learning algorithm may apply deep-learning techniques, wherein such deep-leam- ing techniques may be based on at least one of: Autoencoders, Generative Adversarial Network, weakly supervised learning, boot-strapping, or the like.
  • the supervised learning may further be based on a regression algorithm, a perception algorithm, Bayes-classification, Naiver Bayer classification, next-neighbor classification, artificial neural net work, and the like.
  • the artificial intelligence may, in such embodiments, be fed with ground truth data, which may cor respond to or be based on the predefined object, such that the artificial intelligence may learn to as sign the first and/ or the second feature data to the ground truth data, thereby developing or generating the first and/ or the second feature model.
  • the predefined object is based on a class of predefined objects.
  • a class of predefined objects may, in such embodiments, include a car, a passenger, a road sign, a street, a house, a tree, an animal, a traffic light, and the like.
  • the object detection circuitry may be applied to a surveillance sys tem, for example of a warehouse, or the like, such that the class of predefined objects may include a customer, an employee, a shelf, a product, and the like, without limiting the present disclosure to embodiments pertaining to automotive or surveillance applications.
  • the present disclosure may be applied in any field, in which an object detection may be performed.
  • the object detection circuitry is further configured to: iteratively convolve the first feature data with the first predetermined feature model, thereby generating the first object probability data; iteratively convolve the second feature data with the second predetermined feature model, thereby generating the second object probability data; and iteratively convolve the first and the second object probability data, thereby generating the combined object probability data.
  • a plurality of first feature data may be acquired (e.g. for different heights or layers, for different angles, and the like), which may each be convolved (e.g. compared, as discussed above) with the first predetermined feature model in a consecutive ordering (i.e. iteratively), to which it may in this disclosure be referred to as vertical iteration.
  • a plurality of second feature data may be acquired, which may each be convolved with the second predetermined feature model in a con secutive ordering, as well in a vertical iteration.
  • first generated first object probability data (e.g. from a first measurement of the first sen sor) may be convolved with the first generated second object probability data (e.g. by summing up the respective probabilities), to which it may be referred to as horizontal iteration, whereby the com bined probability data are generated.
  • Some embodiments pertain to an object detection method including: obtaining first feature data which are based on first sensing data of a first sensor; comparing the first feature data to a first pre determined feature model being representative of a predefined object, wherein the first predeter mined feature model is specific for the first sensor, thereby generating first object probability data; obtaining second feature data which based on second sensing data of a second sensor; comparing the second feature data to a second predetermined feature model being representative of the prede fined object, wherein the second predetermined feature model is specific for the second sensor, thereby generating second object probability data; and combining the first and the second object probability data, thereby generating combined probability data for detecting the predefined object, as discussed herein.
  • the object detection method may be performed with an object detection circuitry according to the present disclosure, and the like, such as a processor (or multiple processors) configured to perform the object detection method.
  • an object detection circuitry such as a processor (or multiple processors) configured to perform the object detection method.
  • the method further includes: detecting a plurality of maxima of the com bined probability data being indicative for at least one position parameter of the predefined object; and determining the at least one position parameter of the predefined object, as discussed herein.
  • the at least one position parameter includes at least one of a position, a dis tance, an angle, and a posture, as discussed herein.
  • the method further in cludes: determining at least one of a correctness and a precision of the detection of the predefined object, as discussed herein.
  • the method further includes tracking the detected predefined object, as discussed herein.
  • the method further includes: generat ing a first feature map based on the first sensing data; generating a second feature map based on the second sensing data; and transferring the first and the second feature map into a predefined coordi nate system, as discussed herein.
  • the first and the second sensor include at least one of a radar sensor, a lidar sensor, a camera, or a time-of-flight sensor, as discussed herein.
  • at least one of the first predetermined feature model and the second predeter mined feature model is based on a supervised training of an artificial intelligence, as discussed herein.
  • the predefined object is based on a class of predefined objects, as dis cussed herein.
  • the method further includes: iteratively convolving the first feature data with the first predetermined feature model, thereby generating the first object probabil ity data; iteratively convolving the second feature data with the second predetermined feature model, thereby generating the second object probability data; and iteratively convolving the first and the second convolved feature data, thereby generating the combined probability data, as discussed herein.
  • the methods as described herein are also implemented in some embodiments as a computer pro gram causing a computer and/ or a processor to perform the method, when being carried out on the computer and/ or processor.
  • a non-transitory computer-readable record ing medium is provided that stores therein a computer program product, which, when executed by a processor, such as the processor described above, causes the methods described herein to be per formed.
  • each sensor i.e. the first and the second sensor, but not limited to two sensors
  • 1, i) and a number of measurements not detecting the particular feature, i.e. n(f 0
  • a probability p(0(c, Q) 1
  • fi, ..., fis t ) of an existence of an object in an object parameter space may be determined, wherein O may be the ob ject, c may correspond to coordinates (c x , c z ) of the object, Q may be indicative of the posture of the object (i.e. an angle), (c, Q) being the (three-dimensional) parameter space.
  • c x may include a center of the object in an x-direction
  • c z may include the center of the object in a z-direction
  • Q may include an orientation of the object in an x-z-plane, which may be limited to an interval of values, such as [0, 2p).
  • a discretized representation of the parameter space is generated, which may be represented by a three-dimensional array including a number of discretization points along the x- axis times a number of discretization points along the z-axis times a number of discretization points along an orientation axis (i.e. the angle).
  • a logarithmic value of p(0(c, Q) 1
  • fi, ..., f 3 ⁇ 4i ) may be com puted, which may reduce a number of inference steps to a number of convolutions.
  • the terms p(0(c, Q) 1
  • the terms p(0(c - 1, Q) 1
  • sensor noise may be recognized and/ or filtered out, such that a requirement of sensor noise, as discussed above, may be fulfilled.
  • fi(0) 1) are located in the parameter space while excluding other parameters based on a non-maximum suppression.
  • fi, ..., fin) at the loca tion of the local maximum, wherein a correctness may be given, if p(0(c, Q) 1
  • a requirement of multi-object detection may be fulfilled by providing a list of detection, wherein the precision of the pose may depend on a resolution of the pose space (c, 9).
  • a detection of the pose (or posture) may be refined in a consecutive measurement based on the cor rectness and/ or the precision of the measurement.
  • a first coordinate system 1 including two axes 2a and 2b each indicating positional coordinates.
  • signal distributions 3a, 3b and 3c (first feature data) are shown, which are obtained based on first sensing data by a stereo camera as a first sensor. Only the signal distribution 3a is dis cussed herein, in order to omit a repetitive discussion.
  • the signal distribution 3a has a first sub-signal 4 and a second sub-signal 5, wherein the first sub signal 4 has a stronger signal intensity than the second sub-signal 5.
  • Fig. 2 depicts a second coordinate system 10 including two axes 12a and 12b, which correspond to the axes 2a and 2b Fig. 1.
  • the signal distribution 13a (and correspondingly 13b and 13c although not explicitly discussed again) includes first sub-signals 14, second sub-signals 15, and third sub-signals 16, which are, for illustrative reasons, grouped by different patterns, wherein the pattern strength corresponds to the respective signal strength, i.e. the first sub-signals 14 have the strongest signal intensity of the first to third sub-signals 14 to 16, the second sub-signals 15 have the second strongest signal intensity of the first to third sub-signals 14 to 16, and the third sub-signals 16 have the third strongest signal inten sity of the first to third sub-signals 14 to 16.
  • the signal distributions 3a, 3b, 3c, 13a, 13b, and 13c de scribed with respect to Figs. 1 and 2 are obtained by acquiring first and second sensing data of three cars as predefined objects, i.e. the signal distributions 3a to 3c and 13a to 13c are indicative of cars.
  • Fig. 3 depicts a template (i.e. a feature model) 20 for a stereo camera as a sensor, which is based on a supervised training of an artificial intelligence.
  • a template i.e. a feature model
  • the template 20 represents different postures of cars 24 to 39, wherein each posture is assigned to a probability, as discussed herein.
  • different intensities of the signals are symbolically depicted with different lines 21 to 23, wherein the line 21 represents a strongest intensity, the line 22 the second strongest, and the line 23 the weakest detected intensity, without limiting the present disclosure in that regard, since, as it is generally known, a continuous distribution of intensities may be detected, as well.
  • the assignment is displayed in the following table, wherein each reference sign (corresponding to a posture as depicted) is assigned to a probability and the posture is represented by an angle Q.
  • Fig. 4 depicts a template (i.e. a feature model) 40 for a radar as a sensor, which is based on a super vised training of an artificial intelligence.
  • the template 40 represents different postures of cars, which correspond to the ordering (in terms of probability and angle) of the template 20 of Fig. 3, and, therefore, a repetitive description (together with reference signs) is omitted.
  • Different intensities of the signals are symbolically depicted with differently hatched ellipses 41 to 43, wherein the ellipse 41 represents a strongest intensity, the ellipse 42 the second strongest, and the ellipse 43 the weakest detected intensity, without limiting the present disclosure in that regard, since, as it is generally known, a continuous distribution of intensities may be detected, as well.
  • the logarithmic probabilities of object poses is determined using templates 20 and 40, as discussed herein, and the object pose probability distributions are fused, such that a fused pose probability dis tribution (or a fused feature model) is generated.
  • Fig. 5 depicts such a fused pose probability distribution 50 with the same ordering of postures of the car as in Figs. 3 and 4, such that a repetitive description (together with reference signs) is omitted.
  • Different intensities of the signals are symbolically depicted with differently hatched rectangles 51 to 53, wherein the rectangle 51 represents a strongest intensity, the rectangle 52 the second strongest, and the rectangle 53 the weakest detected intensity, without limiting the present disclosure in that regard, since, as it is generally known, a continuous distribution of intensities may be detected, as well.
  • the logarithmic probability of the first and the second feature data (of in Figs. 1 and 2) are fused into a combined pose probability distribution 50 for each of the three objects (cars) of Figs. 1 and 2, such that poses of the three cars.
  • Fig. 6 depicts a coordinate system 60 including two axes 61a and 61b each indicating positional co ordinates.
  • three detected objects 62a, 62b and 62c are shown, overlaid with summed up features 63 of the feature maps 1 and 10 of Figs. 1 and 2, respectively.
  • the display of the three detected objects further include pointers 64a, 64b, and 64c indicating a posture of each detected object 62a, 62b, and 62c, respectively.
  • Fig. 7 depicts a block diagram of a method 70 according to the present disclosure.
  • a sensor measurement is performed.
  • a stereo camera image pair as first sensing raw data is acquired in 711 and radar data as second sensing raw data is acquired in 712.
  • the low-level processed first and second sensing raw data are transferred into respective fea ture maps, such that the low-level processed first sensing raw data are transferred into a feature map in bird’s eye view in 721, and the low-level processed second sensing raw data are transferred into a feature map in bird’s eye view of 722.
  • a logarithmic probability is determined from the feature maps.
  • a pose logarithmic probability volume of the stereo camera feature map is de termined based on previously trained conditional probabilities, which are being fed to the determina tion of the pose logarithmic probability volume in 732.
  • a pose logarithmic probability volume of the radar sensor is determined based on previously trained conditional probabilities, which are being fed to the determination of the pose logarithmic probability volume in 734.
  • the determined pose logarithmic probability volumes are combined (or fused) in 735, as discussed herein.
  • maxima are detected with a non-maxima suppression algorithm in 741, as discussed herein.
  • a pose, correctness and a precision is determined based on the detected maxima, as discussed herein.
  • a further refinement of the detected poses is performed in 751, such that the detected object (or detected objects) is tracked in 752, as discussed herein.
  • Fig. 8 schematically illustrates a method 80 for generating a feature map.
  • the coordinate system 81 includes a plurality of detections 83, which are overlapping with borders of the cells 82, without limiting the present disclosure in that regard, since a detection 83 may generally be fully inside a cell 82.
  • the detections 83 are assigned to one of the cells 82, in which most of the detection signal lies, i.e. a cell-wise accumulation of the detections 83 is performed in 84.
  • a coordinate system 85 is depicted, which basically corresponds to the coordinate system 81, such that a repetitive description is omitted. However, the coordinate system 85 is different from the coordinate system 81 in that the detections 83 are accumulated to features 86.
  • Fig. 9 schematically illustrates a method 90 for determining object probability data according to the present disclosure.
  • the following description of the method 90 only takes into account a measurement of two sensors 91 and 91’.
  • the disclosure is not limited to two sensors, as it may be envisaged for an arbitrary number of sensors with an arbitrary number of measurements.
  • First feature data which are based on a sensor measurement of a sensor 91 are convolved with a predetermined feature model 92 (i.e. a predetermined conditional probability based on a supervised training being specific for the sensor 91) in 93.
  • a predetermined feature model 92 i.e. a predetermined conditional probability based on a supervised training being specific for the sensor 91
  • Second feature data which are based on a sensor measurement of a sensor 91’ are convolved with a predetermined feature model 92’ (i.e. a predetermined conditional probability based on a supervised training being specific for the sensor 91’) in 93’.
  • a predetermined feature model 92’ i.e. a predetermined conditional probability based on a supervised training being specific for the sensor 91’
  • a sum 94 of the convolved data of 93 and the convolved data of 93’ is determined, which serves as a basis for a further convolution with further sensing data for detecting an object, or, in some embodiments, as a basis for detecting the object.
  • Fig. 10 depicts a block diagram of a vehicle 100, which includes a CPU 101, which is configured to function as an object detection circuitry according to the present disclosure.
  • the vehicle 100 further includes a stereo camera 102 for acquiring first sensing data and a radar 103 for acquiring second sensing data, which are fed to the CPU 101, such that first and second feature data are generated.
  • the first and the second feature data are generated in the first and the second sensor or in specific circuitry for generating the first and/ or the second feature data, as discussed herein.
  • a feature map may be computed in each sensor and may be transferred to a fusion- and-detection unit, which may be configured to fuse the feature maps, determine the probabilities and detect the object.
  • a non-transitory computer-readable recording medium stores therein a computer program product, which, when executed by a processor, such as the pro cessor described above, causes the method described to be performed.
  • the present disclosure although exemplarily described for a de tection of cars and an implementation in a vehicle, is not limited to the described embodiments.
  • the present disclosure may be applied in each case or situation, in which a surrounding scenario can be represented in a two-dimensional coordinate system, e.g. in any land-based or sea- surface based navigation system, such as driver assistance systems in cars, autonomous vehicles, ro bots, boats, ships, and the like.
  • An object detection circuitry configured to: obtain first feature data which are based on first sensing data of a first sensor; compare the first feature data to a first predetermined feature model being representative of a predefined object, wherein the first predetermined feature model is specific for the first sensor, thereby generating first object probability data; obtain second feature data which are based on second sensing data of a second sensor; compare the second feature data to a second predetermined feature model being representa tive of the predefined object, wherein the second predetermined feature model is specific for the second sensor, thereby generating second object probability data; and combine the first and the second object probability data, thereby generating combined prob ability data for detecting the predefined object.
  • the object detection circuitry of (1) further configured to: detect a plurality of maxima of the combined probability data being indicative for at least one position parameter of the predefined object; and determine the at least one position parameter of the predefined object.
  • the object detection circuitry of anyone of (1) to (3) further configured to: determine at least one of a correctness and a precision of the detection of the predefined ob ject.
  • the object detection circuitry of anyone of (1) to (5) further configured to: generate a first feature map based on the first sensing data; generate a second feature map based on the second sensing data; and transfer the first and the second feature map into a predefined coordinate system.
  • An object detection method comprising: obtaining first feature data which are based on first sensing data of a first sensor; comparing the first feature data to a first predetermined feature model being representative of a predefined object, wherein the first predetermined feature model is specific for the first sensor, thereby generating first object probability data; obtaining second feature data which are based on second sensing data of a second sensor; comparing the second feature data to a second predetermined feature model being repre sentative of the predefined object, wherein the second predetermined feature model is specific for the second sensor, thereby generating second object probability data; and combining the first and the second object probability data, thereby generating combined probability data for detecting the predefined object.
  • a computer program comprising program code causing a computer to perform the method according to anyone of (11) to (20), when being carried out on a computer.
  • (22) A non-transitory computer-readable recording medium that stores therein a computer pro gram product, which, when executed by a processor, causes the method according to anyone of (11) to (20) to be performed.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Data Mining & Analysis (AREA)
  • Medical Informatics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Databases & Information Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Probability & Statistics with Applications (AREA)
  • Radar Systems Or Details Thereof (AREA)
  • Image Analysis (AREA)
  • Traffic Control Systems (AREA)

Abstract

The present disclosure generally pertains to an object detection circuitry configured to: obtain first feature data which are based on first sensing data of a first sensor; compare the first feature data to a first predetermined feature model being representative of a predefined object, wherein the first predetermined feature model is specific for the first sensor, thereby generating first object probability data; obtain second feature data which are based on second sensing data of a second sensor; compare the second feature data to a second predetermined feature model being representative of the predefined object, wherein the second predetermined feature model is specific for the second sensor, thereby generating second object probability data; and combine the first and the second object probability data, thereby generating combined probability data for detecting the predefined object.

Description

OBJECT DETECTION CIRCUITRY AND OBJECT DETECTION
METHOD
TECHNICAL FIELD
The present disclosure generally pertains to an object detection circuitry and an object detection method.
TECHNICAL BACKGROUND
Generally, methods and devices for detecting an object are known. For example, in an automotive area (e.g. automotive security, autonomous driving, and the like), it is desirable to have a fast and a reliable method for detecting objects on a street.
Known ways of detecting an object (e.g. a car, a pedestrian, etc.) rely on an output result of one or multiple sensors.
For each sensor, an object detection may be performed and after that, sensor data or the data indica tive of the detected object for each sensor may be fused in order to receive a final object detection.
Hence, such known ways for object detection with multiple sensors may require an object detection for each of the used sensors, which may go in hand with a high complexity in terms of processing and/ or a slow or deteriorated final detection of the object after the sensor data are fused.
Although there exist techniques for detecting an object, it is generally desirable to provide an object detection circuitry and an object detection method.
SUMMARY
According to a first aspect, the disclosure provides an object detection circuitry configured to: obtain first feature data which are based on first sensing data of a first sensor; compare the first feature data to a first predetermined feature model being representative of a predefined object, wherein the first predetermined feature model is specific for the first sensor, thereby generating first object probability data; obtain second feature data which are based on second sensing data of a second sensor; compare the second feature data to a second predetermined feature model being representa tive of the predefined object, wherein the second predetermined feature model is specific for the second sensor, thereby generating second object probability data; and combine the first and the second object probability data, thereby generating combined prob ability data for detecting the predefined object. According to a second aspect, the disclosure provides an object detection method comprising: obtaining first feature data which are based on first sensing data of a first sensor; comparing the first feature data to a first predetermined feature model being representative of a predefined object, wherein the first predetermined feature model is specific for the first sensor, thereby generating first object probability data; obtaining second feature data which are based on second sensing data of a second sensor; comparing the second feature data to a second predetermined feature model being repre sentative of the predefined object, wherein the second predetermined feature model is specific for the second sensor, thereby generating second object probability data; and combining the first and the second object probability data, thereby generating combined probability data for detecting the predefined object.
Further aspects are set forth in the dependent claims, the following description and the drawings.
BRIEF DESCRIPTION OF THE DRAWINGS
Embodiments are explained by way of example with respect to the accompanying drawings, in which:
Fig. 1 depicts a coordinate system including a signal distribution acquired by a stereo camera;
Fig. 2 depicts a coordinate system including a signal distribution acquired by a radar sensor;
Fig. 3 depicts a template (feature model) for a stereo camera;
Fig. 4 depicts a template (feature model) for a radar sensor;
Fig. 5 depicts a fused object pose probability distribution;
Fig. 6 depicts a coordinate system including detected objects based on feature maps of the stereo camera and the radar sensor;
Fig. 7 depicts a block diagram of a method according to the present disclosure;
Fig. 8 schematically illustrates a method for generating a feature map according to the present disclo sure;
Fig. 9 schematically illustrates a method for determining object probability data according to the pre sent disclosure; and
Fig. 10 depicts a block diagram of a vehicle including an object detection circuitry according to the present disclosure. DETAILED DESCRIPTION OF EMBODIMENTS
Before a detailed description of the embodiments under reference of Fig. 1 is given, general explana tions are made.
As mentioned in the outset, known ways of detecting an object may lead to a high processing com plexity, and thereby to a high power consumption and an increase in costs.
It has been recognized that it is generally desirable to reduce a power consumption and costs.
Moreover, it has been recognized, in order to provide (semi-)autonomous driving, a full and reliable understanding of the environment may be necessary, e.g. for a planning of a route, an estimation of a danger, and the like.
Therefore, it is desirable to provide a way to use outputs of several sensors and the fusion of their measurements while at the same time providing a reliable and fast way to detect an object from a predefined list of objects.
It has further been recognized that several demands may be made to a system or a circuitry for providing such a reliable and fast way to detect an object.
One requirement may be abstraction. For example, a low-level sensor measurement (e.g. a distance) may not be suitable for planning purposes (e.g. of a route), such that information which do not con tribute to an object detection may need to be filtered out and relevant information may be bundled in a representation of an environment, wherein the representation may fulfill requirements of com pactness, completeness and preciseness, which may, for example, be a list of objects including pa rameters indicating their position. Moreover, a sensor-specific appearance of the objects must be taken into account (e.g. an object may be represented differendy through a camera than through a radar).
A further requirement may be sensor fusion. Since different sensors may rely on different measure ment principles and output data of each sensor may be different, as well, there may be an increased effort to operate the sensors without increasing a system complexity.
A further requirement may be a fusion of all available measurements. It has been recognized that known devices may translate low-level measurements into high-level abstract representations for each sensor and then fuse the obtained information into a joint representation. However, infor mation, which is discarded in such an abstraction, may not be available during or after the fusion an ymore, such that a quality may be reduced. Therefore, it is desirable to at least reduce an amount of a quality loss. A further requirement may be an overcoming of a view-dependent object appearance. It has been recognized that a predefined object may vary in a sensor appearance, such that an object orientation may depend on a sensor viewing direction (i.e. a viewing angle). For example, there may be a self occlusion effect of the object, such that a sensor may only observe a part of an object. The observed part may appear differently depending of a viewing angle of a sensor. Therefore, it is generally desir able to overcome such an effect for reliably detecting the object.
A further requirement may be in-class object invariance. Different parts of the same object may be considered to belong to different objects. For example, an object may have different shapes and/or colors at different spots, such that a different shape may be recognized as a different object. How ever, it is generally desirable to provide a way to detect an object as one object although there are different shapes and/or colors at the object.
A further requirement may be multi-object detection. For example, it may be necessary to detect more than one object (e.g. the closest one), such as detecting all objects, which are within a detection range.
A further requirement may be sensor noise. In order to avoid a reduction in a measurement quality, it is desirable to provide a way for compensating for (e.g. filtering) sensor noise and/ or a measure ment imperfection without reducing a quality.
Thus, some embodiments pertain to an object detection circuitry configured to: obtain first feature data which are based on first sensing data of a first sensor; compare the first feature data to a first predetermined feature model being representative of a predefined object, wherein the first predeter mined feature model is specific for the first sensor, thereby generating first object probability data; obtain second feature data which based on second sensing data of a second sensor; compare the sec ond feature data to a second predetermined feature model being representative of the predefined object, wherein the second predetermined feature model is specific for the second sensor, thereby generating second object probability data; and combine the first and the second object probability data, thereby generating combined probability data for detecting the predefined object.
The object detection circuitry may include one or multiple processors, such as a CPU (central pro cessing unit), GPU (graphic processing units), one or multiple FPGAs (field programmable gate ar ray), a computer, a server, and the like, also combinations of multiple of the named elements, and the like.
The object detection circuitry may further be included in or associated to an automotive environ ment, such that a detection of objects may be performed according to an operational state of, for example, a car, without limiting the present disclosure in that regard, since the object detection cir cuitry may be applied in any field, in which sensor data are utilized for an object detection, such as (surveillance) camera systems, and the like.
The object detection circuitry may be configured to obtain feature data, i.e. the feature data may be provided to the object detection circuitry by at least one sensor to which the object detection cir cuitry may be coupled, connected, associated, and the like. In some embodiments, the at least one sensor may be included in the object detection circuitry, such that the feature data may be acquired within the object detection circuitry.
In some embodiments, feature data, which are obtained by multiple sensors may be (structurally or logically) distinguished, i.e. feature data from a first sensor may be referred to as first feature data, feature data from a second sensor may be referred to as second feature data, and so on.
In some embodiments, the first feature data may be based on first sensing data. For example, the first sensor may perform a measurement according to an internal configuration; for example it may detect a reflection of a radio wave (e.g. in the case of a radar sensor), such that, for example, an ana log or digital sensor output may be generated, to which it may be referred to as first sensing data.
The first feature data may include an evaluation result of the first sensing data, for example a signal intensity according to a distance in a predefined spatial resolution, such that a feature map, as it is generally known, can be generated.
In some embodiments, the first sensor may be configured to generate the first feature data, whereas in some embodiments, the first feature data are generated in the object detection circuitry.
By obtaining the feature data, a requirement of abstraction, as discussed above, may fulfilled. Thereby, a detection list of objects including the predefined object, including a position, an orienta tion, and a correction estimate may be provided, such that a further pose refinement and/ or a track ing of the object may be carried out.
Moreover, by obtaining feature data multiple times, a requirement of a fusion of all available meas urements, as discussed above, may be fulfilled.
The object detection circuitry may be further configured to compare the first feature data to a first predefined model.
The comparison may be performed by performing a mapping between the first feature data and model feature data (e.g. comparing the generated feature map with a predefined feature map), and detecting common and/ or comparable signals. Hence, the feature model may include data indicative of a predefined feature map, which may be specific for the first sensor in a way that the first feature data may be comparable to the feature model, i.e. a radar feature model may be compared to feature data obtained based on first sensing data of a radar sensor, and the like, without limiting the present disclosure to a radar sensor. For ex ample, also a camera (RGB, infrared, time-of-flight), a lidar sensor, a time-of-flight sensor may con stitute the first sensor, or the first sensing data may be fused sensing data of multiple sensors, such as a fused signal of an RGB and a time-of-flight camera, and the like.
The feature model may be representative of a predefined object. For example, in an automotive en vironment, a predefined object may be a car, a pedestrian, a street sign, and the like, which may be indicated by a specific signal signature, i.e. object specific sensing data may be generated in response to a detection of the predefined object.
The object specific sensing data may be predefined, acquired in a calibration process, a training pro cess, and the like, of the object detection circuitry, such that an assignment between the predefined object and the first feature data can be performed based on the comparison between the first feature data and the first predetermined feature model.
By performing the comparison between the first feature data and the first predetermined feature model, first object probability data may be generated, which may indicate a correspondence between the first feature data and the first predetermined feature model. In other words, it may be deter mined to which extent the first feature data may correspond to the predetermined object or how probable it is that the first feature data indicate the predetermined object.
A similar process may be performed to second feature data based on second sensing data of a sec ond sensor, such that a repetitive description thereof is omitted.
However, it should be noted that, generally, the first and the second sensor may be same or different sensors, e.g. the first sensor may be a radar, the second sensor may be an RGB camera, or the like.
Moreover, the present disclosure is not limiting in that the second sensor is configured to work the same way as the first sensor. For example, in the first sensor the first feature data may be generated based on the first sensing data, whereas in the second sensor only the second sensing data may be acquired and the second feature data may be generated in the object detection circuitry, or vice versa.
The object detection circuitry may be further configured to combine the first and the second object probability data.
Thereby, a requirement of a sensor fusion, as discussed above, may be fulfilled. The combination may be based on a multiplexing, multiplying, adding, and the like, of the first and the second object probability data, which may result in a probability value, but, in some embodi ments, the combination may be based on an abstract data structure, which may be generated by an artificial intelligence, a neural network, based on a supervised or unsupervised training, and the like.
The combination procedure may be followed by a normalization of the combined probability data, as it is generally known.
The predefined object may be detected based on the combined probability data, i.e. a position in a space of the object may be determined. Moreover, a posture of the predefined object may be deter mined, since the first and the second feature data may be acquired for a plurality of locations with a (sensor specific) spatial resolution.
In some embodiments, the object detection circuitry is further configured to: detect a plurality of maxima of the combined probability data being indicative for at least one position parameter of the predefined object; and determine the at least one position parameter of the predefined object.
The combined probability data may represent a distribution of probabilities of features (or occur rences) of the predefined object in a space, such that a high probability (i.e. a probability above a predetermined threshold) may be a maximum and a probability below a predetermined threshold may be a minimum, such that probabilities above the predetermined threshold of the combined probability data may be taken into account for determining the at least one position parameter of the predefined object.
In some embodiments, a maximum may indicate a possibility (or a hypothesis) of an object (e.g. a single or a plurality) at a predetermined location, such that a (symbolic) position of the maximum in a probability space may indicate a (real) spatial position and orientation of the object.
In some embodiments, the detection of a plurality (i.e. at least two) of maxima may be based on known maxima detection processes, e.g. if a maximum is defined to have a Gaussian distribution (or any other distribution, such as Lorentzian, Dirac, Fermi, Poisson, etc.), a Gaussian distribution (or any other) may be detected, as it is generally known.
The plurality of maxima may be detected to be arranged in a spatial order (or pattern), such that a position parameter may be determined.
For example, in a case, in which the predefined object is a car, two maxima may be detected, which are aligned on a horizontal line of a (virtual) coordinate system, wherein the two maxima are two meters apart from each other. It may be a premise that the first sensor and the second sensor may be aligned on a horizontal axis, which may be parallel to the horizontal line. In such a case, it may be detected that a posture of the detected car is parallel to the horizontal axis of the sensors.
Moreover, if the sensors are implemented in a car including the object detection circuitry according to the present disclosure, it may be inferred that the detected car may be perpendicular to the car in cluding the object detection circuitry, such that, for example, a warning may be issued to a driver of the car including the object detection circuitry, or an emergency braking may be issued, and the like, without limiting the present disclosure in that regard.
On the other hand, if two maxima may be detected on a line inclined from the horizontal sensor axis, a posture and/ or a positional angle of the detected car may be determined.
However, the present disclosure is not limited to the position parameter being a posture or an angle, since, in some embodiments, the at least one position parameter includes at least one of a position, a distance, an angle, and a posture.
Hence, a position may be determined and may be represented in a (virtual) coordinate system, which may also be shown to a user of the object detection circuitry on a display, and the like.
A distance may be concluded from the position, without limiting the present disclosure in that re gard, as the distance may be direcdy determined out of the combined probability data.
The posture and/or the angle may further include a direction of the predefined object. For example, if a car is the predefined object, it may be determined where a front and/ or where a back of the car is.
In some embodiments, the object detection circuitry is further configured to determine at least one of a correctness and a precision of the detection of the predefined object.
For example, a measurement may be deteriorated, e.g. by environmental influences (e.g. a tempera ture, a light condition, and the like), a deteriorated calibration of one (or both) of the sensors, due to a movement of the sensors (e.g. if implemented in a car) or of the predefined object (e.g. due to Doppler effect).
Therefore, a precision of the detection may be determined based on an assumed error (function), which may take into account any of the conditions mentioned above, without limiting the present disclosure in that regard.
Moreover, a correctness of the detection may be determined (e.g. based on the precision), which may indicate whether the predefined object corresponds to the detected object.
In some embodiments, the object detection circuitry is further configured to track the detected pre defined object. For example, known object tracking technologies may be applied.
However, in some embodiments the object detection circuitry may perform the same processing for the tracking again as it is applied for the detection of the predefined object, such that a tracking may correspond to a repetitive detecting of the predefined object.
In some embodiments, the object detection circuitry is further configured to: generate a first feature map based on the first sensing data; generate a second feature map based on the second sensing data; and transfer the first and the second feature map into a predefined coordinate system.
A feature map may be a map (e.g. based on the first and/ or the sensing data) depicting the obtained first and second feature data, either in one map or in two maps, for example. The feature map in cludes, in some embodiments, the detected plurality of maxima, and may be represented by a coor dinate system from a bird’s eye perspective, without limiting the present disclosure in that regard.
Moreover, a feature map may be generated for different layers of a space, such that data indicating a vertical direction may be acquired.
Generally, a feature represented in the feature map may be a detected signal (or one maximum or more of the plurality of maxima), which is indicative of a specific property of the predefined object. For example, a feature may be indicated by a detected signal stemming from a specific part of a car, e.g. a window, a back, and the like, which has a specific feature signature.
The predefined coordinate system may be a common coordinate system of the first and the second feature map or it may be different coordinate systems, which, however, may be convertible into each other, and may also be transferred into a common coordinate system.
The predefined coordinate system is not limited to a particular kind of coordinate systems as it may represent Euclidean coordinate, spherical coordinates, cylindrical coordinates, polar coordinates, and the like, and the origin of the predefined coordinate system may, for example be a particular location of one of the two sensors, a detected object, or any other location in a space.
In some embodiments, the first and the second sensor include at least one of a radar sensor, a lidar sensor, a camera, or a time-of-flight sensor, as discussed herein.
In some embodiments, at least one of the first predetermined feature model and the second prede termined feature model is based on a supervised training of an artificial intelligence.
The artificial intelligence (AI) may use machine learning based methods or explicit feature based methods, such as shape matching, for example by edge detection, histogram based methods, tem plate match based methods, color match based methods, or the like. In some embodiments, a ma chine learning algorithm may be used for performing object recognition, e.g. for comparing a detected predefined object with a recognized object to increase a correctness of a detection, which may be based on at least one of: Scale Invariant Feature Transfer (SIFT , Gray Level Co-occurrence Matrix (GLCM), Gabor Features, Tubeness, or the like. Moreover, the machine learning algorithm may be based on a classifier technique, wherein such a machine learning algorithm may be based on least one of: Random Forest; Support Vector Machine; Neural Net, Bayes Net, or the like. Further more, the machine learning algorithm may apply deep-learning techniques, wherein such deep-leam- ing techniques may be based on at least one of: Autoencoders, Generative Adversarial Network, weakly supervised learning, boot-strapping, or the like.
The supervised learning may further be based on a regression algorithm, a perception algorithm, Bayes-classification, Naiver Bayer classification, next-neighbor classification, artificial neural net work, and the like.
The artificial intelligence may, in such embodiments, be fed with ground truth data, which may cor respond to or be based on the predefined object, such that the artificial intelligence may learn to as sign the first and/ or the second feature data to the ground truth data, thereby developing or generating the first and/ or the second feature model.
In some embodiments, the predefined object is based on a class of predefined objects.
As discussed above, the disclosure may pertain to an object detection in an automotive field, such that a class of predefined objects may, in such embodiments, include a car, a passenger, a road sign, a street, a house, a tree, an animal, a traffic light, and the like.
However, in other embodiments, the object detection circuitry may be applied to a surveillance sys tem, for example of a warehouse, or the like, such that the class of predefined objects may include a customer, an employee, a shelf, a product, and the like, without limiting the present disclosure to embodiments pertaining to automotive or surveillance applications. Generally, the present disclosure may be applied in any field, in which an object detection may be performed.
In some embodiments, the object detection circuitry is further configured to: iteratively convolve the first feature data with the first predetermined feature model, thereby generating the first object probability data; iteratively convolve the second feature data with the second predetermined feature model, thereby generating the second object probability data; and iteratively convolve the first and the second object probability data, thereby generating the combined object probability data.
For example, a plurality of first feature data may be acquired (e.g. for different heights or layers, for different angles, and the like), which may each be convolved (e.g. compared, as discussed above) with the first predetermined feature model in a consecutive ordering (i.e. iteratively), to which it may in this disclosure be referred to as vertical iteration. Moreover, a plurality of second feature data may be acquired, which may each be convolved with the second predetermined feature model in a con secutive ordering, as well in a vertical iteration.
Moreover, first generated first object probability data (e.g. from a first measurement of the first sen sor) may be convolved with the first generated second object probability data (e.g. by summing up the respective probabilities), to which it may be referred to as horizontal iteration, whereby the com bined probability data are generated.
Some embodiments pertain to an object detection method including: obtaining first feature data which are based on first sensing data of a first sensor; comparing the first feature data to a first pre determined feature model being representative of a predefined object, wherein the first predeter mined feature model is specific for the first sensor, thereby generating first object probability data; obtaining second feature data which based on second sensing data of a second sensor; comparing the second feature data to a second predetermined feature model being representative of the prede fined object, wherein the second predetermined feature model is specific for the second sensor, thereby generating second object probability data; and combining the first and the second object probability data, thereby generating combined probability data for detecting the predefined object, as discussed herein.
The object detection method may be performed with an object detection circuitry according to the present disclosure, and the like, such as a processor (or multiple processors) configured to perform the object detection method.
In some embodiments, the method further includes: detecting a plurality of maxima of the com bined probability data being indicative for at least one position parameter of the predefined object; and determining the at least one position parameter of the predefined object, as discussed herein. In some embodiments, the at least one position parameter includes at least one of a position, a dis tance, an angle, and a posture, as discussed herein. In some embodiments, the method further in cludes: determining at least one of a correctness and a precision of the detection of the predefined object, as discussed herein. In some embodiment, the method further includes tracking the detected predefined object, as discussed herein. In some embodiments, the method further includes: generat ing a first feature map based on the first sensing data; generating a second feature map based on the second sensing data; and transferring the first and the second feature map into a predefined coordi nate system, as discussed herein. In some embodiments, the first and the second sensor include at least one of a radar sensor, a lidar sensor, a camera, or a time-of-flight sensor, as discussed herein. In some embodiments, at least one of the first predetermined feature model and the second predeter mined feature model is based on a supervised training of an artificial intelligence, as discussed herein. In some embodiments, the predefined object is based on a class of predefined objects, as dis cussed herein. In some embodiments, the method further includes: iteratively convolving the first feature data with the first predetermined feature model, thereby generating the first object probabil ity data; iteratively convolving the second feature data with the second predetermined feature model, thereby generating the second object probability data; and iteratively convolving the first and the second convolved feature data, thereby generating the combined probability data, as discussed herein.
The methods as described herein are also implemented in some embodiments as a computer pro gram causing a computer and/ or a processor to perform the method, when being carried out on the computer and/ or processor. In some embodiments, also a non-transitory computer-readable record ing medium is provided that stores therein a computer program product, which, when executed by a processor, such as the processor described above, causes the methods described herein to be per formed.
Generally, each sensor (i.e. the first and the second sensor, but not limited to two sensors) may pro vide a (predetermined) number of measurements M, wherein in response to each measurement f a feature may be assigned to two coordinates lx and lz (or short 1 = (h, lz)) of a two-dimensional coordi nate system, such that a feature map (or a plurality of feature maps, each measurement being indica tive of a feature map), as discussed herein, may be generated.
Each feature map may have a number of i grid cells, wherein the grid cells may have an identical or a different size and each grid cell may include two values representing the number of measurements detecting a particular feature n, which may be displayed as a function, such as n(f = 1 | 1, i) and a number of measurements not detecting the particular feature, i.e. n(f = 0 | 1, i).
Based on all or a subset of the obtained feature maps, a probability p(0(c, Q) = 1 | fi, ..., fist) of an existence of an object in an object parameter space may be determined, wherein O may be the ob ject, c may correspond to coordinates (cx, cz) of the object, Q may be indicative of the posture of the object (i.e. an angle), (c, Q) being the (three-dimensional) parameter space.
In particular, cx may include a center of the object in an x-direction, cz may include the center of the object in a z-direction and Q may include an orientation of the object in an x-z-plane, which may be limited to an interval of values, such as [0, 2p).
Thereby, a requirement of overcoming a view-dependent object appearance, as discussed above, may be fulfilled.
The probability p may be determined according to the following formula: In some embodiments, a discretized representation of the parameter space is generated, which may be represented by a three-dimensional array including a number of discretization points along the x- axis times a number of discretization points along the z-axis times a number of discretization points along an orientation axis (i.e. the angle).
Moreover, in some embodiments, a logarithmic value of p(0(c, Q) = 1 | fi, ..., f¾i) may be com puted, which may reduce a number of inference steps to a number of convolutions.
In some embodiments, the terms p(0(c, Q) = 1 | f,(l) =1) and p(0(c, Q) = 1 | fi(l) =0) are imple mented shift-invariant in 1, such that p(0(c, Q) = 1 | f(l) =1) = p(0(c - 1, Q) = 1 | fi(0) = 1) and p(0(c, Q) = 1 I fi(l) =0) = p(0(c - 1, Q) = 1 I fi(0) = 0).
Thereby, a requirement of an in-class object invariance, as discussed above, may be fulfilled.
Hence, in such embodiments the terms may be indicative of the object being located at a position with an orientation (c - 1; Q), if a feature is detected at 1 (i.e. f = 1) or not detected at 1 (i.e. f = 0), re spectively.
In some embodiments, the terms p(0(c - 1, Q) = 1 | fi(0) = 1) and p(0(c - 1, Q) = 1 | fi(0) = 0) are determined based on a supervised training from a real or simulated sensor measurement, which, in some embodiments may be noisy, of a predetermined number of predetermined objects with prede termined postures.
Thereby, sensor noise may be recognized and/ or filtered out, such that a requirement of sensor noise, as discussed above, may be fulfilled.
In some embodiments, local maxima, i.e. p(0(c - 1, Q) = 1 | fi(0) = 1) are located in the parameter space while excluding other parameters based on a non-maximum suppression.
In some embodiments, a location (c, Q) of a local maximum may indicate a pose (e.g. a position and an orientation) of the predetermined object (or multiple predetermined objects), wherein a correct ness of the determined pose may be determined by evaluating p(0(c, Q) = 1 | fi, ..., fin) at the loca tion of the local maximum, wherein a correctness may be given, if p(0(c, Q) = 1 | fi, ..., fut) is above a predetermined threshold.
Thereby, a requirement of multi-object detection, as discussed above, may be fulfilled by providing a list of detection, wherein the precision of the pose may depend on a resolution of the pose space (c, 9). In some embodiments, a precision of the determined pose may be determined by evaluating a curva ture of p(0(c, Q) = 1 I fi, _ , f i) at the location of the local maximum, wherein a higher precision may correspond to a higher value of the curvature (in other embodiments, a higher precision may correspond to a lower value of the curvature).
A detection of the pose (or posture) may be refined in a consecutive measurement based on the cor rectness and/ or the precision of the measurement.
Returning to Fig. 1, there is depicted a first coordinate system 1 including two axes 2a and 2b each indicating positional coordinates.
Moreover, three signal distributions 3a, 3b and 3c (first feature data) are shown, which are obtained based on first sensing data by a stereo camera as a first sensor. Only the signal distribution 3a is dis cussed herein, in order to omit a repetitive discussion.
The signal distribution 3a has a first sub-signal 4 and a second sub-signal 5, wherein the first sub signal 4 has a stronger signal intensity than the second sub-signal 5.
Fig. 2 depicts a second coordinate system 10 including two axes 12a and 12b, which correspond to the axes 2a and 2b Fig. 1.
Moreover three signal distributions 13a, 13b and 13c (second feature data) are shown, which are ob tained based on second sensing data by a radar sensor as a second sensor.
The signal distribution 13a (and correspondingly 13b and 13c although not explicitly discussed again) includes first sub-signals 14, second sub-signals 15, and third sub-signals 16, which are, for illustrative reasons, grouped by different patterns, wherein the pattern strength corresponds to the respective signal strength, i.e. the first sub-signals 14 have the strongest signal intensity of the first to third sub-signals 14 to 16, the second sub-signals 15 have the second strongest signal intensity of the first to third sub-signals 14 to 16, and the third sub-signals 16 have the third strongest signal inten sity of the first to third sub-signals 14 to 16.
In the currently described embodiment, the signal distributions 3a, 3b, 3c, 13a, 13b, and 13c de scribed with respect to Figs. 1 and 2 are obtained by acquiring first and second sensing data of three cars as predefined objects, i.e. the signal distributions 3a to 3c and 13a to 13c are indicative of cars.
Fig. 3 depicts a template (i.e. a feature model) 20 for a stereo camera as a sensor, which is based on a supervised training of an artificial intelligence.
The template 20 represents different postures of cars 24 to 39, wherein each posture is assigned to a probability, as discussed herein. Moreover, different intensities of the signals are symbolically depicted with different lines 21 to 23, wherein the line 21 represents a strongest intensity, the line 22 the second strongest, and the line 23 the weakest detected intensity, without limiting the present disclosure in that regard, since, as it is generally known, a continuous distribution of intensities may be detected, as well. The assignment is displayed in the following table, wherein each reference sign (corresponding to a posture as depicted) is assigned to a probability and the posture is represented by an angle Q.
Figure imgf000016_0001
Fig. 4 depicts a template (i.e. a feature model) 40 for a radar as a sensor, which is based on a super vised training of an artificial intelligence. The template 40 represents different postures of cars, which correspond to the ordering (in terms of probability and angle) of the template 20 of Fig. 3, and, therefore, a repetitive description (together with reference signs) is omitted.
Different intensities of the signals are symbolically depicted with differently hatched ellipses 41 to 43, wherein the ellipse 41 represents a strongest intensity, the ellipse 42 the second strongest, and the ellipse 43 the weakest detected intensity, without limiting the present disclosure in that regard, since, as it is generally known, a continuous distribution of intensities may be detected, as well.
The logarithmic probabilities of object poses is determined using templates 20 and 40, as discussed herein, and the object pose probability distributions are fused, such that a fused pose probability dis tribution (or a fused feature model) is generated.
Fig. 5 depicts such a fused pose probability distribution 50 with the same ordering of postures of the car as in Figs. 3 and 4, such that a repetitive description (together with reference signs) is omitted.
Different intensities of the signals are symbolically depicted with differently hatched rectangles 51 to 53, wherein the rectangle 51 represents a strongest intensity, the rectangle 52 the second strongest, and the rectangle 53 the weakest detected intensity, without limiting the present disclosure in that regard, since, as it is generally known, a continuous distribution of intensities may be detected, as well.
The logarithmic probability of the first and the second feature data (of in Figs. 1 and 2) are fused into a combined pose probability distribution 50 for each of the three objects (cars) of Figs. 1 and 2, such that poses of the three cars.
A result of the comparison is displayed in Fig. 6
Fig. 6 depicts a coordinate system 60 including two axes 61a and 61b each indicating positional co ordinates.
Moreover, three detected objects 62a, 62b and 62c are shown, overlaid with summed up features 63 of the feature maps 1 and 10 of Figs. 1 and 2, respectively. The display of the three detected objects further include pointers 64a, 64b, and 64c indicating a posture of each detected object 62a, 62b, and 62c, respectively.
Fig. 7 depicts a block diagram of a method 70 according to the present disclosure.
In 71, a sensor measurement is performed. In this embodiment, a stereo camera image pair as first sensing raw data is acquired in 711 and radar data as second sensing raw data is acquired in 712.
With each of the sensing raw data, a low level processing is performed.
In the case of the stereo camera image pair, a calibration and a disparity estimation is performed in
713. In the case of the radar data, a calibration and a Fast Fourier Transformation is performed in
714.
In 72, the low-level processed first and second sensing raw data are transferred into respective fea ture maps, such that the low-level processed first sensing raw data are transferred into a feature map in bird’s eye view in 721, and the low-level processed second sensing raw data are transferred into a feature map in bird’s eye view of 722.
In 73, a logarithmic probability, as discussed herein, is determined from the feature maps.
In particular, in 731, a pose logarithmic probability volume of the stereo camera feature map is de termined based on previously trained conditional probabilities, which are being fed to the determina tion of the pose logarithmic probability volume in 732.
Moreover, in 733, a pose logarithmic probability volume of the radar sensor is determined based on previously trained conditional probabilities, which are being fed to the determination of the pose logarithmic probability volume in 734.
The determined pose logarithmic probability volumes are combined (or fused) in 735, as discussed herein.
In 74, local maxima are determined and processed, as discussed herein.
Based on the combined probability, maxima are detected with a non-maxima suppression algorithm in 741, as discussed herein.
In 742, a pose, correctness and a precision is determined based on the detected maxima, as discussed herein.
In 75, further processing is performed based on the determined poses (i.e. position and orientation), correctness, and precision, as discussed herein.
In particular, a further refinement of the detected poses is performed in 751, such that the detected object (or detected objects) is tracked in 752, as discussed herein.
Fig. 8 schematically illustrates a method 80 for generating a feature map.
A coordinate system 81 including a plurality of cells 82 being defined by coordinates 1K and 12 (or (lx, lz)) is depicted.
Moreover, the coordinate system 81 includes a plurality of detections 83, which are overlapping with borders of the cells 82, without limiting the present disclosure in that regard, since a detection 83 may generally be fully inside a cell 82.
The detections 83 are assigned to one of the cells 82, in which most of the detection signal lies, i.e. a cell-wise accumulation of the detections 83 is performed in 84.
A coordinate system 85 is depicted, which basically corresponds to the coordinate system 81, such that a repetitive description is omitted. However, the coordinate system 85 is different from the coordinate system 81 in that the detections 83 are accumulated to features 86.
For each cell corresponding to a specific coordinate, a number of detections is determined, such that an empty cell, for example the cell (4, 3) can be described as n(f = 1 11 = ((4, 3), i) = 0, a cell with one detection or feature, for example the cells (5, 16), (23, 13), and (17, 16) can be described as n(f = 1 11 = ((5, 16), i) = 1, n(f = 1 11 = ((23, 13), i) = 1, and n(f = 1 11 = ((16, 17), i) = 1, respectively, and a cell with two detections or features, for example the cell (19, 5) can be described as n(f = 1 11 = ((19, 5), i) = 2, without limiting the present disclosure to a maximum number of to de tections in one cell.
Fig. 9 schematically illustrates a method 90 for determining object probability data according to the present disclosure.
The following description of the method 90 only takes into account a measurement of two sensors 91 and 91’. However, as the skilled person may take from Fig. 9, the disclosure is not limited to two sensors, as it may be envisaged for an arbitrary number of sensors with an arbitrary number of measurements.
First feature data, which are based on a sensor measurement of a sensor 91 are convolved with a predetermined feature model 92 (i.e. a predetermined conditional probability based on a supervised training being specific for the sensor 91) in 93.
Second feature data, which are based on a sensor measurement of a sensor 91’ are convolved with a predetermined feature model 92’ (i.e. a predetermined conditional probability based on a supervised training being specific for the sensor 91’) in 93’.
Moreover, a sum 94 of the convolved data of 93 and the convolved data of 93’ is determined, which serves as a basis for a further convolution with further sensing data for detecting an object, or, in some embodiments, as a basis for detecting the object.
Fig. 10 depicts a block diagram of a vehicle 100, which includes a CPU 101, which is configured to function as an object detection circuitry according to the present disclosure.
The vehicle 100 further includes a stereo camera 102 for acquiring first sensing data and a radar 103 for acquiring second sensing data, which are fed to the CPU 101, such that first and second feature data are generated.
However, in some embodiments, the first and the second feature data are generated in the first and the second sensor or in specific circuitry for generating the first and/ or the second feature data, as discussed herein. For example, a feature map may be computed in each sensor and may be transferred to a fusion- and-detection unit, which may be configured to fuse the feature maps, determine the probabilities and detect the object.
It should be recognized that the embodiments describe methods with an exemplary ordering of method steps. The specific ordering of method steps is however given for illustrative purposes only and should not be construed as binding. For example the ordering of 91 and 91’ in the embodiment of Fig. 9 may be exchanged. Also, the ordering of 711 and 712 in the embodiment of Fig. 7 may be exchanged. Further, also the ordering of 721 and 722 in the embodiment of Fig. 7 may be ex changed. Other changes of the ordering of method steps may be apparent to the skilled person.
In some embodiments, also a non-transitory computer-readable recording medium is provided that stores therein a computer program product, which, when executed by a processor, such as the pro cessor described above, causes the method described to be performed.
All units and entities described in this specification and claimed in the appended claims can, if not stated otherwise, be implemented as integrated circuit logic, for example on a chip, and functionality provided by such units and entities can, if not stated otherwise, be implemented by software.
In so far as the embodiments of the disclosure described above are implemented, at least in part, us ing software-controlled data processing apparatus, it will be appreciated that a computer program providing such software control and a transmission, storage or other medium by which such a com puter program is provided are envisaged as aspects of the present disclosure.
It should, moreover, be noted that the present disclosure, although exemplarily described for a de tection of cars and an implementation in a vehicle, is not limited to the described embodiments. In particular, the present disclosure may be applied in each case or situation, in which a surrounding scenario can be represented in a two-dimensional coordinate system, e.g. in any land-based or sea- surface based navigation system, such as driver assistance systems in cars, autonomous vehicles, ro bots, boats, ships, and the like.
Note that the present technology can also be configured as described below.
(1) An object detection circuitry configured to: obtain first feature data which are based on first sensing data of a first sensor; compare the first feature data to a first predetermined feature model being representative of a predefined object, wherein the first predetermined feature model is specific for the first sensor, thereby generating first object probability data; obtain second feature data which are based on second sensing data of a second sensor; compare the second feature data to a second predetermined feature model being representa tive of the predefined object, wherein the second predetermined feature model is specific for the second sensor, thereby generating second object probability data; and combine the first and the second object probability data, thereby generating combined prob ability data for detecting the predefined object.
(2) The object detection circuitry of (1), further configured to: detect a plurality of maxima of the combined probability data being indicative for at least one position parameter of the predefined object; and determine the at least one position parameter of the predefined object.
(3) The object detection circuitry of anyone of (1) or (2), wherein the at least one position pa rameter includes at least one of a position, a distance, an angle, and a posture.
(4) The object detection circuitry of anyone of (1) to (3), further configured to: determine at least one of a correctness and a precision of the detection of the predefined ob ject.
(5) The object detection circuitry of anyone of (1) to (4), further configured to: track the detected predefined object.
(6) The object detection circuitry of anyone of (1) to (5), further configured to: generate a first feature map based on the first sensing data; generate a second feature map based on the second sensing data; and transfer the first and the second feature map into a predefined coordinate system.
(7) The object detection circuitry of anyone of (1) to (6), wherein the first and the second sensor include at least one of a radar sensor, a lidar sensor, a camera, or a time-of-flight sensor.
(8) The object detection circuitry of anyone of (1) to (7), wherein at least one of the first prede termined feature model and the second predetermined feature model is based on a supervised train ing of an artificial intelligence.
(9) The object detection circuitry of anyone of (1) to (8), wherein the predefined object is based on a class of predefined objects.
(10) The object detection circuitry of anyone of (1) to (9), further configured to: iteratively convolve the first feature data with the first predetermined feature model, thereby generating the first object probability data; iteratively convolve the second feature data with the second predetermined feature model, thereby generating the second object probability data; and iteratively convolve the first and the object probability data, thereby generating the combined object probability data.
(11) An object detection method comprising: obtaining first feature data which are based on first sensing data of a first sensor; comparing the first feature data to a first predetermined feature model being representative of a predefined object, wherein the first predetermined feature model is specific for the first sensor, thereby generating first object probability data; obtaining second feature data which are based on second sensing data of a second sensor; comparing the second feature data to a second predetermined feature model being repre sentative of the predefined object, wherein the second predetermined feature model is specific for the second sensor, thereby generating second object probability data; and combining the first and the second object probability data, thereby generating combined probability data for detecting the predefined object.
(12) The object detection method of (11), further comprising: detecting a plurality of maxima of the combined probability data being indicative for at least one position parameter of the predefined object; and determining the at least one position parameter of the predefined object.
(13) The object detection method of anyone of (11) and (12), wherein the at least one position parameter includes at least one of a position, a distance, an angle, and a posture.
(14) The object detection method of anyone of (11) to (13), further comprising: determining at least one of a correctness and a precision of the detection of the predefined object.
(15) The object detection method of anyone of (11) to (14), further comprising: tracking the detected predefined object.
(16) The object detection method of anyone of (11) to (15), further comprising: generating a first feature map based on the first sensing data; generating a second feature map based on the second sensing data; and transferring the first and the second feature map into a predefined coordinate system.
(17) The object detection method of anyone of (11) to (16), wherein the first and the second sen sor include at least one of a radar sensor, a lidar sensor, a camera, or a time-of-flight sensor.
(18) The object detection method of anyone of (11) to (17), wherein at least one of the first pre determined feature model and the second predetermined feature model is based on a supervised training of an artificial intelligence. (19) The object detection method of anyone of (11) to (18), wherein the predefined object is based on a class of predefined objects.
(20) The object detection method of anyone of (11) to (19), further comprising: iteratively convolving the first feature data with the first predetermined feature model, thereby generating the first object probability data; iteratively convolving the second feature data with the second predetermined feature model, thereby generating the second object probability data; and iteratively convolving the first and the second object probability data, thereby generating the combined probability data. (21) A computer program comprising program code causing a computer to perform the method according to anyone of (11) to (20), when being carried out on a computer.
(22) A non-transitory computer-readable recording medium that stores therein a computer pro gram product, which, when executed by a processor, causes the method according to anyone of (11) to (20) to be performed.

Claims

1. An object detection circuitry configured to: obtain first feature data which are based on first sensing data of a first sensor; compare the first feature data to a first predetermined feature model being representative of a predefined object, wherein the first predetermined feature model is specific for the first sensor, thereby generating first object probability data; obtain second feature data which are based on second sensing data of a second sensor; compare the second feature data to a second predetermined feature model being representa tive of the predefined object, wherein the second predetermined feature model is specific for the second sensor, thereby generating second object probability data; and combine the first and the second object probability data, thereby generating combined prob ability data for detecting the predefined object.
2. The object detection circuitry of claim 1, further configured to: detect a plurality of maxima of the combined probability data being indicative for at least one position parameter of the predefined object; and determine the at least one position parameter of the predefined object.
3. The object detection circuitry of claim 2, wherein the at least one position parameter in cludes at least one of a position, a distance, an angle, and a posture.
4. The object detection circuitry of claim 2, further configured to: determine at least one of a correctness and a precision of the detection of the predefined ob ject.
5. The object detection circuitry of claim 2, further configured to: track the detected predefined object.
6. The object detection circuitry of claim 1, further configured to: generate a first feature map based on the first sensing data; generate a second feature map based on the second sensing data; and transfer the first and the second feature map into a predefined coordinate system.
7. The object detection circuitry of claim 1, wherein the first and the second sensor include at least one of a radar sensor, a lidar sensor, a camera, or a time-of-flight sensor.
8. The object detection circuitry of claim 1, wherein at least one of the first predetermined fea ture model and the second predetermined feature model is based on a supervised training of an arti ficial intelligence.
9. The object detection circuitry of claim 1, wherein the predefined object is based on a class of predefined objects.
10. The object detection circuitry of claim 1, further configured to: iteratively convolve the first feature data with the first predetermined feature model, thereby generating the first object probability data; iteratively convolve the second feature data with the second predetermined feature model, thereby generating the second object probability data; and iteratively convolve the first and the object probability data, thereby generating the combined object probability data.
11. An object detection method comprising: obtaining first feature data which are based on first sensing data of a first sensor; comparing the first feature data to a first predetermined feature model being representative of a predefined object, wherein the first predetermined feature model is specific for the first sensor, thereby generating first object probability data; obtaining second feature data which are based on second sensing data of a second sensor; comparing the second feature data to a second predetermined feature model being repre sentative of the predefined object, wherein the second predetermined feature model is specific for the second sensor, thereby generating second object probability data; and combining the first and the second object probability data, thereby generating combined probability data for detecting the predefined object.
12. The object detection method of claim 11, further comprising: detecting a plurality of maxima of the combined probability data being indicative for at least one position parameter of the predefined object; and determining the at least one position parameter of the predefined object.
13. The object detection method of claim 12, wherein the at least one position parameter in cludes at least one of a position, a distance, an angle, and a posture.
14. The object detection method of claim 12, further comprising: determining at least one of a correctness and a precision of the detection of the predefined object.
15. The object detection method of claim 12, further comprising: tracking the detected predefined object.
16. The object detection method of claim 11, further comprising: generating a first feature map based on the first sensing data; generating a second feature map based on the second sensing data; and transferring the first and the second feature map into a predefined coordinate system.
17. The object detection method of claim 11, wherein the first and the second sensor include at least one of a radar sensor, a lidar sensor, a camera, or a time-of-flight sensor.
18. The object detection method of claim 11, wherein at least one of the first predetermined fea ture model and the second predetermined feature model is based on a supervised training of an arti ficial intelligence.
19. The object detection method of claim 11, wherein the predefined object is based on a class of predefined objects.
20. The object detection method of claim 11, further comprising: iteratively convolving the first feature data with the first predetermined feature model, thereby generating the first object probability data; iteratively convolving the second feature data with the second predetermined feature model, thereby generating the second object probability data; and iteratively convolving the first and the second object probability data, thereby generating the combined probability data.
PCT/EP2020/075417 2019-09-25 2020-09-10 Object detection circuitry and object detection method WO2021058300A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
EP20768050.5A EP4035059A1 (en) 2019-09-25 2020-09-10 Object detection circuitry and object detection method
CN202080065788.8A CN114424254A (en) 2019-09-25 2020-09-10 Object detection circuit and object detection method
US17/635,385 US20220406044A1 (en) 2019-09-25 2020-09-10 Object detection circuitry and object detection method
JP2022517741A JP7586173B2 (en) 2019-09-25 2020-09-10 OBJECT DETECTION CIRCUIT AND OBJECT DETECTION METHOD

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP19199664 2019-09-25
EP19199664.4 2019-09-25

Publications (1)

Publication Number Publication Date
WO2021058300A1 true WO2021058300A1 (en) 2021-04-01

Family

ID=68069601

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2020/075417 WO2021058300A1 (en) 2019-09-25 2020-09-10 Object detection circuitry and object detection method

Country Status (5)

Country Link
US (1) US20220406044A1 (en)
EP (1) EP4035059A1 (en)
JP (1) JP7586173B2 (en)
CN (1) CN114424254A (en)
WO (1) WO2021058300A1 (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9684836B1 (en) * 2013-07-31 2017-06-20 Waymo Llc Combining multiple estimates of an environment into a consolidated estimate for an autonomous vehicle
US20180068206A1 (en) * 2016-09-08 2018-03-08 Mentor Graphics Corporation Object recognition and classification using multiple sensor modalities

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9542626B2 (en) 2013-09-06 2017-01-10 Toyota Jidosha Kabushiki Kaisha Augmenting layer-based object detection with deep convolutional neural networks
US20160299229A1 (en) * 2015-04-09 2016-10-13 Sharper Shape Oy Method and system for detecting objects
CA2999697A1 (en) 2015-09-30 2017-04-06 Sony Corporation Information processing apparatus, information processing method and program
US10417816B2 (en) * 2017-06-16 2019-09-17 Nauto, Inc. System and method for digital environment reconstruction
US10733755B2 (en) * 2017-07-18 2020-08-04 Qualcomm Incorporated Learning geometric differentials for matching 3D models to objects in a 2D image
US10726275B2 (en) * 2018-03-07 2020-07-28 Visteon Global Technologies, Inc. System and method for correlating vehicular sensor data
EP3702802A1 (en) * 2019-03-01 2020-09-02 Aptiv Technologies Limited Method of multi-sensor data fusion
US11062454B1 (en) * 2019-04-16 2021-07-13 Zoox, Inc. Multi-modal sensor data association architecture

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9684836B1 (en) * 2013-07-31 2017-06-20 Waymo Llc Combining multiple estimates of an environment into a consolidated estimate for an autonomous vehicle
US20180068206A1 (en) * 2016-09-08 2018-03-08 Mentor Graphics Corporation Object recognition and classification using multiple sensor modalities

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
BECKER J C: "Fusion of data from the object-detecting sensors of an autonomous vehicle", INTELLIGENT TRANSPORTATION SYSTEMS, 1999. PROCEEDINGS. 1999 IEEE/IEEJ/ JSAI INTERNATIONAL CONFERENCE ON TOKYO, JAPAN 5-8 OCT. 1999, PISCATAWAY, NJ, USA,IEEE, US, 5 October 1999 (1999-10-05), pages 362 - 367, XP010369870, ISBN: 978-0-7803-4975-9, DOI: 10.1109/ITSC.1999.821082 *
MICHAEL AEBERHARD: "Object-Level Fusion for Surround Environment Perception in Automated Driving Applications", PHD THESIS, 31 May 2017 (2017-05-31), Faculty of Electrical Engineering and Information Technology at Technische Universit"at Dortmund, XP055663290, Retrieved from the Internet <URL:https://eldorado.tu-dortmund.de/bitstream/2003/36011/1/Aeberhard_Dissertation.pdf> [retrieved on 20200129] *
RICARDO OMAR ET AL: "Multiple Sensor Fusion for Detection, Classification and Tracking of Moving Objects in Driving Environments", THESIS, 25 September 2014 (2014-09-25), pages 1 - 176, XP055643862, Retrieved from the Internet <URL:https://www.theses.fr/2014GRENM034.pdf> [retrieved on 20191119] *
See also references of EP4035059A1 *

Also Published As

Publication number Publication date
JP2022550285A (en) 2022-12-01
CN114424254A (en) 2022-04-29
JP7586173B2 (en) 2024-11-19
EP4035059A1 (en) 2022-08-03
US20220406044A1 (en) 2022-12-22

Similar Documents

Publication Publication Date Title
WO2022012158A1 (en) Target determination method and target determination device
US7680300B2 (en) Visual object recognition and tracking
Zhu et al. Reliable detection of overtaking vehicles using robust information fusion
CN111913177A (en) Method and device for detecting target object and storage medium
CN113743385A (en) Unmanned ship water surface target detection method and device and unmanned ship
US12198381B2 (en) Hand pose estimation method, device and storage medium
CN110826450A (en) Automatic suspicious article detection method based on millimeter wave image
CN114970705A (en) Driving state analysis method, device, equipment and medium based on multi-sensing data
CN106296708B (en) Car tracing method and apparatus
CN110555339A (en) target detection method, system, device and storage medium
Kumar et al. Deep learning-based multi-object tracking
Long et al. Object detection research of SAR image using improved faster region-based convolutional neural network
CN116844124A (en) Three-dimensional target detection frame annotation method, device, electronic equipment and storage medium
WO2022083529A1 (en) Data processing method and apparatus
CN111507938B (en) Human body dangerous goods detection method and system
CN118397588A (en) Camera scene analysis method, system, equipment and medium for intelligent driving automobile
EP4035059A1 (en) Object detection circuitry and object detection method
WO2021109033A1 (en) Apparatus and method for collecting and auto-labelling measurement data in traffic scenario
xi Wang et al. Detection of MMW radar target based on Doppler characteristics and deep learning
CN113793250B (en) Pose evaluation method, pose determination method, corresponding device and electronic equipment
CN112927338A (en) Simulation method based on three-dimensional contour, storage medium and computer equipment
CN114627365A (en) Scene re-recognition method and device, electronic equipment and storage medium
Yang et al. Locator slope calculation via deep representations based on monocular vision
Al-Temeemy et al. Laser radar invariant spatial chromatic image descriptor
Dwijayanti et al. Real-time object detection and distance measurement for humanoid robot using you only look once

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20768050

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2022517741

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 2020768050

Country of ref document: EP

Effective date: 20220425