US8824548B2 - Object detecting with 1D range sensors - Google Patents

Object detecting with 1D range sensors Download PDF

Info

Publication number
US8824548B2
US8824548B2 US13/092,408 US201113092408A US8824548B2 US 8824548 B2 US8824548 B2 US 8824548B2 US 201113092408 A US201113092408 A US 201113092408A US 8824548 B2 US8824548 B2 US 8824548B2
Authority
US
United States
Prior art keywords
sequence
image
classifier
labels
scanner
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US13/092,408
Other versions
US20110200229A1 (en
Inventor
Cuneyt Oncel Tuzel
Gungor Polatkan
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Mitsubishi Electric Research Laboratories Inc
Original Assignee
Mitsubishi Electric Research Laboratories Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US11/385,620 external-priority patent/US7903737B2/en
Application filed by Mitsubishi Electric Research Laboratories Inc filed Critical Mitsubishi Electric Research Laboratories Inc
Priority to US13/092,408 priority Critical patent/US8824548B2/en
Publication of US20110200229A1 publication Critical patent/US20110200229A1/en
Assigned to MITSUBISHI ELECTRIC RESEARCH LABORATORIES, INC. reassignment MITSUBISHI ELECTRIC RESEARCH LABORATORIES, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: POLATKAN, GUNGOR, TUZEL, CUNEYT ONCEL
Assigned to MITSUBISHI ELECTRIC RESEARCH LABORATORIES, INC reassignment MITSUBISHI ELECTRIC RESEARCH LABORATORIES, INC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: POLATKAN, GUNGOR, TUZEL, CUNEYT ONCEL
Priority to JP2012090679A priority patent/JP5773935B2/en
Application granted granted Critical
Publication of US8824548B2 publication Critical patent/US8824548B2/en
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • H04N19/615Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding using motion compensated temporal filtering [MCTF]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/103Selection of coding mode or of prediction mode
    • H04N19/105Selection of the reference unit for prediction within a chosen coding or prediction mode, e.g. adaptive choice of position and number of pixels used for prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/61Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding in combination with predictive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/60Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding
    • H04N19/63Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using transform coding using sub-band based transform, e.g. wavelets
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/13Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]

Definitions

  • This invention relates generally to image processing, and more particularly to classifying objects using range scanners in computer vision applications.
  • Object classification is widely used in computer vision applications. While most common applications use 2D camera images, there is a need for accurate classification methods for 3D range data. For example, the objects can be part moving on an assembly line.
  • object classification can use several type of data acquisition techniques such as inductive loop detector, video detector, acoustic detector, range sensor, and infrared detector.
  • data acquisition techniques such as inductive loop detector, video detector, acoustic detector, range sensor, and infrared detector.
  • One system uses a laser sensor that outputs range and intensity information for object detection and classification.
  • the embodiments of the invention provide a method and system for classifying objects based on maximum margin classification and discriminative probabilistic sequential modeling of range data acquired by a scanner with a set of one or more 1D laser line scanner.
  • the method includes pre-processing and classification phases. Different techniques, such as median filtering, background and foreground detection, 3D reconstruction and object prior information, are used during pre-processing steps to denoise the range data, and extract the most discriminative features. Then, a classifier is trained.
  • the classifier is composed of an appearance classifier, a sequence classifier with different inference techniques, and state machine enforcement.
  • FIG. 1 is a block diagram of object classification according to embodiments of the invention.
  • FIG. 2 is a schematic of a scanner with a 1D laser line scanners according to embodiments of the invention.
  • FIG. 1 shows a system and method for classifying an object 80 according to embodiments of our invention.
  • Range data 101 are acquired by a scanner 90 from the object 80 as input for the method.
  • the scanner 90 includes a 1D laser line sensor.
  • the scanner is arranged a on pole 202 near the object is to be identified. It is understood that the invention can be worked with just one sensor.
  • FIG. 2 also shows the field of view 203 for each sensor.
  • the sensor acquires one or more side views of the object.
  • the 1D (line) measurements of the range data are accumulated over time, and 2D images of range profile of the object are constructed.
  • the 2D range image is used for object type classification.
  • Output is a class 109 of the object.
  • the method includes a preprocessing phase, and a classifying phase.
  • preprocessing we denoise 110 the range data, remove 120 irrelevant background information, 3D project 130 the remaining foreground pixels using range information, and sensor scanning geometries, correct 140 the range, and extract features 155 .
  • classification 170 we use outputs of a appearance classifier such as multi-class support vector machine (SVM) as features for a sequence classifier such as a conditional random field (CRF) classification to obtain initial class labels, enforce 180 object structure using discriminative properties of objects and feature attributes, and the sequential structure, and finally obtain the object class 109 .
  • SVM multi-class support vector machine
  • CRF conditional random field
  • the noise level of the measurements changes based on the surface reflectance. For example, a black object can result in noisy measurements.
  • each column of measurement comes from a vertical line in 3D space.
  • different lines of scans can have different depth values (such as pole and body can be at different depth values).
  • Classification is performed by the following steps. First, the height features are classified in the appearance classification 160 , and the appearance classification output is denoised using a sequence classification 170 . This approach is highly accurate because it benefits from both the maximum-margin nature of the appearance classification such as SVM and the power of discriminative probabilistic sequential model such as CRF. At last, we use a structure enforcement using a finite state machine to prevent invalid predictions, e.g. a object with only a single tire.
  • the multi-class max-margin classifier SVM assigns initial labels to each time step of the image sequence.
  • the sequential structure of the data is not taken into account during learning in this step except the windowing procedure in feature extraction.
  • SVM takes the 70 ⁇ 11 dimensional height feature described above, and labels each features as either an object, body, tire or pole.
  • the window with length 11 is shifted along the time axis, and each column of the range data is classified in that manner during testing.
  • the SVM assigns initial labels but does not consider the sequential structure of the object. Therefore, we use the CRFs as an additional layer to exploit the sequential correlations between time steps. This stage performs as a denoising part on the predictions of SVMs, removing inconsistencies.
  • MEMM Maximum Entropy Markov Model
  • an inference process labels a test sequence.
  • accurately predicting whole label sequence is very difficult so that individual predictions are used. This is achieved via predicting y i,t from a marginal distribution p(y i,t
  • ⁇ t ⁇ ( j ) ⁇ i ⁇ ⁇ ⁇ ⁇ ( j , i , x t ) ⁇ ⁇ t - 1 ⁇ ( i ) , where ⁇ t (j) are the forward variables.
  • the backward recursion is
  • ⁇ t ⁇ ( i ) ⁇ j ⁇ ⁇ ⁇ t + 1 ⁇ ( j , i , x t + 1 ) ⁇ ⁇ t + 1 ⁇ ( j ) , where ⁇ t (i) are the backward variables, from which the marginal probabilities can be determined.
  • the final step of classification is the enforcement of object constraints.
  • This module takes output of the CRF. If labels do not correspond to a valid object, in other words, the labels do not correspond to some finite state machine. We convert the labels to labels of a most similar valid object model defined in an object grammar. If the CRF result is valid, this means there is no need for any correction. This is the case for a great majority of objects.
  • the process is an error correcting regular grammar parser.

Abstract

Moving objects are classified based on maximum margin classification and discriminative probabilistic sequential modeling of range data acquired by a scanner with a set of one or more 1D laser line scanner. The range data in the form of 2D images is pre-processed and then classified. The classifier is composed of appearance classifiers, sequence classifiers with different inference techniques, and state machine enforcement of a structure of the objects.

Description

FIELD OF THE INVENTION
This invention relates generally to image processing, and more particularly to classifying objects using range scanners in computer vision applications.
BACKGROUND OF THE INVENTION
Object classification is widely used in computer vision applications. While most common applications use 2D camera images, there is a need for accurate classification methods for 3D range data. For example, the objects can be part moving on an assembly line.
The innovation of new sensor technologies results in new types of data collection techniques. In conjunction, new applications of automations appear and machines are substituted for more and more human labor.
Generally, object classification can use several type of data acquisition techniques such as inductive loop detector, video detector, acoustic detector, range sensor, and infrared detector. One system uses a laser sensor that outputs range and intensity information for object detection and classification.
SUMMARY OF THE INVENTION
The embodiments of the invention provide a method and system for classifying objects based on maximum margin classification and discriminative probabilistic sequential modeling of range data acquired by a scanner with a set of one or more 1D laser line scanner.
The method includes pre-processing and classification phases. Different techniques, such as median filtering, background and foreground detection, 3D reconstruction and object prior information, are used during pre-processing steps to denoise the range data, and extract the most discriminative features. Then, a classifier is trained. The classifier is composed of an appearance classifier, a sequence classifier with different inference techniques, and state machine enforcement.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a block diagram of object classification according to embodiments of the invention; and
FIG. 2 is a schematic of a scanner with a 1D laser line scanners according to embodiments of the invention.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT
Notation
We use the following notations to represent all the variables described herein, which are either explicitly defined or obvious from the description. We use bold character to represent vectors, i.e., data sequences in this case, and unbold character to represent single variables. For example, xi=
Figure US08824548-20140902-P00001
χi,1, χi,2, . . . , xi,Ti
Figure US08824548-20140902-P00002
represents the sequence indexed by i, and xi,j represents the single variable of the sequence i at time step j. For an arbitrary single sequence, we skip the sequence index i and write the sequence as Xi=
Figure US08824548-20140902-P00001
χ, χ2, . . . , χT
Figure US08824548-20140902-P00002
.
Overview
FIG. 1 shows a system and method for classifying an object 80 according to embodiments of our invention. Range data 101 are acquired by a scanner 90 from the object 80 as input for the method.
As shown in FIG. 2, the scanner 90 includes a 1D laser line sensor. The scanner is arranged a on pole 202 near the object is to be identified. It is understood that the invention can be worked with just one sensor.
FIG. 2 also shows the field of view 203 for each sensor. The sensor acquires one or more side views of the object.
The 1D (line) measurements of the range data are accumulated over time, and 2D images of range profile of the object are constructed. The 2D range image is used for object type classification. Output is a class 109 of the object.
The above steps can be performed in a processor 199 connected to memory and input/output interfaces as known in the art.
The method includes a preprocessing phase, and a classifying phase. During preprocessing, we denoise 110 the range data, remove 120 irrelevant background information, 3D project 130 the remaining foreground pixels using range information, and sensor scanning geometries, correct 140 the range, and extract features 155.
For classification 170, we use outputs of a appearance classifier such as multi-class support vector machine (SVM) as features for a sequence classifier such as a conditional random field (CRF) classification to obtain initial class labels, enforce 180 object structure using discriminative properties of objects and feature attributes, and the sequential structure, and finally obtain the object class 109.
Preprocessing
Initial Denoising p One major problem with the range data is the noise due to non-zero angle of incidence, reflectance of an objects surfaces, imperfect operation of scanner, and interfering noise from the environment. Therefore, we first denoise the range data.
We use a 2D median filter to denoise the range data. Median filtering tends to preserve detail information, e.g., edges, while denoising the signal. We use an M×N neighborhood window around a corresponding pixel in the input image to be filtered, where M and N are specified empirically from the data. Median filtering reduces noise significantly even with a relatively small neighborhood. The tradeoff between the detail information and the amount of denoising is balanced by the order of the filter. The higher the order the higher the noise reduction, but less detail remains in the image.
Background Estimation and Removal
Some pixels can be totally corrupted during acquisition. Because of that, at the first step of background estimation, we determine “good” pixels and “bad” pixels based on a median amplitude of each row of pixels. Then, we use pixel based background estimation by fitting a single Gaussian distribution on the history of the range values of each good pixel when there is no object in the scene. At each new test sample from the same pixel, the determination is based on hypothesis testing as either foreground or background. For bad pixels, the decision is based on the hypothesis testing using the amplitude values of the signal. Finally, we use median filtering on the background mapping in order to remove irrelevant regions of noisy pixels.
3D Projection
Depending on environmental conditions and deployment errors, positions and orientations of the sensors relative to object can be inaccurate. To solve this issue, we back project good foreground pixels to 3D using initial sensor information, and fit a plane to a ground plane. We use a RANdom SAmple Consensus (RANSAC) process for plane fitting. This plane modifies the sensor location and orientation. The estimated base plane is assumed to correspond to the y=0 plane of a world coordinate system. Given the relative locations and orientations of the sensor with respect to base plane and the sensor field of view, we determine the 3D coordinates of each sensor measurement in a world coordinate system with back projection. The 3D projection is helpful in the following ways. We extract planar side view information from 3D values, which we use during range correction, and features. In addition, unlike 2D images, which are subject to perspective deformation of the world to image plane, the features we obtain from the 3D values are scale invariant and more informative.
Range Correction
The noise level of the measurements changes based on the surface reflectance. For example, a black object can result in noisy measurements. We exploit 3D information and planar side structure of the object to further correct range values. We assume that each column of measurement comes from a vertical line in 3D space. However different lines of scans can have different depth values (such as pole and body can be at different depth values). We initially determine the top 30% of depth values for each column of measurement.
Next, we median filter these measurements over time with an empirically specified filter order and obtain the depth values of each column of measurement. The larger the median filter order, the larger the area is assumed to have the same depth. Then, we correct outlier range values with the ones projected to the estimated plane. After range correction the noisy samples are relocated to correct positions and the object has a smooth structure.
Features
We use a binary height map as our features, which is equal to the quantized side view of the 3D projection. Initially, we take a part of the object above the base plane, and quantize such that each pixel corresponds to a small height value. For some objects, due to background removal, parts of the object that touch the base are removed. Therefore, we first detect the bottom of the object in the side view and shift the object to touch the base. Moreover, to incorporate partial temporal information, we take overlapping 70×11 patches of pixels using a sliding window technique. One patch is taken for each column on the image. Then, this patch is passed to classifying phase as a feature to obtain a classification of the center column.
Classification
Classification is performed by the following steps. First, the height features are classified in the appearance classification 160, and the appearance classification output is denoised using a sequence classification 170. This approach is highly accurate because it benefits from both the maximum-margin nature of the appearance classification such as SVM and the power of discriminative probabilistic sequential model such as CRF. At last, we use a structure enforcement using a finite state machine to prevent invalid predictions, e.g. a object with only a single tire.
Appearance Classification
The multi-class max-margin classifier SVM assigns initial labels to each time step of the image sequence. The sequential structure of the data is not taken into account during learning in this step except the windowing procedure in feature extraction. SVM takes the 70×11 dimensional height feature described above, and labels each features as either an object, body, tire or pole. The window with length 11 is shifted along the time axis, and each column of the range data is classified in that manner during testing. We use a linear kernel SVM, which enables fast processing.
Sequence Classification
The SVM assigns initial labels but does not consider the sequential structure of the object. Therefore, we use the CRFs as an additional layer to exploit the sequential correlations between time steps. This stage performs as a denoising part on the predictions of SVMs, removing inconsistencies. A sequential learning problem can be formulated as finding the optimal fiuictionf that can predict y=f(x), given N training sequences
{(xi,yi)}i N=1, where xi=
Figure US08824548-20140902-P00001
χi,1, χi,2, . . . , χi,Ti
Figure US08824548-20140902-P00002
,
and
yi=
Figure US08824548-20140902-P00001
yi,1,yi,2, . . . , yi,Ti
Figure US08824548-20140902-P00002

is the label sequence.
One common approach to solve the sequence labeling problem using probabilistic sequential modeling is to use generative models to sequence labeling problem, such as hidden Markov models (HMM). Another common approach is to use discriminative models. One such model is the Maximum Entropy Markov Model (MEMM). In addition to being a discriminative model, MEMMs provide the ability to model arbitrary features of observation sequences. One can handle overlapping features in this way. However, the label bias problem limits the performance of MEMMs.
Therefore, we use the CRFs as the sequence labeler to smooth noisy SVM outputs. A linear chain conditional random field is defined as
p ( y | x ) = 1 Z ( x ) t = 1 T Ψ ( y t , y t - 1 , x t ) , ( 1 ) Ψ ( y t , y t - 1 , x t ) exp { j λ j g j ( y t - 1 , y t , x ) + k μ k u k ( y t , x ) } . ( 2 )
where
Ψ(yt,y t−1, χt)
is the potential function,
gj(y t−1,yt,x)
is the transition feature function from state, and
y t−1 to yt: μk(yt,x),
is state feature function at state yi; λj and a μk are the parameters estimated at the learning process, and Z(x) is the normalization factor as a function of the observation sequence. Maximum likelihood parameter estimation of the above exponential family distribution corresponds to the maximum entropy solution.
Inference
After the model parameters are learned, an inference process labels a test sequence. We give a brief overview of conventional inference methods on probabilistic sequential models. One way of labeling a test sequence is the most likely labeling using the joint density y*=arg max), p(y|x). The solution can be efficiently determined via a Viterbi process using recursion
t(j)=maxiΨ(j,i,χ tt−1(i),
which propagates the most likely path based on the maximum product rule. However, in many applications, accurately predicting whole label sequence is very difficult so that individual predictions are used. This is achieved via predicting yi,t from a marginal distribution p(yi,t|xi) using a dynamic programming forward-backward procedure,
The forward recursion is
α t ( j ) = i Ψ ( j , i , x t ) α t - 1 ( i ) ,
where αt(j) are the forward variables. The backward recursion is
β t ( i ) = j Ψ t + 1 ( j , i , x t + 1 ) β t + 1 ( j ) ,
where βt(i) are the backward variables, from which the marginal probabilities can be determined.
Structure Enforcement
The final step of classification is the enforcement of object constraints. This module takes output of the CRF. If labels do not correspond to a valid object, in other words, the labels do not correspond to some finite state machine. We convert the labels to labels of a most similar valid object model defined in an object grammar. If the CRF result is valid, this means there is no need for any correction. This is the case for a great majority of objects. The process is an error correcting regular grammar parser.
Although the invention has been described by way of examples of preferred embodiments, it is to be understood that various other adaptations and modifications can be made within the spirit and scope of the invention. Therefore, it is the object of the appended claims to cover all such variations and modifications as come within the true spirit and scope of the invention.

Claims (15)

We claim:
1. A method for classifying an object in a scene, comprising the steps of:
preprocessing a sequence of images, wherein each image is acquired of the object in the scene by a scanner, wherein the scanner includes a 1D laser line sensor, and each image includes columns of pixels, and each pixel has an associated depth value such that each image is a range image, wherein the preprocessing further comprises;
denoising each image in the sequence;
removing background pixels from each image;
projecting, in 3D, each image to a 3D world coordinate system;
correcting depth values;
extracting features; and
classifying the sequence of images, wherein the classifying further comprises:
applying an appearance classifier to the features to obtain labels;
applying a sequence classifier to smooth the labels; and
enforcing a structure of the object to determine a class of the object, wherein the steps are performed in a processor.
2. The method of claim 1, wherein 1D laser line scanner scans a side of the object.
3. The method of claim 1, wherein the denoising uses a 2D median filter.
4. The method of claim 1, wherein a background of the scene is modeled with a Gaussian distribution for each pixel.
5. The method of claim 1, wherein a ground plane in the scene is estimated using a RANdom SAmple Consensus (RANSAC) process.
6. The method of claim 1, wherein noisy range measurements on surfaces of the object are corrected by fitting vertical planes to each column.
7. The method of claim 1, wherein the features are scale invariant.
8. The method of claim 1, wherein the features are binary height maps equal to a quantized side view of the 3D projection.
9. The method of claim 1, wherein the scanner is mounted on a pole near the object.
10. The method of claim 1, wherein the labels are determined by evaluating with the appearance classifier by a sliding window technique along a time axis.
11. The method of claim 1, wherein the appearance classifier is a support vector machine.
12. The method of claim 1, wherein the outputs of the appearance classifier are smoothed with the sequence classifier.
13. The method of claim 1, wherein the sequence classifier uses a conditional random field model or a hidden Markov model.
14. The method of claim 1, wherein the structure enforcing converts the labels to labels of a most similar valid object model defined in a object grammar.
15. The method of claim 1, wherein the structure enforcing uses an error correcting regular grammar parser.
US13/092,408 2006-03-21 2011-04-22 Object detecting with 1D range sensors Expired - Fee Related US8824548B2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US13/092,408 US8824548B2 (en) 2006-03-21 2011-04-22 Object detecting with 1D range sensors
JP2012090679A JP5773935B2 (en) 2011-04-22 2012-04-12 How to classify objects in a scene

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US11/385,620 US7903737B2 (en) 2005-11-30 2006-03-21 Method and system for randomly accessing multiview videos with known prediction dependency
US13/092,408 US8824548B2 (en) 2006-03-21 2011-04-22 Object detecting with 1D range sensors

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US11/385,620 Division US7903737B2 (en) 2005-11-30 2006-03-21 Method and system for randomly accessing multiview videos with known prediction dependency

Publications (2)

Publication Number Publication Date
US20110200229A1 US20110200229A1 (en) 2011-08-18
US8824548B2 true US8824548B2 (en) 2014-09-02

Family

ID=47470252

Family Applications (1)

Application Number Title Priority Date Filing Date
US13/092,408 Expired - Fee Related US8824548B2 (en) 2006-03-21 2011-04-22 Object detecting with 1D range sensors

Country Status (1)

Country Link
US (1) US8824548B2 (en)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9740937B2 (en) 2012-01-17 2017-08-22 Avigilon Fortress Corporation System and method for monitoring a retail environment using video content analysis with depth sensing
US9858923B2 (en) * 2015-09-24 2018-01-02 Intel Corporation Dynamic adaptation of language models and semantic tracking for automatic speech recognition
US10371512B2 (en) * 2016-04-08 2019-08-06 Otis Elevator Company Method and system for multiple 3D sensor calibration
US10729382B2 (en) * 2016-12-19 2020-08-04 Mitsubishi Electric Research Laboratories, Inc. Methods and systems to predict a state of the machine using time series data of the machine
CN114205621A (en) * 2018-02-28 2022-03-18 三星电子株式会社 Encoding method and device, and decoding method and device
CN110751188B (en) * 2019-09-26 2020-10-09 华南师范大学 User label prediction method, system and storage medium based on multi-label learning

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020159628A1 (en) * 2001-04-26 2002-10-31 Mitsubishi Electric Research Laboratories, Inc Image-based 3D digitizer
US7043084B2 (en) * 2002-07-30 2006-05-09 Mitsubishi Electric Research Laboratories, Inc. Wheelchair detection using stereo vision
US20080063264A1 (en) * 2006-09-08 2008-03-13 Porikli Fatih M Method for classifying data using an analytic manifold
US20080063285A1 (en) * 2006-09-08 2008-03-13 Porikli Fatih M Detecting Moving Objects in Video by Classifying on Riemannian Manifolds
US7599555B2 (en) * 2005-03-29 2009-10-06 Mitsubishi Electric Research Laboratories, Inc. System and method for image matting
US20090315981A1 (en) * 2008-06-24 2009-12-24 Samsung Electronics Co., Ltd. Image processing method and apparatus
US7835568B2 (en) * 2003-08-29 2010-11-16 Samsung Electronics Co., Ltd. Method and apparatus for image-based photorealistic 3D face modeling
US7903737B2 (en) * 2005-11-30 2011-03-08 Mitsubishi Electric Research Laboratories, Inc. Method and system for randomly accessing multiview videos with known prediction dependency

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020159628A1 (en) * 2001-04-26 2002-10-31 Mitsubishi Electric Research Laboratories, Inc Image-based 3D digitizer
US7043084B2 (en) * 2002-07-30 2006-05-09 Mitsubishi Electric Research Laboratories, Inc. Wheelchair detection using stereo vision
US7835568B2 (en) * 2003-08-29 2010-11-16 Samsung Electronics Co., Ltd. Method and apparatus for image-based photorealistic 3D face modeling
US7599555B2 (en) * 2005-03-29 2009-10-06 Mitsubishi Electric Research Laboratories, Inc. System and method for image matting
US7903737B2 (en) * 2005-11-30 2011-03-08 Mitsubishi Electric Research Laboratories, Inc. Method and system for randomly accessing multiview videos with known prediction dependency
US20080063264A1 (en) * 2006-09-08 2008-03-13 Porikli Fatih M Method for classifying data using an analytic manifold
US20080063285A1 (en) * 2006-09-08 2008-03-13 Porikli Fatih M Detecting Moving Objects in Video by Classifying on Riemannian Manifolds
US20090315981A1 (en) * 2008-06-24 2009-12-24 Samsung Electronics Co., Ltd. Image processing method and apparatus

Also Published As

Publication number Publication date
US20110200229A1 (en) 2011-08-18

Similar Documents

Publication Publication Date Title
US8594431B2 (en) Adaptive partial character recognition
US8824548B2 (en) Object detecting with 1D range sensors
Schindler An overview and comparison of smooth labeling methods for land-cover classification
Moser et al. Dictionary-based stochastic expectation-maximization for SAR amplitude probability density function estimation
US7813581B1 (en) Bayesian methods for noise reduction in image processing
Korus et al. Evaluation of random field models in multi-modal unsupervised tampering localization
Bentabet et al. Road vectors update using SAR imagery: a snake-based method
Zhang et al. Hierarchical conditional random fields model for semisupervised SAR image segmentation
CN107680120A (en) Tracking Method of IR Small Target based on rarefaction representation and transfer confined-particle filtering
CN105590020B (en) Improved data comparison method
US20070127817A1 (en) Change region detection device and change region detecting method
US20030215155A1 (en) Calculating noise estimates of a digital image using gradient analysis
US20110274356A1 (en) Image pattern recognition
US11663840B2 (en) Method and system for removing noise in documents for image processing
US20180122097A1 (en) Apparatus, method, and non-transitory computer-readable storage medium for storing program for position and orientation estimation
Hong et al. Selective image registration for efficient visual SLAM on planar surface structures in underwater environment
Mohammad et al. Contour-based character segmentation for printed Arabic text with diacritics
Nguyen et al. UnfairGAN: An enhanced generative adversarial network for raindrop removal from a single image
Jana et al. A fuzzy C-means based approach towards efficient document image binarization
CN113313179A (en) Noise image classification method based on l2p norm robust least square method
Ghoshal et al. An improved scene text and document image binarization scheme
CN113239828A (en) Face recognition method and device based on TOF camera module
JP5773935B2 (en) How to classify objects in a scene
Gómez-Moreno et al. A “salt and pepper” noise reduction scheme for digital images based on support vector machines classification and regression
Sharma et al. A Noise-Resilient Super-Resolution framework to boost OCR performance

Legal Events

Date Code Title Description
AS Assignment

Owner name: MITSUBISHI ELECTRIC RESEARCH LABORATORIES, INC., M

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:TUZEL, CUNEYT ONCEL;POLATKAN, GUNGOR;SIGNING DATES FROM 20120302 TO 20120307;REEL/FRAME:027883/0406

Owner name: MITSUBISHI ELECTRIC RESEARCH LABORATORIES, INC, MA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:TUZEL, CUNEYT ONCEL;POLATKAN, GUNGOR;SIGNING DATES FROM 20120302 TO 20120307;REEL/FRAME:027883/0469

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551)

Year of fee payment: 4

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20220902