US8542276B2 - Object Tracking method and apparatus for a non-overlapping-sensor network - Google Patents
Object Tracking method and apparatus for a non-overlapping-sensor network Download PDFInfo
- Publication number
- US8542276B2 US8542276B2 US12/758,812 US75881210A US8542276B2 US 8542276 B2 US8542276 B2 US 8542276B2 US 75881210 A US75881210 A US 75881210A US 8542276 B2 US8542276 B2 US 8542276B2
- Authority
- US
- United States
- Prior art keywords
- sensors
- probability function
- exit
- entrance
- time difference
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/292—Multi-camera tracking
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
- G06F18/2155—Generating training patterns; Bootstrap methods, e.g. bagging or boosting characterised by the incorporation of unlabelled data, e.g. multiple instance learning [MIL], semi-supervised techniques using expectation-maximisation [EM] or naïve labelling
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
- G06V10/7753—Incorporation of unlabelled data, e.g. multiple instance learning [MIL]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/94—Hardware or software architectures specially adapted for image or video understanding
- G06V10/95—Hardware or software architectures specially adapted for image or video understanding structured as a network, e.g. client-server architectures
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/52—Surveillance or monitoring of activities, e.g. for recognising suspicious objects
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
Definitions
- the disclosure generally relates to an object tracking method and apparatus for a non-overlapping-sensor network.
- Video surveillance system detects the occurrence of abnormal security events by analyzing the behavior of the moving people in monitored video, and effectively notifies the security staff to handle.
- the basic issues of video surveillance systems such as, background subtraction, moving object detection and tracking, shadow removal, and so on, are all well researched and documented.
- High-level event detection such as, behavior analysis, unattended object detection, loitering detection or jam detection, i.e., automatic and intelligent behavior analysis, is also expected to be in high demand.
- a steady moving object tracking technology is the basic element of the intelligent video surveillance system.
- Taiwan Patent Publication No. 200806020 discloses a video tracking technology by using a fixed camera with pre-set priority and a PTZ camera cooperatively tracking an object. When the camera with priority detects moving object, PTZ camera is activated to track the moving object so that the field of view covers the field of view of fixed camera.
- Taiwan Patent Publication No. 200708102 discloses a video surveillance system merging data from a plurality of surveillance cameras to monitor a large-area scene, and providing scene map and scale map of the monitored scene, and sensor network model information of the scene to the monitored scene. For example, as shown in FIG. 1 , these types of information may be stored in map-FOV image 104 , human scale map 108 and camera network model 112 , and may be generated and managed by map basic calibrator 102 , FOV basic calibrator 106 and camera network model manager 110 .
- U.S. Pat. No. 7,149,325 discloses a cooperative camera network architecture for recording color characteristic of pedestrians and storing in a database for human identification, where only when the person is in the overlapped part of the cameras, the moving object can be tracked.
- U.S. Pat. No. 7,394,916 discloses a method for target tracking, aiming at the situation when a human figure appearing in different cameras, comparing the likelihoods of transition of the scene and the other scenes of the previous human figures departing for the basis as human tracking.
- the likelihoods of transition aim at the blueprint of scene, speed of moving object and the distance to entrances and exits or traffic condition, and are set by the user.
- China Patent Publication No. 101,142,593A discloses a method for tracking target in a video sequence. This method compares the changes of appearance feature of the foreground appearing in different cameras. When comparing the different foreground objects, extra comparison is performed when different foreground objects show the state of engagement so as to eliminate the condition that the correct corresponding foreground object cannot be found when the foreground object is in the state of engagement. When comparing different foreground objects in different cameras, the combination of foreground color distribution and edge density information is used to compute the correlation of the foregrounds.
- China Patent Publication No. 101,090,485A discloses an image surveillance system and object tracking method, where the functional module of image processing unit 200 is shown as FIG. 2 .
- the image processing unit executes the object detection processing and object tracking processing in detecting moving object of the image.
- this unit uses a unique label to correlate the current object and the previous object.
- the tracking processing will keep the label assigned to the invisible object and the label will be assigned to the object when the object is visible again.
- the exemplary embodiments may provide an object tracking method and apparatus for a non-overlapping-sensor network, applicable to a sensor network with a plurality of sensors.
- the disclosed relates to an object tracking method for a non-overlapping-sensor network.
- the method comprises a training phase and a detection phase.
- a training phase a plurality of data measured by the sensors in the sensor network is used as training samples.
- At least an entrance/exit is marked out within the measurement range of each sensor.
- At least three characteristic functions related to an object to be tracked including sensor spatial relation among the sensors in the sensor network, time difference of movement and similarity in appearance, are estimated by an automatic learning method.
- the at least three characteristic functions are used as the principles for tracking the object and linking relationship of said object in the detection phase.
- the disclosed relates to an object detection system for a non-overlapping-sensor network.
- the system comprises a plurality of sensors forming a sensor network, a training-phase processing module, a characteristic function estimating and updating module and a detection-phase tracking module, where at least an entrance/exit is marked out within the measurement range of the plurality of sensors.
- the training-phase processing module obtains a plurality of measured data by the sensors in the sensor network and used as training samples, and records all the departure events within a previous duration in a training sample space for the entering events in each entrance/exit of each sensor.
- the characteristic function estimating and updating module uses an automatic learning method and the existing samples in the training sample space to estimate at least three characteristic functions of the object correlation, including function of sensor spatial relation among the sensors in the sensor network, function of time difference of movement and function of similarity in appearance.
- the detection-phase tracking module uses the at least three characteristic functions as the principles for object tracking and relationship linking in the detection phase.
- FIG. 1 shows an exemplary schematic view of the scene model manager of a large-area scene-based video surveillance system.
- FIG. 2 shows an exemplary schematic view of the functional modules of an image processing unit of an image surveillance system.
- FIG. 3A shows an exemplary schematic view of the FOV and entrance/exit of each camera of a camera network, consistent with certain disclosed embodiments.
- FIG. 3B shows an exemplary schematic view of the equivalence of the non-overlapping-camera object tracking problem and the object correlation problem of objects observed at different time and different entrance/exit, consistent with certain disclosed embodiments.
- FIG. 4 shows an exemplary schematic view of the basis of tracking a person of the leaving entrance/exit when the person entering entrance/exit to enter the camera FOV, consistent with certain disclosed embodiments.
- FIG. 5 shows an exemplary flowchart of an object tracking method for a non-overlapping sensor network, consistent with certain disclosed embodiments.
- FIG. 6A-FIG . 6 C show an exemplar of the entire training phase, consistent with certain disclosed embodiments.
- FIG. 7 shows an exemplary schematic block diagram of the design of recursive learning strategy, consistent with certain disclosed embodiments.
- FIG. 8 shows an exemplary flowchart of the method of recursive learning strategy, consistent with certain disclosed embodiments.
- FIG. 9A-FIG . 9 B show an exemplary experiment scene and an exemplary camera deployment of a camera network, respectively, consistent with certain disclosed embodiments.
- FIG. 10 shows an exemplary training result with actual correct corresponding relation of FIG. 9 , where FIG. 10A and FIG. 10B are bar charts H( ⁇ A) and H( ⁇ t), respectively, and FIG. 10C is the mixture Gaussian model approximating bar chart H( ⁇ t), consistent with certain disclosed embodiments.
- FIG. 11 shows an exemplary training result without actual correct corresponding relation of FIG. 9 , where FIG. 11A and FIG. 11B are bar charts H( ⁇ A) and H( ⁇ t), respectively, and FIG. 11C is the mixture Gaussian model approximating bar chart H( ⁇ t), consistent with certain disclosed embodiments.
- FIG. 12 shows an exemplar describing the inquiry result of finding the correct related events, where FIG. 12A is the person entering event of inquiry, FIG. 12B , FIG. 12C , FIG. 12D are three possible related person leaving events, consistent with certain disclosed embodiments.
- FIG. 13 shows an exemplar describing the inquiry result of inability to find the correct related events, where FIG. 13A is the person entering event of inquiry, FIG. 13B , FIG. 13C , are two events with very low correlation values, consistent with certain disclosed embodiments.
- FIG. 14 shows an exemplary schematic view of an object tracking system for a non-overlapping sensor network, consistent with certain disclosed embodiments.
- Moving object tracking for non-overlapping sensors is defined for a sensor network having k sensors (e.g., sensor C — 1, sensor C — 2, . . . , sensor C_k), and each sensor including nk entrances/exits.
- sensors e.g., sensor C — 1, sensor C — 2, . . . , sensor C_k
- each sensor including nk entrances/exits.
- nk entrances/exits For example, in the measurement range of sensor C — 1, there exist a — 1, a — 2, . . . , a_n1 entrances/exits, in the measurement range of sensor C — 2, there exist b — 1, b — 2, . . . , b_n1 entrances/exits, and so on.
- An entrance/exit is where an object appears in for disappears from the region of the measurement range of the sensor.
- the object tracking in non-overlapping sensors may be viewed as solving the related problem of objects entering and leaving different entrances/exits and the objects observed at different entrances/exits at different time.
- the senor may be a sensor of various types, such as, color video camera, with a camera network to track the object movement, but is not limited to this type of sensor.
- the sensor may be black-and-white camera, heat-sensor camera, infrared camera, microphone, supersonic, laser distance-measurement instrument, weight scale, and so on.
- FOV field of view
- FOV A includes entrances/exits A1, A2
- FOV B includes entrances/exits B1, B2
- B3 FOV C includes entrances/exits C1, C2, as shown in FIG. 3A .
- 14 different object images are observed at different time at the entrances/exits, as shown in FIG. 3B .
- object image 312 leaves FOV A1 and disappears from FOV A
- object image 321 enters B1 and appears in FOV B
- object image 322 leaves B3 and disappears from FOV B
- object image 331 enters C1 and appears in FOV C.
- the object tracking for non-overlapping cameras is to establish the correct correlation (dash line 310 ) between object image 312 and object image 321 and correlation (dash line 320 ) between object image 322 and object image 331 .
- the object tracking for non-overlapping cameras may be accomplished by establishing the correct correlation between the object images (shown as the connections with the dash lines).
- O_i_p represents a person p is observed at entrance/exit I
- the appearance characteristic O_i_p(a) and the time difference between leaving an entrance/exit and entering another entrance/exit another characteristic O_i_p(t) as the related basis to accomplish object tracking.
- the person correlation problem may be expressed as the conditional probability P(M((i,j),(p,q))
- the appearance characteristic difference ⁇ a and time difference ⁇ t of person moving cross-cameras may be computed as characteristics, and assume that the moving person does not change much in appearance when moving from camera to camera and most people move roughly at the same speed, which are both reasonable assumptions.
- O_i_p,O_j_q) of equation (1) may be rewritten as follows according to Bayes rule:
- equation (2) implies that when a person q enters a camera FOV at an entrance/exit, where does the person q leave from?
- the basis of tracking person q is as follows: backtrack all the persons left each camera FOV in previous ⁇ T time, and maximize equation (2) and P( ⁇ a(p,q)
- FIG. 4 shows how a person entering FOV of a camera at entrance/exit A2, and the basis of tracking the person's exit point, consistent with certain disclosed embodiments.
- Correlation P1 is proportional to appearance similarity, camera spatial correlation M(A1, B2) and moving time difference (t ⁇ t1) of camera sample 411 ;
- correlation P2 is proportional to appearance similarity, camera spatial correlation M(A2, B3) and moving time difference (t ⁇ t2) of camera sample 412 ;
- correlation P3 is proportional to appearance similarity, camera spatial correlation M(A2, C1) and moving time difference (t ⁇ t3) of camera sample 413 .
- the person leaving event with the highest similarity may be selected from P1, P2, P3 as the correct correlated event.
- the disclosure is to provide moving object tracking in a non-overlapping sensor network, without the need of information of the deployment blueprint of the sensors and any operator involvement in the learning phase.
- the disclosed exemplary embodiments may automatically learn the P( ⁇ a(p,q)
- the disclosed exemplary embodiments provide an automatic learning method to estimate the required probability function. This automatic learning method neither needs to specify the number of samples appearing in the training data nor requires to manually label the related persons. This method may be a recursive training method, which will be described momentarily.
- FIG. 5 shows an exemplary flowchart of the method for object tracking in a non-overlapping sensor network, consistent with certain disclosed embodiments.
- the method uses a plurality of sensors in the camera network to obtain a plurality of sensor measurements data as training samples, as shown in step 510 .
- step 520 at least an entrance/exit is marked out within the measurement coverage range of each of the plurality of sensors in the sensor network.
- an automatic learning method is used to estimate at least three characteristic functions related to an object to be tracked, including spatial con-elation function of the sensors in the sensor network, time difference function of the object leaving and entering the measurement coverage range of different sensors, and similarity difference function of the object appearance.
- the at least three functions may be used as the baseline for tracking the object and linking relationship of the object, as shown in step 540 .
- the automatic learning method may be a recursive learning strategy.
- FIGS. 6A-6C show the entire training phase, consistent with certain disclosed embodiments.
- an n ⁇ n training sample space may be allocated in a memory, where n is the total number of the entrances/exits in the entire sensor network.
- Each field of this space is for storing two related entering and exiting events.
- the space may be represented by an n ⁇ n matrix, where the field (d,b) of the n ⁇ n matrix represents the event of leaving entrance/exit b during the past period when an object is observed to enter entrance/exit d.
- each field in the sample space contains a spatial correlation between the sensors.
- FIG. 6A shows an exemplary schematic view of allocating a 7 ⁇ 7 training sample space represented by a 7 ⁇ 7 matrix 630 , consistent with certain disclosed embodiments.
- the FOV of the three cameras in the camera network are A, B, C, and the camera network includes 7 entrances/exits.
- FOV A has entrances/exits A1, A2,
- FOV B has entrances/exits B1, B2, B3 and
- FOV C has entrances/exits C1, C2.
- all training sample data 610 include a plurality of object images, such as, object image 1 , object image 2 , object image 3 , and so on.
- table 615 lists all sample data entering A2, such as, object image 1 , object image 5 , and so on, and all sample data leaving B2, such as, object image 6 , object image 7 , object image 10 , and so on.
- Field (A2, B2) of matrix 630 indicates all the leaving B2 events during the past period ⁇ T when an object is observed entering A2, such as, (object image 1 , object image 6 ), (object image 1 , object image 7 ), (object image 1 , object image 6 ), and so on, marked as 620 .
- each field of matrix 630 contains the correlation between the entrances/exits of the three cameras.
- the events stored at location (d,b) are used for training the corresponding probability distribution functions P( ⁇ a
- M(d,b)) the probability distribution functions
- M(d,b)) the probability distribution functions
- the disclosed exemplary embodiments target at each field and uses bar chart to represent the appearance difference and time difference of person crossing sensors.
- the first step is to eliminate anomaly of the statistic distribution of the appearance similarity in each possible link.
- the second step is to find the data with more obvious time difference among the high appearance similarity data. After repeating the above two steps for many times, if the link does exist, the convergence distribution of time difference and the appearance characteristic difference may be found.
- FIG. 6B Take field (A2, B2) and field (A2, C2) as an example.
- the exemplar in FIG. 6B makes a bar statistic chart out of the characteristics of appearance characteristic difference, with x-axis as appearance similarity ⁇ A and y-axis as the statistic distribution of appearance similarity ⁇ A, i.e., H( ⁇ A). After eliminating the anomaly, that is, filtering out the potential outliers, i.e., the rightmost data in the bar chart.
- 6C shows the finding of the data with obvious time difference from the remaining data with high similarity and using the moving time difference of the object crossing sensors as the characteristic to draw the bar chart, with x-axis as moving time difference ⁇ t, and y-axis as bar chart, i.e., H( ⁇ t).
- Symbol x indicates the removal of the data.
- Bar charts H( ⁇ A) and H( ⁇ t) may be approximated with a mixture Gaussian model.
- a Gaussian model with a smaller mean and variance and other Gaussian models with a larger mean and variance are expected to exist because the coherence of the moving object appearance will make the correct matching to lower the appearance similarity ⁇ A, i.e., corresponding to Gaussian model with a smaller mean and variance.
- Gaussian model with a larger mean and variance will correspond to the sample outliers which are the part requiring further elimination.
- the final correct sample may be found by recursive filtering.
- bar charts H( ⁇ A) and H( ⁇ t) are made for all possible sample statistics. Some potential outliers are filtered from H( ⁇ A), i.e., the rightmost data in the bar chart.
- the Gauss value in H( ⁇ t) is updated and observed to see whether a concentrate trend exists; if so, continue filtering H( ⁇ A) and updating H( ⁇ t) until the similarity distribution function converges; otherwise, no concentrate trend exists and P(M) is relatively smaller than other combination, this indicates that these two entrances/exits have no physical spatial link.
- FIG. 7 shows an exemplary schematic block diagram of the design of the recursive learning strategy, consistent with certain disclosed embodiments.
- all the possible corresponding leaving events are collected to establish an event pool 710 for any two entrances/exits d, b, i.e., all the fields (d,b) of n ⁇ n matrix, according to the exemplary embodiment of FIG. 6A .
- event pool 710 appearance similarity difference probability function P( ⁇ A) and moving time difference probability function P( ⁇ t) are estimated and updated, where the estimation and updating of appearance similarity difference probability function includes estimating P( ⁇ A), data trimming, and updating P( ⁇ A).
- a mixture Gaussian model G1( ⁇ A) may be used to approximate appearance similarity difference probability function P( ⁇ A) and remove the outliers.
- the estimation and updating of moving time difference probability function includes estimating P( ⁇ t), data trimming, and updating P( ⁇ t), where the data trimming may use another mixture Gaussian model G2( ⁇ t) to approximate moving time difference probability function P( ⁇ t) and remove the data without the trend to concentrate.
- the next step may determine whether moving time difference probability function converges; if not converging, the process returns to event pool 710 and continues estimating and updating appearance similarity difference probability function P( ⁇ A) and moving time difference probability function P( ⁇ t); otherwise, the process ends.
- Removing the outliers may be based on whether the conditional probability function P( ⁇ A
- Removing the data without the trend to concentrate may also be based on whether the conditional probability function P( ⁇ t
- the condition of convergence for moving time difference probability function P( ⁇ t) is, for example, the number of the removed events is less than a preset value K3.
- K1, K2 are, the higher the ratio of the data removal. Thus, the condition of convergence may be reached faster. However, if K1 and K2 are set to be high, too many events may be removed. The higher K3 is set, the easier the condition of convergence may be met. But, too many events without physical links may remain.
- the settings of K1, K2, K3 may depend on the actual application, for example, the experience from the experiments.
- FIG. 8 shows an exemplary flowchart illustrating the steps of the recursive learning strategy, consistent with certain disclosed embodiments.
- step 810 all the leaving events during a past period in a training sample space are recorded for each entering event at each entrance/exit of each sensor, as shown in step 810 .
- step 820 with the samples existing in the training sample space, it is to estimate the entrance/exit correlation probability function, moving time difference probability function and appearance similarity difference probability function.
- step 830 it is to observe the appearance similarity difference function and remove the data belonging to the statistical outliers. With the remaining data, it is to update the moving time difference probability function and appearance similarity difference probability function, as shown in step 840 .
- Steps 830 - 840 are repeated until the moving time difference probability function converges, as shown in step 850 .
- step 830 it may use a mixture Gaussian model to approximate the appearance similarity difference function.
- step 840 before updating the moving time difference probability function, another mixture Gaussian model may be used to approximate the moving time difference probability function and observe whether or not to remove the data without concentrate trend.
- step 850 the convergence of the moving time difference probability function may be determined by, for example, whether the number of removed events is less than a preset number. After step 850 , the data of the remaining events may be used to estimate the entrance/exit correlation probability function.
- the experiment scene and the camera deployment of the camera network are shown in FIG. 9A and FIG. 9B , respectively.
- the experiment scene of FIG. 9A is an office and four non-overlapping cameras A, B, C, D, i.e., the areas outlined by dashed lines.
- FOV 910 of camera A has an entrance/exit a1
- FOV 920 of camera B has three entrances/exits b1, b2, b3,
- FOV 930 of camera C has two entrances/exits c1, c2
- FOV 940 of camera D has two entrances/exits d1, d2.
- a video clip is used with the first 7 minutes as the training phase and the last minute as the detection phase.
- the appearance change and time different of each entrance/exit are estimated.
- the detection phase when using a person entering event to inquire, the person leaving events with higher similarity to the person entering event will be listed. All the person leaving events occurred during the time interval (t ⁇ t_Max, t), where t is the time the person enters.
- FIG. 10A and FIG. 10B show the bar charts H( ⁇ A) and H( ⁇ t), respectively, and FIG. 10C shows the mixture Gaussian model approximating bar chart H( ⁇ t).
- the x-axis of FIG. 10A is the person similarity in the events, with 0 indicating the highest similarity.
- the x-axis of FIG. 10B and FIG. 10C is the time of person entering/leaving in the events, with second as unit.
- the y-axis of FIG. 10 is the number of events.
- FIG. 11 The training result without actual corresponding relation may be shown in FIG. 11 .
- FIG. 11C shows the approximation probability model, and P(M) approximates 0.
- the results may be either that (1) the person leaving event with highest similarity is the correct correlation event, or (2) no related leaving event is found.
- the exemplar of FIG. 12 shows the inquiry result of finding the correct related event, where FIG. 12A is camera image of the person entering event (entering b2), and FIG. 12B , FIG. 12C , FIG. 12D are the camera images of found possible person leaving events, i.e., leaving a1, c2, d2, respectively.
- 12D may be computed with equation (2), i.e., the product of P( ⁇ A(p,q)
- the leaving person in FIG. 12B and the inquired entering person is the same one, which indicates the correlation of the correct related event is higher than the others and the inquiry result is correct.
- object tracking and correlation rule are positively proportional to the aforementioned appearance similarity, moving time difference and the spatial correlation of cameras, i.e., as equation (2).
- FIG. 13 shows the no corresponding leaving person being found as the inquiry result, where FIG. 13A indicates the person entering event for the inquiry. Because there is no actual related leaving event, the found two related events, i.e., FIG. 13B and FIG. 13C have low correlation values of 7.86*10 ⁇ 4 and 3.83*10 ⁇ 138 , respectively.
- the disclosed exemplary embodiments may also be applied to other types of sensor networks, such as, black and white camera, thermal sensor camera, infrared camera, microphone, supersonic, laser distance measuring instruments, weight scale, and so on.
- the above method may be successfully applied.
- the appearance characteristic may be the texture or the gray scale intensity distribution of the moving object.
- the thermal sensor camera the appearance characteristic may be the object temperature or temperature distribution.
- the microphone the appearance characteristic may be audio frequency or the tone of the sounds of the objects.
- the sensor is the supersonic, laser distance measurement instrument or weight scale
- the appearance characteristic may be the height or weight of the moving object.
- object tracking system 1400 may comprise m sensors of the sensor network, a training-phase processing module 1410 , a characteristic function estimating and updating module 1420 and a detection-phase tracking module 1430 , where m sensors are denoted as sensor 1 to sensor m, m ⁇ 2.
- the coverage range of each sensor j is configured to include at least an entrance/exit, 1 ⁇ j ⁇ m.
- Training-phase processing module 1410 obtains a plurality of measured data through each sensor j as training samples and, for entering events at each entrance/exit of each sensor j, records all leaving events during the past period in a training sample space 1410 a .
- characteristic function estimating and updating module 1420 may estimate at least three characteristic functions related to an object to be tracked, including sensor spatial correlation function 1421 , moving time different function 1422 and appearance similarity function 1423 , via an automatic learning method.
- Detection-phase tracking module 1430 may use the estimated three characteristic functions as the basis for tracking the object and linking relationship of the object.
- the plurality of sensors may be deployed in a non-overlapping sensor network.
- an n ⁇ n training sample space may be allocated in a memory.
- such a training sample space may be represented by an n ⁇ n matrix, where field (d,b) of the n ⁇ n matrix indicates the leaving b events during a specific past period when an object entering d event is observed.
- the above three characteristic functions may be estimated with the above probability function.
- the disclosed exemplary embodiments may provide a system and method for moving object tracking in a non-overlapping sensor network.
- Object tracking may be performed in a sensor network with non-overlapping measurement range, and the disclosed exemplary embodiments do not need to know the scene blueprint of sensor deployment, and require no manual intervention during the learning process.
- the disclosed exemplary embodiments observe a large amount of samples and automatically estimate camera spatial correlation, the distribution of leaving and entering time difference and the distribution of color difference of the object appearance and use the above as the basis for object tracking.
- the sensors may also be other types of sensors.
- the sensors may be color camera, and a camera network is constructed to track the moving object moving within the camera network.
- the sensor may also be black and white camera, thermal sensor camera, infrared camera, microphone, supersonic, laser distance measuring instrument, weight scale, and so on.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Databases & Information Systems (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Medical Informatics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- General Engineering & Computer Science (AREA)
- Image Analysis (AREA)
Abstract
Description
O — i — p=argmaxP(M((i,j),(p,q))|O — i — p,O — j — q), ∀O — i — pεE (1)
O — i — p=argmaxP(Δa(p,q)|M((i,j),(p,q)))P(Δt(p,q)|M((i,j),(p,q)))P(M(i,j)),
∀O — i — pεE (2)
Claims (16)
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
TW098142321A TWI416068B (en) | 2009-12-10 | 2009-12-10 | Object tracking method and apparatus for a non-overlapping-sensor network |
TW98142321A | 2009-12-10 | ||
TW098142321 | 2009-12-10 |
Publications (2)
Publication Number | Publication Date |
---|---|
US20110141288A1 US20110141288A1 (en) | 2011-06-16 |
US8542276B2 true US8542276B2 (en) | 2013-09-24 |
Family
ID=44142473
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/758,812 Active 2032-02-24 US8542276B2 (en) | 2009-12-10 | 2010-04-13 | Object Tracking method and apparatus for a non-overlapping-sensor network |
Country Status (2)
Country | Link |
---|---|
US (1) | US8542276B2 (en) |
TW (1) | TWI416068B (en) |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090244281A1 (en) * | 2008-03-26 | 2009-10-01 | Canon Kabushiki Kaisha | Monitoring apparatus and display processing method for the monitoring apparatus |
US20130101167A1 (en) * | 2011-10-19 | 2013-04-25 | Lee F. Holeva | Identifying, matching and tracking multiple objects in a sequence of images |
US20140043480A1 (en) * | 2011-04-18 | 2014-02-13 | Zte Corporation | Video monitoring system and method |
US8988662B1 (en) * | 2012-10-01 | 2015-03-24 | Rawles Llc | Time-of-flight calculations using a shared light source |
US20150235237A1 (en) * | 2012-02-29 | 2015-08-20 | RetailNext, Inc. | Methods and systems for excluding individuals from retail analytics |
WO2016205951A1 (en) * | 2015-06-25 | 2016-12-29 | Appropolis Inc. | A system and a method for tracking mobile objects using cameras and tag devices |
US20180107898A1 (en) * | 2015-09-30 | 2018-04-19 | International Business Machines Corporation | Classifying and Grouping Electronic Images |
US9990535B2 (en) | 2016-04-27 | 2018-06-05 | Crown Equipment Corporation | Pallet detection using units of physical length |
US10083378B2 (en) | 2015-12-28 | 2018-09-25 | Qualcomm Incorporated | Automatic detection of objects in video images |
US10088549B2 (en) * | 2015-06-25 | 2018-10-02 | Appropolis Inc. | System and a method for tracking mobile objects using cameras and tag devices |
US10514256B1 (en) | 2013-05-06 | 2019-12-24 | Amazon Technologies, Inc. | Single source multi camera vision system |
CN111310535A (en) * | 2018-12-11 | 2020-06-19 | 财团法人工业技术研究院 | Object detection method and object detection device using convolutional neural network model |
US10748033B2 (en) * | 2018-12-11 | 2020-08-18 | Industrial Technology Research Institute | Object detection method using CNN model and object detection apparatus using the same |
US11200683B2 (en) * | 2017-08-22 | 2021-12-14 | Mitsubishi Electric Corporation | Image processing device and image processing method |
US12253620B2 (en) | 2017-02-14 | 2025-03-18 | Microsoft Technology Licensing, Llc | Multi-user intelligent assistance |
Families Citing this family (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9280833B2 (en) | 2013-03-05 | 2016-03-08 | International Business Machines Corporation | Topology determination for non-overlapping camera network |
US8913791B2 (en) | 2013-03-28 | 2014-12-16 | International Business Machines Corporation | Automatically determining field of view overlap among multiple cameras |
TWI489415B (en) * | 2013-07-17 | 2015-06-21 | Vivotek Inc | Video processing method and system |
US9607245B2 (en) * | 2014-12-02 | 2017-03-28 | Xerox Corporation | Adapted vocabularies for matching image signatures with fisher vectors |
US10438277B1 (en) * | 2014-12-23 | 2019-10-08 | Amazon Technologies, Inc. | Determining an item involved in an event |
US10838837B2 (en) * | 2016-06-24 | 2020-11-17 | International Business Machines Corporation | Sensor based system state prediction |
GB2553108B (en) * | 2016-08-22 | 2020-07-15 | Canon Kk | Method, processing device and system for managing copies of media samples in a system comprising a plurality of interconnected network cameras |
JPWO2018128138A1 (en) * | 2017-01-05 | 2019-11-07 | 日本電気株式会社 | Image processing apparatus, video monitoring system, image processing method, and program |
CN108460864B (en) * | 2017-02-22 | 2022-09-02 | 开利公司 | People flow estimation system and fault processing method for same |
TWI638329B (en) * | 2017-06-23 | 2018-10-11 | 修平學校財團法人修平科技大學 | The device and method for detecting customer tracks |
CN109241928B (en) * | 2018-09-19 | 2021-02-26 | 释码融和(上海)信息科技有限公司 | Method and computing device for recognizing heterogeneous irises |
CN110931988A (en) * | 2018-09-20 | 2020-03-27 | 中车株洲电力机车研究所有限公司 | Train sensor network wireless energy emission method and device |
EP3734500A1 (en) * | 2019-05-03 | 2020-11-04 | Axis AB | Method for re-identification of observations |
JP7578107B2 (en) * | 2019-09-25 | 2024-11-06 | 日本電気株式会社 | Item management device, item management system, item management method, and recording medium |
CN113591527B (en) * | 2021-01-14 | 2025-07-08 | 腾讯科技(深圳)有限公司 | Object track recognition method and device, electronic equipment and storage medium |
CN116311019A (en) * | 2022-12-02 | 2023-06-23 | 中国兵器装备集团自动化研究所有限公司 | Personnel loitering detection method, device, equipment and storage medium |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2002086831A2 (en) | 2001-04-19 | 2002-10-31 | Honeywell International Inc. | Method and apparatus for tracking with identification |
TW200708102A (en) | 2005-04-05 | 2007-02-16 | Objectvideo Inc | Wide-area site-based video surveillance system |
CN101090485A (en) | 2006-06-15 | 2007-12-19 | 索尼株式会社 | Image monitoring system and object area tracking method |
TW200806020A (en) | 2006-07-07 | 2008-01-16 | Wen Wen Hung Tech Corp | Image tracking method |
CN101142593A (en) | 2005-03-17 | 2008-03-12 | 英国电讯有限公司 | Method for tracking objects in video sequences |
TW200824433A (en) | 2006-11-22 | 2008-06-01 | Univ Nat Chiao Tung | Intelligent monitoring and analyzing system and method thereof |
US7394916B2 (en) | 2003-02-10 | 2008-07-01 | Activeye, Inc. | Linking tracked objects that undergo temporary occlusion |
TW200907826A (en) | 2007-05-29 | 2009-02-16 | Cognex Corp | System and method for locating a three-dimensional object using machine vision |
TW200915852A (en) | 2007-09-26 | 2009-04-01 | Huper Lab Co Ltd | Object corresponding method capable of real-time and accurate operating for multi-camera construction |
CN101520502A (en) | 2009-03-24 | 2009-09-02 | 中国航空无线电电子研究所 | Method for tracking and positioning mobile node of wireless sensor network |
-
2009
- 2009-12-10 TW TW098142321A patent/TWI416068B/en active
-
2010
- 2010-04-13 US US12/758,812 patent/US8542276B2/en active Active
Patent Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2002086831A2 (en) | 2001-04-19 | 2002-10-31 | Honeywell International Inc. | Method and apparatus for tracking with identification |
US20030040815A1 (en) * | 2001-04-19 | 2003-02-27 | Honeywell International Inc. | Cooperative camera network |
US7149325B2 (en) * | 2001-04-19 | 2006-12-12 | Honeywell International Inc. | Cooperative camera network |
US7394916B2 (en) | 2003-02-10 | 2008-07-01 | Activeye, Inc. | Linking tracked objects that undergo temporary occlusion |
CN101142593A (en) | 2005-03-17 | 2008-03-12 | 英国电讯有限公司 | Method for tracking objects in video sequences |
TW200708102A (en) | 2005-04-05 | 2007-02-16 | Objectvideo Inc | Wide-area site-based video surveillance system |
CN101090485A (en) | 2006-06-15 | 2007-12-19 | 索尼株式会社 | Image monitoring system and object area tracking method |
TW200806020A (en) | 2006-07-07 | 2008-01-16 | Wen Wen Hung Tech Corp | Image tracking method |
TW200824433A (en) | 2006-11-22 | 2008-06-01 | Univ Nat Chiao Tung | Intelligent monitoring and analyzing system and method thereof |
TW200907826A (en) | 2007-05-29 | 2009-02-16 | Cognex Corp | System and method for locating a three-dimensional object using machine vision |
TW200915852A (en) | 2007-09-26 | 2009-04-01 | Huper Lab Co Ltd | Object corresponding method capable of real-time and accurate operating for multi-camera construction |
CN101520502A (en) | 2009-03-24 | 2009-09-02 | 中国航空无线电电子研究所 | Method for tracking and positioning mobile node of wireless sensor network |
Non-Patent Citations (9)
Title |
---|
"Robust Real-Time Unusual Event Detection Using Multiple Fixed-Location Monitors", Amit Adam, Ehud Rivlin, Ilan Shimshoni David Reinitz, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 30, No. 3, Mar. 2008. |
"Video-Based Human Movement Analysis and Its Application to Surveillance Systems", Jun-Wei Hsieh, Yung-Tai Hsu, Hong-Yuan Mark Liao and Chih-Chiang Chen, IEEE Transaction on Multimedia, vol. 10, No. 3 Apr. 2008. |
China Patent Office, Notice of Allowance, Patent Application Serial No. CN201010002105.2, Oct. 8, 2012, China. |
Huang, and S. Russell, "Object identification in a Bayesian context." In Proceedings of IJCAI, 1997. |
Javed, K. Shafique, and M. Shah, "Appearance Modeling for Tracking in Multiple Non-overlapping Cameras." In CVPR, 2005. |
Javed, Z. Rasheed, K. Shafique, and M. Shah, "Tracking across multiple cameras with disjoint views." In ICCV, 2003. |
Makris, T. Ellis, and J. Black, "Bridging the gaps between cameras." In Computer Vision and Pattern Recognition, 2004. |
Rahimi, B. Dunagan, and T. Darrell, "Simultaneous calibration and tracking with a network of non-overlapping sensors," In Computer Vision and Pattern Recognition, 2004. |
Taiwan Patent Office, Office Action, Patent Application Serial No. TW098142321, Apr. 2, 2013, Taiwan. |
Cited By (26)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090244281A1 (en) * | 2008-03-26 | 2009-10-01 | Canon Kabushiki Kaisha | Monitoring apparatus and display processing method for the monitoring apparatus |
US20140043480A1 (en) * | 2011-04-18 | 2014-02-13 | Zte Corporation | Video monitoring system and method |
US8885948B2 (en) | 2011-10-19 | 2014-11-11 | Crown Equipment Corporation | Identifying and evaluating potential center stringers of a pallet in an image scene |
US8934672B2 (en) | 2011-10-19 | 2015-01-13 | Crown Equipment Corporation | Evaluating features in an image possibly corresponding to an intersection of a pallet stringer and a pallet board |
US8938126B2 (en) | 2011-10-19 | 2015-01-20 | Crown Equipment Corporation | Selecting objects within a vertical range of one another corresponding to pallets in an image scene |
US20130101167A1 (en) * | 2011-10-19 | 2013-04-25 | Lee F. Holeva | Identifying, matching and tracking multiple objects in a sequence of images |
US8995743B2 (en) | 2011-10-19 | 2015-03-31 | Crown Equipment Corporation | Identifying and locating possible lines corresponding to pallet structure in an image |
US9025886B2 (en) | 2011-10-19 | 2015-05-05 | Crown Equipment Corporation | Identifying and selecting objects that may correspond to pallets in an image scene |
US9025827B2 (en) | 2011-10-19 | 2015-05-05 | Crown Equipment Corporation | Controlling truck forks based on identifying and tracking multiple objects in an image scene |
US9082195B2 (en) | 2011-10-19 | 2015-07-14 | Crown Equipment Corporation | Generating a composite score for a possible pallet in an image scene |
US9087384B2 (en) * | 2011-10-19 | 2015-07-21 | Crown Equipment Corporation | Identifying, matching and tracking multiple objects in a sequence of images |
US9569786B2 (en) * | 2012-02-29 | 2017-02-14 | RetailNext, Inc. | Methods and systems for excluding individuals from retail analytics |
US20150235237A1 (en) * | 2012-02-29 | 2015-08-20 | RetailNext, Inc. | Methods and systems for excluding individuals from retail analytics |
US8988662B1 (en) * | 2012-10-01 | 2015-03-24 | Rawles Llc | Time-of-flight calculations using a shared light source |
US10514256B1 (en) | 2013-05-06 | 2019-12-24 | Amazon Technologies, Inc. | Single source multi camera vision system |
WO2016205951A1 (en) * | 2015-06-25 | 2016-12-29 | Appropolis Inc. | A system and a method for tracking mobile objects using cameras and tag devices |
US10088549B2 (en) * | 2015-06-25 | 2018-10-02 | Appropolis Inc. | System and a method for tracking mobile objects using cameras and tag devices |
US20180107898A1 (en) * | 2015-09-30 | 2018-04-19 | International Business Machines Corporation | Classifying and Grouping Electronic Images |
US10534978B2 (en) * | 2015-09-30 | 2020-01-14 | International Business Machines Corporation | Classifying and grouping electronic images |
US10083378B2 (en) | 2015-12-28 | 2018-09-25 | Qualcomm Incorporated | Automatic detection of objects in video images |
US9990535B2 (en) | 2016-04-27 | 2018-06-05 | Crown Equipment Corporation | Pallet detection using units of physical length |
US12253620B2 (en) | 2017-02-14 | 2025-03-18 | Microsoft Technology Licensing, Llc | Multi-user intelligent assistance |
US11200683B2 (en) * | 2017-08-22 | 2021-12-14 | Mitsubishi Electric Corporation | Image processing device and image processing method |
CN111310535A (en) * | 2018-12-11 | 2020-06-19 | 财团法人工业技术研究院 | Object detection method and object detection device using convolutional neural network model |
US10748033B2 (en) * | 2018-12-11 | 2020-08-18 | Industrial Technology Research Institute | Object detection method using CNN model and object detection apparatus using the same |
CN111310535B (en) * | 2018-12-11 | 2023-07-14 | 财团法人工业技术研究院 | Object detection method and object detection device using convolutional neural network model |
Also Published As
Publication number | Publication date |
---|---|
US20110141288A1 (en) | 2011-06-16 |
TW201120408A (en) | 2011-06-16 |
TWI416068B (en) | 2013-11-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8542276B2 (en) | Object Tracking method and apparatus for a non-overlapping-sensor network | |
EP1668921B1 (en) | Computerized method and apparatus for determining field-of-view relationships among multiple image sensors | |
US10346688B2 (en) | Congestion-state-monitoring system | |
Vargas et al. | An enhanced background estimation algorithm for vehicle detection in urban traffic scenes | |
US9161084B1 (en) | Method and system for media audience measurement by viewership extrapolation based on site, display, and crowd characterization | |
US9639747B2 (en) | Online learning method for people detection and counting for retail stores | |
US8320613B2 (en) | Detecting and tracking targets in images based on estimated target geometry | |
EP3531386A1 (en) | Fire monitoring system | |
US9443144B2 (en) | Methods and systems for measuring group behavior | |
Sjarif et al. | Detection of abnormal behaviors in crowd scene: a review | |
US20090296989A1 (en) | Method for Automatic Detection and Tracking of Multiple Objects | |
MX2007016406A (en) | Target detection and tracking from overhead video streams. | |
JPH0795625A (en) | Audience measurement system and method | |
CN101965576A (en) | Object matching for tracking, indexing, and search | |
Abdelkader et al. | Integrated motion detection and tracking for visual surveillance | |
WO2013160688A1 (en) | Abnormal object track determination using a gaussian processes based variational bayes expectation maximisation | |
GB2443739A (en) | Detecting image regions of salient motion | |
CN109255360B (en) | A target classification method, device and system | |
US20060072010A1 (en) | Target property maps for surveillance systems | |
WO2009039350A1 (en) | System and method for estimating characteristics of persons or things | |
CN113011371A (en) | Target detection method, device, equipment and storage medium | |
Higa et al. | Robust estimation of product amount on store shelves from a surveillance camera for improving on-shelf availability | |
Daniyal et al. | Content and task-based view selection from multiple video streams | |
Elguebaly et al. | A nonparametric Bayesian approach for enhanced pedestrian detection and foreground segmentation | |
Li et al. | Evaluation of an ivs system for abandoned object detection on pets 2006 datasets |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: INDUSTRIAL TECHNOLOGY RESEARCH INSTITUTE, TAIWAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:HUANG, CHUNG-HSIEN;CHOU, CHENG-CHUAN;WU, RUEI-CHENG;REEL/FRAME:024220/0644 Effective date: 20100401 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |
|
FEPP | Fee payment procedure |
Free format text: 11.5 YR SURCHARGE- LATE PMT W/IN 6 MO, LARGE ENTITY (ORIGINAL EVENT CODE: M1556); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 12 |