US20200334472A1 - Movement state estimation device, movement state estimation method and program recording medium - Google Patents

Movement state estimation device, movement state estimation method and program recording medium Download PDF

Info

Publication number
US20200334472A1
US20200334472A1 US16/921,447 US202016921447A US2020334472A1 US 20200334472 A1 US20200334472 A1 US 20200334472A1 US 202016921447 A US202016921447 A US 202016921447A US 2020334472 A1 US2020334472 A1 US 2020334472A1
Authority
US
United States
Prior art keywords
persons
state
unit
movement
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US16/921,447
Inventor
Hiroyoshi Miyano
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NEC Corp
Original Assignee
NEC Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by NEC Corp filed Critical NEC Corp
Priority to US16/921,447 priority Critical patent/US20200334472A1/en
Publication of US20200334472A1 publication Critical patent/US20200334472A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • G06K9/00778
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • G06V20/53Recognition of crowd images, e.g. recognition of crowd congestion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • G06K9/00362
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • G06T7/248Analysis of motion using feature-based methods, e.g. the tracking of corners or segments involving reference images or patches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/254Analysis of motion involving subtraction of images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30242Counting objects in image
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/01Detecting movement of traffic to be counted or controlled
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/01Detecting movement of traffic to be counted or controlled
    • G08G1/04Detecting movement of traffic to be counted or controlled using optical or ultrasonic detectors

Definitions

  • the present invention relates to a movement state estimation device, a movement state estimation method, and a program recording medium.
  • PTL 1 describes a number-of-persons measurement device that measures the number of persons from a shot video of a crowd.
  • the number-of-persons measurement device described in PTL 1 extracts a head of a person included in the image, based on a head model, connects head positions determined as the same person between frames by using a feature quantity such as position information and color distribution, and measures the number of persons from the connection result.
  • NPL 1 describes a method for estimating the number of persons in a crowd.
  • the method described in NPL 1 seizes a crowd state including overlap of persons, by a crowd-patch that represents the crowd state by a local image, and performs recursive learning of the number of persons in the patch to thereby estimate the number of persons from a static image.
  • PTL 2 describes a traffic quantity measurement system capable of acquiring traffic quantity data at an examination target spot.
  • the system described in PTL 2 identifies, from a captured image of a predetermined examination target region, a passerby in the examination target region, and determines the number of the passersby.
  • NPL 1 uses a crowd-patch described in NPL 1 to enable a crowd in an image to be recognized without depending on a frame rate.
  • a crowd-patch described in NPL 1 it is possible to predict the number of persons that may exist in a predetermined region in an image, but it is difficult to estimate how many of such observing target objects are moving.
  • One of exemplified objects of the present invention is to provide a movement state estimation device, a movement state estimation method, and a program recording medium that can accurately estimate a movement state of a monitoring target even in a crowded environment.
  • a movement state estimation device is characterized by including: quantity estimation means for, by using a plurality of temporally sequential images, estimating a quantity of monitoring targets for each local region in each of the plurality of images; and movement state estimation means for estimating a movement state of the monitoring target from a time-series change in the quantity estimated in each of the local regions.
  • a movement state estimation method is characterized by including: by using a plurality of temporally sequential images, estimating a quantity of monitoring targets for each local region in each of the images; and estimating a movement state of the monitoring target from a time-series change in the quantity estimated in each of the local regions.
  • a program recording medium is characterized by recording a program for causing a computer to execute: a quantity estimation process of, by using a plurality of temporally sequential images, estimating a quantity of monitoring targets for each of local regions in the plurality of images; and a movement state estimation process of estimating a movement state of the monitoring target from time-series change in the quantity estimated in each of the local regions.
  • a movement state of a monitoring target can be accurately estimated even in a crowded environment.
  • FIG. 1 is a block diagram exemplifying one example embodiment of a movement state estimation device.
  • FIG. 2 is an illustration depicting an example of a process of estimating a quantity of monitoring targets.
  • FIG. 3 is an illustration depicting an example of a movement state of monitoring targets.
  • FIG. 4 is an illustration depicting relation between a local region for which a quantity of monitoring targets has been estimated and particles existing in this local region.
  • FIG. 5 is an illustration depicting an example of a process of updating weight values when local regions overlap each other.
  • FIG. 6 is an illustration depicting an example of a condition in which detection probability varies.
  • FIG. 7 is an illustration depicting an example of a process of calculating a quantity of monitoring targets that have crossed over a predetermined location.
  • FIG. 8 is a flowchart illustrating an operation example of a movement state estimation device.
  • FIG. 9 is a block diagram exemplifying an outline of a movement state estimation device.
  • FIG. 10 is a block diagram exemplifying a hardware configuration of a computer device.
  • a monitoring target mentioned in the present invention is exemplified by a person in the following description of the example embodiment, but may also be an object other than a person, such as a bicycle and an automobile.
  • FIG. 1 is a block diagram illustrating one example embodiment of a movement state estimation device according to the present invention.
  • the movement state estimation device 100 of the present example embodiment includes an image input unit 11 , a number-of-persons estimation unit 12 , a flow calculation unit 13 , a state storage device 14 , a state prediction unit 15 , a staying information calculation unit 16 , an individual person detection unit 17 , a state update unit 18 , and a number-of-persons output unit 19 .
  • the arrows illustrated in this drawing indicate one example of flows of data. Flows of data in the movement state estimation device 100 are not limited to particular directions.
  • the image input unit 11 acquires, from a video (a moving image), an image at a time point of certain processing time. In the following, an image acquired by the image input unit 11 is written as “target image”. The image input unit 11 receives input of a plurality of time sequence target images.
  • the number-of-persons estimation unit 12 estimates the number of photographed persons for each local region in a target image. In other words, the number-of-persons estimation unit 12 estimates the number of persons (a quantity of monitoring targets) for each local region of an input target image.
  • a method by which the number-of-persons estimation unit 12 estimates the number of persons is not particularly limited.
  • the number-of-persons estimation unit 12 may compare a crowd-patch described in NPL 1 with a local region of a target image to estimate the number of persons, or may estimate the number of persons by using a plurality of images including a target image as in a method described in PTL 1.
  • FIG. 2 is an illustration depicting an example of a process of estimating the number of persons.
  • the number-of-persons estimation unit 12 extracts a local region 21 from a target image 20 , and estimates the number of persons included in the local region 21 .
  • the number-of-persons estimation unit 12 estimates that four persons exist in the local region 21 .
  • the number-of-persons output unit 19 may output an image in which the number of persons estimated for each local region is expressed by using a mode (color, tone, or the like) associated with this number of persons, as exemplified by a density map 30 in FIG. 2 .
  • a mode color, tone, or the like
  • a concrete process performed by the number-of-persons output unit 19 is described below.
  • the number-of-persons estimation unit 12 can detect a fluctuation (time-series change) in the number of persons in each local region. It is considered that particularly in a crowded environment, persons move in a certain degree of group. For this reason, the number-of-persons estimation unit 12 estimates a movement state of persons from transition of the number of persons in a local region.
  • the number-of-persons estimation unit 12 predicts future locations of persons, from the movement of the persons at the time when the number of persons is estimated.
  • the number-of-persons estimation unit 12 may assume that monitoring targets move at the same probability and at the same speed in all directions, in an initial state.
  • the number-of-persons estimation unit 12 estimates the future number of persons in each local region, as well. For the convenience of description, to estimate the future number of persons by the number-of-persons estimation unit 12 is referred to as “predict” in the following.
  • the number-of-persons estimation unit 12 may assume that each person moves at the same speed, or may use a below-described prediction result of the state prediction unit 15 .
  • the number-of-persons estimation unit 12 compares the number of persons previously predicted for each local region with the number of persons estimated for each local region from a target image at this time point. Then, the number-of-persons estimation unit 12 estimates a movement state of persons, placing importance on the local region where a difference in the number of persons is smaller. Concretely, the number-of-persons estimation unit 12 may estimate that a person existing in a particular local region has moved to another local region that is one of a plurality of local regions near the particular local region and in which a difference in the number of persons estimated from the target image is the smallest.
  • the density map 30 exemplified in FIG. 2 represents a movement state of persons.
  • FIG. 3 is an illustration depicting an example of a movement state of persons.
  • a density map exemplified in FIG. 3 indicates that more persons exist in a lighter color region.
  • the number-of-persons estimation unit 12 estimates that many persons exist in the region 31 a and the region 31 b in a target image at a certain time point. It is assumed that the number-of-persons estimation unit 12 continues to perform the estimation, and as a result, many persons move to the region 32 a and the region 32 b in a target image at another time point. It becomes possible to detect, from this time-series change, a movement (the arrow 33 a ) of persons from the region 31 a to the region 32 a , and a movement (the arrow 33 b ) of persons from the region 31 b to the region 32 b.
  • the number-of-persons estimation unit 12 may determine whether a person included in a target image at a certain time point is a person who newly appears at this time point, or an already existing person who has been present in the target image before this time point (i.e., a person who has moved within the shot region). For example, when estimating existence of a person at a location to which a person has not been predicted to move from the target image, the number-of-persons estimation unit 12 may determine that this person is a newly appearing person.
  • the flow calculation unit 13 calculates an optical flow for target images.
  • a method for calculating an optical flow is not limited particularly.
  • the flow calculation unit 13 may calculate an optical flow by using a feature point as in the Lucas-Kanade method, or may calculate an optical flow by using the calculus of variations as in the Horn-Schunck method.
  • the state storage device 14 stores past states of persons.
  • a state of a person includes a location, a velocity, and likelihood (a weight value) of a person at a past time point.
  • a state of a person may include a behavior state (a moving state or a stationary state) of a person.
  • a state of one monitoring target is represented by a plurality of particles.
  • the state storage device 14 may hold, as a state of each particle, a variable s i expressing whether a person is in a moving state or in a stationary state.
  • the total sum of weight values w i of the particles corresponds to a quantity of monitoring targets, i.e., the number of persons. For example, in a PHD particle filter, when a hundred of particles are newly scattered for one monitoring target, a weight value of each particle set at this time is “0.01”.
  • the state prediction unit 15 predicts a state of a person at the time that the image input unit 11 acquires an image, by using a past state of the person stored in the state storage device 14 . Other than a location, a velocity, and a weight value of the person, the state prediction unit 15 may predict whether a behavior state of the person is a moving state or a stationary state.
  • the state prediction unit 15 predicts a future state of a person by using a plurality of particles expressing a state of this person.
  • the description is made on a method in which the state prediction unit 15 predicts a state of an object by using a PHD particle filter.
  • the state prediction unit 15 predicts a state of a person at the time that the image input unit 11 acquires a target image.
  • a state of one monitoring target is expressed by a plurality of particles.
  • a state of the particle may be expressed further by a variable s i expressing whether a person is in a moving state or in a stationary state.
  • a method by which the state prediction unit 15 predicts a location and a velocity is similar to a prediction method performed in a general particle filter.
  • a location of a particle at a certain time point is x
  • a location of this particle after a lapse of dt is expressed by x+v ⁇ dt+e.
  • e expresses a noise that cannot be expressed in a constant velocity motion model, and for example, e is a random number value generated based on normal distribution of predetermined standard deviation.
  • v a velocity of a particle at the location x
  • v a velocity of this particle after a lapse of dt
  • f expresses a noise that cannot be expressed in a constant velocity motion model, and for example, f is a random number value generated based on normal distribution of predetermined standard deviation.
  • the following describes a method in which the state prediction unit 15 predicts a behavior state.
  • a behavior state of a person changes from a stationary state to a moving state in accordance with predetermined probability P, and a behavior state changes from a moving state to a stationary state based on predetermined probability Q.
  • the state prediction unit 15 When s i expresses a stationary state, the state prediction unit 15 generates a uniform random number from 0 to 1, and when this random number value is equal to or smaller than P, the state prediction unit 15 changes s i to a value indicating a moving state. Meanwhile, when s i expresses a moving state, the state prediction unit 15 generates a uniform random number from 0 to 1, and when this random number value is equal to or smaller than Q, the state prediction unit 15 changes s i to a value indicating a stationary state.
  • the state prediction unit 15 may predict a state of a variable s i based on a past history and a statistical result. For example, it is assumed that a person of a tracking target has changed from being in a stationary state to being in a moving state, and s i has become a value indicating the moving state. In the case where this change has been made recently, and a predetermined period is yet to lapse from the change to the moving state, it can be supposed that this tracking target does not return to being in a stationary state soon. For this reason, in this case, the state prediction unit 15 may hold the behavior state in the moving state for a fixed period.
  • the state prediction unit 15 may hold the behavior state in the stationary state for a fixed period.
  • the state prediction unit 15 may change a condition for changing a behavior state, depending on a location in an image. For example, in a region in an image where a pathway exists, the state prediction unit 15 may set, at a small value, change probability Q of change from a moving state to a stationary state. In a region in an image including an area often used for a rendezvous, the state prediction unit 15 may set, at a large value, the change probability Q.
  • the staying information calculation unit 16 extracts a region determined as a staying object, by using a target image acquired by the image input unit 11 .
  • the staying information calculation unit 16 may determine, for each pixel, whether or not an object is a temporarily stationary one, by using a method for detecting an object stationary for a long period, and may detect a region that is determined as the temporarily stationary object, by a labeling process.
  • the individual person detection unit 17 detects an individual person from a target image. For example, a person existing on a front side in an image can be often detected by a general detection method because of being less obstructed by other objects. For this reason, the individual person detection unit 17 sets a region (referred to as an individual detection region in the following) where a person can be individually detected, and then detects a person from the individual detection region in a target image.
  • an individual detection region referred to as an individual detection region in the following
  • the state update unit 18 updates a state of a person predicted by the state prediction unit 15 , based on an estimation result of the number-of-persons estimation unit 12 .
  • the state update unit 18 may update a state of a person based on process results of the flow calculation unit 13 , the staying information calculation unit 16 , and the individual person detection unit 17 .
  • the state update unit 18 updates a state of a person in the corresponding region.
  • the state update unit 18 updates a weight value w i of a particle predicted to exist in the corresponding region.
  • the state update unit 18 updates a weight value w i of each particle such that the total sum of weights in the local region becomes equal to the number of persons.
  • the state update unit 18 may update a weight value w i of each particle by using a, a predetermined value between 0 and 1, such that the total sum of updated weights becomes (1 ⁇ ) ⁇ (the last total sum of weights)+ ⁇ (the number of persons) ⁇ so as to cause a weight value w i to change gradually.
  • the state update unit 18 may update weight values of particles such that the total sum of weight values approaches the number of persons in the local region.
  • FIG. 4 is an illustration depicting a relation between a local region whose number of persons have been estimated and particles existing in this local region.
  • white and black circles represent particles.
  • the black particles represent particles whose moving directions (the arrows in the drawing) are the same as the specific direction or within a predetermined angle from the specific direction.
  • persons existing in a local region 41 a of a target image and having moving directions that are the same as the specific direction or within the predetermined angle from the specific direction are four persons.
  • Persons existing in a local region 41 b and having moving directions that are the same as the specific direction or within the predetermined angle from the specific direction are three persons.
  • the state update unit 18 updates a weight value of each particle such that the total sum of the weight values becomes two.
  • the state update unit 18 may update weight values of particles individually for each local region, or may update weight values of particles all together for one region into which the local regions overlapping each other are combined. When combining the regions into one to update weight values, the state update unit 18 may set the number of persons in the overlapped region at the number of persons taking into account the number of persons for each local region (e.g., the average number of persons of the overlapped local regions).
  • FIG. 5 is an illustration depicting an example of a process of updating weight values when local regions overlap each other.
  • the example illustrated in FIG. 5 represents the case of estimating the number of persons for three local regions partially overlapping each other.
  • the state update unit 18 may update weight values of particles all together for one region into which the local region 42 a , the local region 42 b , and the local region 42 c are combined.
  • the state update unit 18 may update weights of particles included in each local region individually for the local region 42 a , the local region 42 b , and the local region 42 c , respectively.
  • the state update unit 18 may update weight values of particles included in a local region, at the same ratio for the all or at ratios different from particle to particle. For example, depending on detection probability of persons, the state update unit 18 may change weight values of particles to be updated. Concretely, when the detection probability of each particle is Pi, and the estimated number of persons is H, the state update unit 18 may calculate a weight value w i of each particle to be updated, as [(1 ⁇ P i )*w i +H*(P i *w i )/ ⁇ the total sum of (P i *w i ) ⁇ ]
  • the state update unit 18 can set a weight value that is more influenced by the estimated number H of persons as detection probability becomes higher. In a case with a location where no persons are detected, since detection probability becomes zero, a weight value does not change. Changing a weight value to be updated, depending on detection probability, enables the update process to be performed prioritizing information of a region having higher detection probability, for example for an image shot from a low depression angle.
  • FIG. 6 is an illustration depicting an example of a condition in which detection probability varies.
  • a far end side in a target image is often shot at a low depression angle.
  • detection probability of a circumference 44 of a local region 43 decreases from a near side toward a back side.
  • the state update unit 18 updates a weight of a particle existing on the front side to be higher, and updates a weight of a particle existing on the back side to be lower.
  • the state update unit 18 When no particles exist at a location corresponding to a local region for which the number-of-persons estimation unit 12 has estimated the number of persons, the state update unit 18 newly generates a particle at this location. Concretely, the state update unit 18 may newly generate a particle randomly near this location in accordance with normal distribution of predetermined standard deviation.
  • the state update unit 18 may perform update so as to increase a weight value w i of a particle whose movement is close to an optical flow of target images. At this time, the state update unit 18 update a weight value w i so as not to change the total sum of the weights.
  • the state update unit 18 may update a weight value of each particle by [L i *w i / ⁇ the total sum of (L i *w i ) ⁇ ]*S. By performing such update, a particle whose movement is closer to the optical flow has a larger weight value w i .
  • a method for updating a weight value of a particle whose movement is closer to the optical flow is not limited to the above-described method.
  • the state update unit 18 may simply multiply a weight value of the particle by a positive constant, and then normalize each weight value such that the total sum of the weight values multiplied by the constant becomes equal to the total sum S of the original weight values.
  • the state update unit 18 may determine, as a particle of close movement, a particle for which a distance between the vectors is equal to or smaller than a threshold value.
  • the state update unit 18 may update a weight value of a particle depending on a closeness degree between a region determined as a staying object by the staying information calculation unit 16 and a behavior state of the particle. Concretely, the state update unit 18 may perform update so as to more increase a weight value of a particle as a distance between a region determined as a staying object by the staying information calculation unit 16 and the particle whose behavior state has been predicted as a stationary state is shorter. At this time, the state update unit 18 may update a weight value of the particle, for example by using a method similar to the method of updating a weight value of a particle whose movement is close to the optical flow.
  • the state update unit 18 may update weight values of particles existing in a region associated with the detected person, in a general method performed in a PHD particle filter. In other words, for an area where the individual person detection unit 17 has detected persons, the state update unit 18 may update weight values of particles included in each local region, by a method other than the method using the estimated number of persons in each local region. Selecting the more suitable prediction method depending on a region in a shot image in this manner makes it possible to improve accuracy in tracking a monitoring target.
  • the state update unit 18 deletes a particle whose weight value w i is equal to or smaller than a threshold value.
  • the state update unit 18 updates location information of a person in a manner similar to a typical tracking process.
  • the state update unit 18 records an updated state of a person in the state storage device 14 .
  • the state update unit 18 may perform re-sampling of scattering particles again depending on weight values of the respective particles.
  • the number-of-persons output unit 19 outputs the number of persons in a shot area, based on states of persons. Concretely, the number-of-persons output unit 19 outputs the number of persons included in a target image, by using weight values w i of particles updated by the state update unit 18 .
  • the number-of-persons output unit 19 may calculate the total sum of weight values w i of particles and output the number of persons in a shot area.
  • the number-of-persons output unit 19 may identify particles existing in this region, and may calculate the total sum of weight values w i of the identified particles.
  • the number-of-persons output unit 19 may calculate the total sum of weight values w i of the particles in a stationary state, and may output the number of persons staying in a shot area, for example.
  • the number-of-persons output unit 19 may output the number of persons that have moved in a specific direction, and the number of persons that have crossed over a line set at a predetermined location in a shot area (i.e., the number of persons that have crossed over the predetermined line), by using states of current particles and states of past particles. Concretely, the number-of-persons output unit 19 outputs the number of persons that have crossed over the predetermined location, based on weighs of particles that have crossed over this location. The following describes a method of calculating the number of persons by using temporally changing particles.
  • the number-of-persons output unit 19 may identify particles whose past and current locations are connected by a line whose direction is the same as the specific direction or makes a predetermined angle or less with the specific direction, and may calculate the total sum of weight values of these particles. At the time of calculating the total sum of weight values of the particles, the number-of-persons output unit 19 may use weight values of past particles, may use weight values of current particles, or may use an average of the weight values of both of the particles.
  • the number-of-persons output unit 19 When calculating the number of persons that have crossed over the predetermined line, the number-of-persons output unit 19 identifies particles that have crossed over the predetermined line in the course of moving from the last location to the present location, and calculates the total sum of weight values of these particles. For example, when among particles that cross over the predetermined line, only the particles that have moved in a specific direction are targeted, the number-of-persons output unit 19 may target particles whose moving directions are each expressed by a vector that makes an inner product equal to or larger than zero, with a normal vector of the predetermined line.
  • the number-of-persons output unit 19 calculates an inner product of a normal vector (a vector directed just below in this case) to the horizontal straight line and a vector expressing a moving direction of each particle, and targets particles for which the inner product is equal to or larger than zero.
  • the number-of-persons output unit 19 can output the number of persons that have crossed over the predetermined line in a predetermined period, by integrating the total sums of weight values in the predetermined period.
  • the number-of-persons output unit 19 can use current weight values, past weight values, or both thereof, as weight values of particles used in calculating the total sums of weight values, in the same manner as in the case of calculating the number of persons that have moved in the specific direction.
  • Particles newly generated at the current time point do not have past states, and for this reason, are not included in targets for calculating these numbers of persons.
  • FIG. 7 is an illustration depicting an example of a process of calculating a quantity of monitoring targets that have crossed over a predetermined location.
  • the circles of the broken lines exemplified in FIG. 7 represent past particles, the circles of the solid lines represent current particles.
  • the arrows exemplified in FIG. 7 represent a movement state of particles from the past time to the current time.
  • the number-of-persons output unit 19 identifies particles that have crossed over the solid line 52 from above to below, and calculates the total sum of weights of the identified particles.
  • the particle 50 b , the particle 50 c , and the particle 50 d cross over the solid line 52 from above to below.
  • the number-of-persons output unit 19 calculates the total sum of particle weights of the particle 50 b , the particle 50 c , and the particle 50 d , and outputs the total sum as the passing number of persons.
  • the number-of-persons output unit 19 may integrate the total sums of weights of particles in the fixed period.
  • the image input unit 11 , the number-of-persons estimation unit 12 , the flow calculation unit 13 , the state prediction unit 15 , the staying information calculation unit 16 , the individual person detection unit 17 , the state update unit 18 , and the number-of-persons output unit 19 can be implemented by programs. These respective units can be implemented by a processor of a computer operating in accordance with the programs.
  • FIG. 10 is a block diagram exemplifying a hardware configuration of a computer device 200 implementing a movement state estimation device 100 .
  • the computer device 200 includes a central processing unit (CPU) 201 , a read-only memory (ROM) 202 , a random access memory (RAM) 203 , a storage device 204 , a drive device 205 , a communication interface 206 , and input-output interface 207 .
  • the movement state estimation device 100 can be implemented by the configuration (or a part thereof) illustrated in FIG. 10 .
  • the CPU 201 executes a program 208 by using the RAM 203 .
  • the program 208 may be stored in the ROM 202 .
  • the program 208 may be recorded in a recording medium 209 such as a flash memory, and may be read out by the drive device 205 , or may be transmitted from an external device via a network 210 .
  • the communication interface 206 exchanges data with the external device via the network 210 .
  • the input-output interface 207 exchanges data with peripheral devices (an input device, a display device, and the like).
  • the communication interface 206 and the input-output interface 207 can function as means for acquiring or outputting data.
  • the movement state estimation device 100 may be configured by a single piece of circuitry (a processor or the like), or may be configured by a combination of a plurality of pieces of circuitry.
  • the circuitry mentioned here may be dedicated or multipurpose circuitry.
  • the CPU 201 may function as the image input unit 11 , the number-of-persons estimation unit 12 , the flow calculation unit 13 , the state prediction unit 15 , the staying information calculation unit 16 , the individual person detection unit 17 , the state update unit 18 , and the number-of-persons output unit 19 , in accordance with the program 208 .
  • Each of the image input unit 11 , the number-of-persons estimation unit 12 , the flow calculation unit 13 , the state prediction unit 15 , the staying information calculation unit 16 , the individual person detection unit 17 , the state update unit 18 , and the number-of-persons output unit 19 may be implemented by dedicated hardware.
  • the state storage device 14 may be implemented by the storage device 204 , or may be an external device connected via the communication interface 206 .
  • FIG. 8 is a flowchart illustrating an operation example of the movement state estimation device 100 according to the present example embodiment.
  • the number-of-persons estimation unit 12 estimates the number of persons for each local region in each image using a plurality of temporally sequential images (step S 11 ). Then, the number-of-persons estimation unit 12 estimates a movement state of persons, from time-series change of the quantity estimated for each local region (step S 12 ).
  • the number-of-persons estimation unit 12 predicts future locations of persons, from the movement state of persons at the time point that the estimation is made, and predicts the future number of persons in each local region.
  • the movement state of persons can be determined from states of particles expressing persons, for example.
  • the number-of-persons estimation unit 12 can predict the future number of persons from future locations of particles predicted by the state prediction unit 15 , for example. Further, for each local region, the number-of-persons estimation unit 12 places more importance on the local region where a difference between the predicted number of persons and the estimated number of persons is small.
  • the number-of-persons estimation unit 12 estimates the number of persons for each local region in each image by using a plurality of the temporally sequential images, and estimates a movement state of persons from the time-series change in the number of persons estimated for each local region. For this reason, according to the present example embodiment, even in a crowded environment where it is difficult to track individual persons, a movement state of persons can be accurately estimated.
  • the state prediction unit 15 predicts future states of persons by using a plurality of particles expressing states of persons, and the number-of-persons output unit 19 calculates the number of persons by targeting particles having weights.
  • states of monitoring targets are tracked by using particles having weight values so that the number of persons that have crossed over a specific place can be counted.
  • not only the number of persons that have simply crossed over a specific place, but also the number of persons that move in a specific direction can be counted. Thereby, it becomes possible to measure not only a simple crowdedness degree in a certain location, but also a flow quantity (a person flow) at this location.
  • FIG. 9 is a block diagram illustrating the outline of a movement state estimation device in the present example embodiment.
  • the movement state estimation device illustrated in FIG. 9 includes a quantity estimation unit 81 estimating a quantity of monitoring targets for each local region in each image by using a plurality of temporally sequential images, and a movement state estimation unit 82 estimating a movement state of the monitoring targets from time-series change in the estimated quantity for each local region.
  • the quantity estimation unit 81 and the movement state estimation unit 82 correspond to the number-of-persons estimation unit 12 in the above-described example embodiment.
  • the thus-configured movement state estimation device can accurately estimate a movement state of monitoring targets even in a crowded environment.
  • the movement state estimation unit 82 may predict future locations of monitoring targets, from a movement state of the monitoring targets at the time point that the estimation is made, and may predict a future quantity of the monitoring targets in each local region.
  • the movement state estimation unit 82 may compare the future quantity of the monitoring targets predicted for each local region, with the quantity of the monitoring targets estimated for each local region from a target image at the future time point, and may estimate a movement state of the monitoring targets, placing importance on the local region where the difference in the quantity is smaller.
  • the movement state estimation unit 82 may determine whether a shot monitoring target in a target image is a new monitoring target or a monitoring target that has moved within the target image, to estimate a movement state of the monitoring targets.
  • the movement state estimation device may include a prediction unit (e.g., the state prediction unit 15 ) predicting future states of monitoring targets by using a plurality of particles expressing states of the monitoring targets.
  • the movement state estimation device may include an update unit (e.g., the state update unit 18 ) updating, depending on an estimated quantity of monitoring targets in each local region, weight values of the particles predicted to exist in each corresponding local region.
  • the movement state estimation unit 82 may estimate a movement state of monitoring targets from time-series change in the total sum of weight values of particles included in each local region. According to such a configuration, it becomes possible to deal with various states of movement performed by monitoring targets.
  • the update unit may update weight values to be set for particles, such that the total sum of weight values to be set for particles predicted to be included in a local region approaches an estimated quantity of monitoring targets in the corresponding local region.
  • weight values are set so as to amount to one in total.
  • For particles expressing a state of a monitoring target at least locations and velocities for the monitoring target are set. Based on the locations and the velocities set for the particles, the prediction unit predicts a future location of the monitoring target.
  • the movement state estimation device may include a quantity output unit (e.g., the number-of-persons output unit 19 ) outputting a quantity of monitoring targets depending on weights set for particles.
  • a quantity output unit e.g., the number-of-persons output unit 19
  • the quantity output unit may output a quantity of monitoring targets that have crossed over a predetermined location, based on weights of particles that have crossed over this location.
  • the movement state estimation device may include a flow calculation unit (e.g., the flow calculation unit 13 ) calculating an optical flow of target images.
  • the update unit may perform update so as to increase weight values of particles whose movement is close to the optical flow. According to such a configuration, it becomes possible to estimate a movement state, placing importance on particles whose movement is close to movement estimated from images.
  • the movement state estimation device may include a staying information calculation unit (e.g., the staying information calculation unit 16 ) extracting a region determined as a staying object from target images.
  • the prediction unit may predict future behavior states of monitoring targets, and as the particles whose behavior states have been predicted to be a stationary state are closer to a region determined as a staying object, the update unit may update weight values of these particles so as to more increase. According to such a configuration, states of monitoring targets that are not moving can be determined appropriately.
  • the movement state estimation device may include a monitoring target detection unit detecting a monitoring target from a target image (concretely, an individual detection region set, as an area where a monitoring target can be individually detected, in the target image).
  • the update unit may update weight values set for particles predicted to be included in each local region thereof, by a method other than a method using an estimated quantity of monitoring targets in each local region. Examples of such a method include a method in which a general PHD particle filter updates weight values.
  • the movement state estimation method according to the present invention is combined with a method actually detecting a monitoring target so that accuracy in estimating a movement state of monitoring targets can be improved.
  • the present invention is described by citing as a typical example the above-described example embodiment.
  • the present invention is not limited to the above-described example embodiment.
  • various configurations that can be understood by those skilled in the art can be adopted in the present invention within the scope of the present invention.
  • the present invention is applicable to a movement state estimation device estimating a quantity of moving objects.
  • the present invention is applicable to a device estimating a flow of objects such as persons or vehicles, or the number of objects crossing over a specific spot, from a video taken with a camera in a monitoring system or the like using a fixed camera or the like.

Abstract

[Problem] To provide a motion condition estimation device, a motion condition estimation method and a motion condition estimation program capable of accurately estimating the motion condition of monitored subjects even in a crowded environment. [Solution] A motion condition estimation device according to the present invention is provided with a quantity estimating means and a motion condition estimating means. The quantity estimating means uses a plurality of chronologically consecutive images to estimate a quantity of monitored subjects for each local region in each image. The motion condition estimating means estimates the motion condition of the monitored subjects from chronological changes in the quantities estimated in each local region.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of U.S. application Ser. No. 16/296,516, filed Mar. 8, 2019, which is a continuation of U.S. application Ser. No. 15/543,408, now U.S. Pat. No. 10,325,160, filed Jul. 13, 2017, based on National Stage of International Application No. PCT/JP2016/000146 filed Jan. 13, 2016, claiming priority based on Japanese Patent Application No. 2015-004963 filed Jan. 14, 2015, the contents of all of which are incorporated herein by reference in their entirety.
  • TECHNICAL FIELD
  • The present invention relates to a movement state estimation device, a movement state estimation method, and a program recording medium.
  • BACKGROUND ART
  • Various techniques for recognizing a crowd in an image have been proposed. For example, PTL 1 describes a number-of-persons measurement device that measures the number of persons from a shot video of a crowd. The number-of-persons measurement device described in PTL 1 extracts a head of a person included in the image, based on a head model, connects head positions determined as the same person between frames by using a feature quantity such as position information and color distribution, and measures the number of persons from the connection result.
  • NPL 1 describes a method for estimating the number of persons in a crowd. The method described in NPL 1 seizes a crowd state including overlap of persons, by a crowd-patch that represents the crowd state by a local image, and performs recursive learning of the number of persons in the patch to thereby estimate the number of persons from a static image.
  • PTL 2 describes a traffic quantity measurement system capable of acquiring traffic quantity data at an examination target spot. The system described in PTL 2 identifies, from a captured image of a predetermined examination target region, a passerby in the examination target region, and determines the number of the passersby.
  • CITATION LIST Patent Literature
    • [PTL 1] Japanese Laid-open Patent Publication No. 2010-198566
    • [PTL 2] Japanese Laid-open Patent Publication No. 2001-76291
    Non Patent Literature
    • [NPL 1] Ikeda Hiroo, Oami Ryoma, “Number of People Estimation Based on Crowd-Patch Learning”, Proceedings of the 12th Forum on Information Technology, Information Processing Society of Japan, pp. 129-130, September 2013
    SUMMARY OF INVENTION Technical Problem
  • In a crowded environment, overlaps of persons in a photographed image are large, and it is difficult to detect individual persons and track individual persons. For this reason, it is difficult to determine how many persons are moving. For example, in a case that the method described in PTL 1 is used, a head position that has been determined as the same person cannot be tracked if a frame rate is low. Thus, it is difficult to measure a flow of photographed persons and the number of the photographed persons. The system described in PTL 2 also has difficulty in recognizing individual passersby in an examination target region in a crowded environment. For this reason, the system described in PTL 2 has difficulty in appropriately grasping a quantity of passersby in the region.
  • Meanwhile, using a crowd-patch described in NPL 1 enables a crowd in an image to be recognized without depending on a frame rate. However, when a crowd-patch described in NPL 1 is used, it is possible to predict the number of persons that may exist in a predetermined region in an image, but it is difficult to estimate how many of such observing target objects are moving.
  • One of exemplified objects of the present invention is to provide a movement state estimation device, a movement state estimation method, and a program recording medium that can accurately estimate a movement state of a monitoring target even in a crowded environment.
  • Solution to Problem
  • A movement state estimation device according to the present invention is characterized by including: quantity estimation means for, by using a plurality of temporally sequential images, estimating a quantity of monitoring targets for each local region in each of the plurality of images; and movement state estimation means for estimating a movement state of the monitoring target from a time-series change in the quantity estimated in each of the local regions.
  • A movement state estimation method according to the present invention is characterized by including: by using a plurality of temporally sequential images, estimating a quantity of monitoring targets for each local region in each of the images; and estimating a movement state of the monitoring target from a time-series change in the quantity estimated in each of the local regions.
  • A program recording medium according to the present invention is characterized by recording a program for causing a computer to execute: a quantity estimation process of, by using a plurality of temporally sequential images, estimating a quantity of monitoring targets for each of local regions in the plurality of images; and a movement state estimation process of estimating a movement state of the monitoring target from time-series change in the quantity estimated in each of the local regions.
  • Advantageous Effects of Invention
  • According to the present invention, a movement state of a monitoring target can be accurately estimated even in a crowded environment.
  • BRIEF DESCRIPTION OF DRAWINGS
  • FIG. 1 is a block diagram exemplifying one example embodiment of a movement state estimation device.
  • FIG. 2 is an illustration depicting an example of a process of estimating a quantity of monitoring targets.
  • FIG. 3 is an illustration depicting an example of a movement state of monitoring targets.
  • FIG. 4 is an illustration depicting relation between a local region for which a quantity of monitoring targets has been estimated and particles existing in this local region.
  • FIG. 5 is an illustration depicting an example of a process of updating weight values when local regions overlap each other.
  • FIG. 6 is an illustration depicting an example of a condition in which detection probability varies.
  • FIG. 7 is an illustration depicting an example of a process of calculating a quantity of monitoring targets that have crossed over a predetermined location.
  • FIG. 8 is a flowchart illustrating an operation example of a movement state estimation device.
  • FIG. 9 is a block diagram exemplifying an outline of a movement state estimation device.
  • FIG. 10 is a block diagram exemplifying a hardware configuration of a computer device.
  • DESCRIPTION OF EMBODIMENTS Example Embodiment
  • The following describes an example embodiment of the present invention with reference to the drawings. A monitoring target mentioned in the present invention is exemplified by a person in the following description of the example embodiment, but may also be an object other than a person, such as a bicycle and an automobile.
  • FIG. 1 is a block diagram illustrating one example embodiment of a movement state estimation device according to the present invention. The movement state estimation device 100 of the present example embodiment includes an image input unit 11, a number-of-persons estimation unit 12, a flow calculation unit 13, a state storage device 14, a state prediction unit 15, a staying information calculation unit 16, an individual person detection unit 17, a state update unit 18, and a number-of-persons output unit 19. The arrows illustrated in this drawing indicate one example of flows of data. Flows of data in the movement state estimation device 100 are not limited to particular directions.
  • The image input unit 11 acquires, from a video (a moving image), an image at a time point of certain processing time. In the following, an image acquired by the image input unit 11 is written as “target image”. The image input unit 11 receives input of a plurality of time sequence target images.
  • The number-of-persons estimation unit 12 estimates the number of photographed persons for each local region in a target image. In other words, the number-of-persons estimation unit 12 estimates the number of persons (a quantity of monitoring targets) for each local region of an input target image.
  • A method by which the number-of-persons estimation unit 12 estimates the number of persons is not particularly limited. For example, the number-of-persons estimation unit 12 may compare a crowd-patch described in NPL 1 with a local region of a target image to estimate the number of persons, or may estimate the number of persons by using a plurality of images including a target image as in a method described in PTL 1.
  • FIG. 2 is an illustration depicting an example of a process of estimating the number of persons. As exemplified in FIG. 2, the number-of-persons estimation unit 12 extracts a local region 21 from a target image 20, and estimates the number of persons included in the local region 21. In an example illustrated in FIG. 2, the number-of-persons estimation unit 12 estimates that four persons exist in the local region 21.
  • The number-of-persons output unit 19 may output an image in which the number of persons estimated for each local region is expressed by using a mode (color, tone, or the like) associated with this number of persons, as exemplified by a density map 30 in FIG. 2. A concrete process performed by the number-of-persons output unit 19 is described below.
  • As described above, since a plurality of temporally sequential target images are input to the number-of-persons estimation unit 12 from the image input unit 11, the number-of-persons estimation unit 12 can detect a fluctuation (time-series change) in the number of persons in each local region. It is considered that particularly in a crowded environment, persons move in a certain degree of group. For this reason, the number-of-persons estimation unit 12 estimates a movement state of persons from transition of the number of persons in a local region.
  • Concretely, the number-of-persons estimation unit 12 predicts future locations of persons, from the movement of the persons at the time when the number of persons is estimated. The number-of-persons estimation unit 12 may assume that monitoring targets move at the same probability and at the same speed in all directions, in an initial state.
  • Further, based on predicted future locations of persons, the number-of-persons estimation unit 12 estimates the future number of persons in each local region, as well. For the convenience of description, to estimate the future number of persons by the number-of-persons estimation unit 12 is referred to as “predict” in the following.
  • When predicting locations of persons, the number-of-persons estimation unit 12 may assume that each person moves at the same speed, or may use a below-described prediction result of the state prediction unit 15.
  • At a certain future time point, the number-of-persons estimation unit 12 compares the number of persons previously predicted for each local region with the number of persons estimated for each local region from a target image at this time point. Then, the number-of-persons estimation unit 12 estimates a movement state of persons, placing importance on the local region where a difference in the number of persons is smaller. Concretely, the number-of-persons estimation unit 12 may estimate that a person existing in a particular local region has moved to another local region that is one of a plurality of local regions near the particular local region and in which a difference in the number of persons estimated from the target image is the smallest.
  • The density map 30 exemplified in FIG. 2 represents a movement state of persons. FIG. 3 is an illustration depicting an example of a movement state of persons. A density map exemplified in FIG. 3 indicates that more persons exist in a lighter color region.
  • For example, it is assumed that the number-of-persons estimation unit 12 estimates that many persons exist in the region 31 a and the region 31 b in a target image at a certain time point. It is assumed that the number-of-persons estimation unit 12 continues to perform the estimation, and as a result, many persons move to the region 32 a and the region 32 b in a target image at another time point. It becomes possible to detect, from this time-series change, a movement (the arrow 33 a) of persons from the region 31 a to the region 32 a, and a movement (the arrow 33 b) of persons from the region 31 b to the region 32 b.
  • The number-of-persons estimation unit 12 may determine whether a person included in a target image at a certain time point is a person who newly appears at this time point, or an already existing person who has been present in the target image before this time point (i.e., a person who has moved within the shot region). For example, when estimating existence of a person at a location to which a person has not been predicted to move from the target image, the number-of-persons estimation unit 12 may determine that this person is a newly appearing person.
  • The flow calculation unit 13 calculates an optical flow for target images. A method for calculating an optical flow is not limited particularly. For example, the flow calculation unit 13 may calculate an optical flow by using a feature point as in the Lucas-Kanade method, or may calculate an optical flow by using the calculus of variations as in the Horn-Schunck method.
  • The state storage device 14 stores past states of persons. Here, a state of a person includes a location, a velocity, and likelihood (a weight value) of a person at a past time point. Further, a state of a person may include a behavior state (a moving state or a stationary state) of a person.
  • For example, in the case of using a probability hypothesis density (PHD) particle filter, a state of one monitoring target is represented by a plurality of particles. In this case, when the number of particles is N, the state storage device 14 holds, as a state of the i-th (i=1, . . . , N) particle, a location xi, a velocity vi, and a weight value wi which are information held in a typical particle filter. In addition to these pieces of information, the state storage device 14 may hold, as a state of each particle, a variable si expressing whether a person is in a moving state or in a stationary state.
  • The total sum of weight values wi of the particles corresponds to a quantity of monitoring targets, i.e., the number of persons. For example, in a PHD particle filter, when a hundred of particles are newly scattered for one monitoring target, a weight value of each particle set at this time is “0.01”.
  • The state prediction unit 15 predicts a state of a person at the time that the image input unit 11 acquires an image, by using a past state of the person stored in the state storage device 14. Other than a location, a velocity, and a weight value of the person, the state prediction unit 15 may predict whether a behavior state of the person is a moving state or a stationary state.
  • Concretely, the state prediction unit 15 predicts a future state of a person by using a plurality of particles expressing a state of this person. In the present example embodiment, the description is made on a method in which the state prediction unit 15 predicts a state of an object by using a PHD particle filter. The state prediction unit 15 predicts a state of a person at the time that the image input unit 11 acquires a target image.
  • As in the above-described example, in the case of using a PHD particle filter, a state of one monitoring target is expressed by a plurality of particles. When the number of particles is N, a state of the i-th (i=1, . . . , N) particle is expressed by using a location xi, a velocity vi, and a weight value wi which are information held in a typical PHD particle filter. A state of the particle may be expressed further by a variable si expressing whether a person is in a moving state or in a stationary state.
  • A method by which the state prediction unit 15 predicts a location and a velocity is similar to a prediction method performed in a general particle filter. For example, when a location of a particle at a certain time point is x, a location of this particle after a lapse of dt is expressed by x+v×dt+e. Here, e expresses a noise that cannot be expressed in a constant velocity motion model, and for example, e is a random number value generated based on normal distribution of predetermined standard deviation. When a velocity of a particle at the location x is v, a velocity of this particle after a lapse of dt is expressed by v+f. Here, f expresses a noise that cannot be expressed in a constant velocity motion model, and for example, f is a random number value generated based on normal distribution of predetermined standard deviation.
  • The following describes a method in which the state prediction unit 15 predicts a behavior state. Here, it is assumed that a behavior state of a person changes from a stationary state to a moving state in accordance with predetermined probability P, and a behavior state changes from a moving state to a stationary state based on predetermined probability Q.
  • When si expresses a stationary state, the state prediction unit 15 generates a uniform random number from 0 to 1, and when this random number value is equal to or smaller than P, the state prediction unit 15 changes si to a value indicating a moving state. Meanwhile, when si expresses a moving state, the state prediction unit 15 generates a uniform random number from 0 to 1, and when this random number value is equal to or smaller than Q, the state prediction unit 15 changes si to a value indicating a stationary state.
  • In addition, the state prediction unit 15 may predict a state of a variable si based on a past history and a statistical result. For example, it is assumed that a person of a tracking target has changed from being in a stationary state to being in a moving state, and si has become a value indicating the moving state. In the case where this change has been made recently, and a predetermined period is yet to lapse from the change to the moving state, it can be supposed that this tracking target does not return to being in a stationary state soon. For this reason, in this case, the state prediction unit 15 may hold the behavior state in the moving state for a fixed period.
  • Similarly, it is assumed that a person of a tracking target has changed from being in a moving state to being in a stationary state, and si has become a value indicating the stationary state. In the case where this change has been made recently, and a predetermined period is yet to lapse from the change to the stationary state, it can be supposed that this tracking target does not return to being in a moving state soon. For this reason, in this case, the state prediction unit 15 may hold the behavior state in the stationary state for a fixed period.
  • In addition, the state prediction unit 15 may change a condition for changing a behavior state, depending on a location in an image. For example, in a region in an image where a pathway exists, the state prediction unit 15 may set, at a small value, change probability Q of change from a moving state to a stationary state. In a region in an image including an area often used for a rendezvous, the state prediction unit 15 may set, at a large value, the change probability Q.
  • The staying information calculation unit 16 extracts a region determined as a staying object, by using a target image acquired by the image input unit 11. For example, the staying information calculation unit 16 may determine, for each pixel, whether or not an object is a temporarily stationary one, by using a method for detecting an object stationary for a long period, and may detect a region that is determined as the temporarily stationary object, by a labeling process.
  • The individual person detection unit 17 detects an individual person from a target image. For example, a person existing on a front side in an image can be often detected by a general detection method because of being less obstructed by other objects. For this reason, the individual person detection unit 17 sets a region (referred to as an individual detection region in the following) where a person can be individually detected, and then detects a person from the individual detection region in a target image.
  • The state update unit 18 updates a state of a person predicted by the state prediction unit 15, based on an estimation result of the number-of-persons estimation unit 12. The state update unit 18 may update a state of a person based on process results of the flow calculation unit 13, the staying information calculation unit 16, and the individual person detection unit 17. The following concretely describes a method by which the state update unit 18 updates a monitoring target.
  • Depending on the number of persons in each local region estimated by the number-of-persons estimation unit 12, the state update unit 18 updates a state of a person in the corresponding region. In the present example embodiment, depending on the number of persons in each local region estimated by the number-of-persons estimation unit 12, the state update unit 18 updates a weight value wi of a particle predicted to exist in the corresponding region.
  • Concretely, the state update unit 18 updates a weight value wi of each particle such that the total sum of weights in the local region becomes equal to the number of persons. At this time, the state update unit 18 may update a weight value wi of each particle by using a, a predetermined value between 0 and 1, such that the total sum of updated weights becomes (1−α)×(the last total sum of weights)+{α×(the number of persons)} so as to cause a weight value wi to change gradually. In this manner, the state update unit 18 may update weight values of particles such that the total sum of weight values approaches the number of persons in the local region.
  • FIG. 4 is an illustration depicting a relation between a local region whose number of persons have been estimated and particles existing in this local region. In FIG. 4, white and black circles represent particles. The black particles represent particles whose moving directions (the arrows in the drawing) are the same as the specific direction or within a predetermined angle from the specific direction. In an example illustrated in FIG. 4, persons existing in a local region 41 a of a target image and having moving directions that are the same as the specific direction or within the predetermined angle from the specific direction are four persons. Persons existing in a local region 41 b and having moving directions that are the same as the specific direction or within the predetermined angle from the specific direction are three persons. Persons existing in a local region 41 c and having moving directions that are the same as the specific direction or within the predetermined angle from the specific direction are two persons. For example, for three particles included in the local region 41 c, the state update unit 18 updates a weight value of each particle such that the total sum of the weight values becomes two.
  • When the number-of-persons estimation unit 12 estimates the number of persons for local regions overlapping each other, the state update unit 18 may update weight values of particles individually for each local region, or may update weight values of particles all together for one region into which the local regions overlapping each other are combined. When combining the regions into one to update weight values, the state update unit 18 may set the number of persons in the overlapped region at the number of persons taking into account the number of persons for each local region (e.g., the average number of persons of the overlapped local regions).
  • FIG. 5 is an illustration depicting an example of a process of updating weight values when local regions overlap each other. The example illustrated in FIG. 5 represents the case of estimating the number of persons for three local regions partially overlapping each other. In this case, the state update unit 18 may update weight values of particles all together for one region into which the local region 42 a, the local region 42 b, and the local region 42 c are combined. Alternatively, the state update unit 18 may update weights of particles included in each local region individually for the local region 42 a, the local region 42 b, and the local region 42 c, respectively.
  • The state update unit 18 may update weight values of particles included in a local region, at the same ratio for the all or at ratios different from particle to particle. For example, depending on detection probability of persons, the state update unit 18 may change weight values of particles to be updated. Concretely, when the detection probability of each particle is Pi, and the estimated number of persons is H, the state update unit 18 may calculate a weight value wi of each particle to be updated, as [(1−Pi)*wi+H*(Pi*wi)/{the total sum of (Pi*wi)}]
  • By adopting such a weight value wi, the state update unit 18 can set a weight value that is more influenced by the estimated number H of persons as detection probability becomes higher. In a case with a location where no persons are detected, since detection probability becomes zero, a weight value does not change. Changing a weight value to be updated, depending on detection probability, enables the update process to be performed prioritizing information of a region having higher detection probability, for example for an image shot from a low depression angle.
  • FIG. 6 is an illustration depicting an example of a condition in which detection probability varies. As exemplified in FIG. 6, a far end side in a target image is often shot at a low depression angle. For example, it is assumed that detection probability of a circumference 44 of a local region 43 decreases from a near side toward a back side. For this reason, the state update unit 18 updates a weight of a particle existing on the front side to be higher, and updates a weight of a particle existing on the back side to be lower.
  • When no particles exist at a location corresponding to a local region for which the number-of-persons estimation unit 12 has estimated the number of persons, the state update unit 18 newly generates a particle at this location. Concretely, the state update unit 18 may newly generate a particle randomly near this location in accordance with normal distribution of predetermined standard deviation.
  • The state update unit 18 may perform update so as to increase a weight value wi of a particle whose movement is close to an optical flow of target images. At this time, the state update unit 18 update a weight value wi so as not to change the total sum of the weights.
  • For example, assuming that an angle between a vector of the optical flow and a vector expressing movement of a particle i is θ, a closeness degree Li in movement between these two vectors is calculated by Li=(cos θ+1)/2. Li is a value that becomes larger for a particle whose movement is closer to the optical flow. When the total sum of weight values is S, the state update unit 18 may update a weight value of each particle by [Li*wi/{the total sum of (Li*wi)}]*S. By performing such update, a particle whose movement is closer to the optical flow has a larger weight value wi.
  • A method for updating a weight value of a particle whose movement is closer to the optical flow is not limited to the above-described method. Alternatively, when the angle θ between these vectors is equal to or smaller than a threshold value, the state update unit 18 may simply multiply a weight value of the particle by a positive constant, and then normalize each weight value such that the total sum of the weight values multiplied by the constant becomes equal to the total sum S of the original weight values. At the time of determining a closeness degree in movement, the state update unit 18 may determine, as a particle of close movement, a particle for which a distance between the vectors is equal to or smaller than a threshold value.
  • The state update unit 18 may update a weight value of a particle depending on a closeness degree between a region determined as a staying object by the staying information calculation unit 16 and a behavior state of the particle. Concretely, the state update unit 18 may perform update so as to more increase a weight value of a particle as a distance between a region determined as a staying object by the staying information calculation unit 16 and the particle whose behavior state has been predicted as a stationary state is shorter. At this time, the state update unit 18 may update a weight value of the particle, for example by using a method similar to the method of updating a weight value of a particle whose movement is close to the optical flow.
  • When the individual person detection unit 17 detects a person in an individual detection region, the state update unit 18 may update weight values of particles existing in a region associated with the detected person, in a general method performed in a PHD particle filter. In other words, for an area where the individual person detection unit 17 has detected persons, the state update unit 18 may update weight values of particles included in each local region, by a method other than the method using the estimated number of persons in each local region. Selecting the more suitable prediction method depending on a region in a shot image in this manner makes it possible to improve accuracy in tracking a monitoring target.
  • The state update unit 18 deletes a particle whose weight value wi is equal to or smaller than a threshold value. The state update unit 18 updates location information of a person in a manner similar to a typical tracking process. The state update unit 18 records an updated state of a person in the state storage device 14. The state update unit 18 may perform re-sampling of scattering particles again depending on weight values of the respective particles.
  • The number-of-persons output unit 19 outputs the number of persons in a shot area, based on states of persons. Concretely, the number-of-persons output unit 19 outputs the number of persons included in a target image, by using weight values wi of particles updated by the state update unit 18.
  • As described above, the total sum of weight values wi of particles corresponds to the number of persons. For this reason, the number-of-persons output unit 19 may calculate the total sum of weight values wi of particles and output the number of persons in a shot area. When outputting the number of persons in a predetermined region, the number-of-persons output unit 19 may identify particles existing in this region, and may calculate the total sum of weight values wi of the identified particles.
  • When a state of each particle includes a behavior state, the number-of-persons output unit 19 may calculate the total sum of weight values wi of the particles in a stationary state, and may output the number of persons staying in a shot area, for example.
  • The number-of-persons output unit 19 may output the number of persons that have moved in a specific direction, and the number of persons that have crossed over a line set at a predetermined location in a shot area (i.e., the number of persons that have crossed over the predetermined line), by using states of current particles and states of past particles. Concretely, the number-of-persons output unit 19 outputs the number of persons that have crossed over the predetermined location, based on weighs of particles that have crossed over this location. The following describes a method of calculating the number of persons by using temporally changing particles.
  • When calculating the number of persons that have moved in a specific direction, for example, the number-of-persons output unit 19 may identify particles whose past and current locations are connected by a line whose direction is the same as the specific direction or makes a predetermined angle or less with the specific direction, and may calculate the total sum of weight values of these particles. At the time of calculating the total sum of weight values of the particles, the number-of-persons output unit 19 may use weight values of past particles, may use weight values of current particles, or may use an average of the weight values of both of the particles.
  • When calculating the number of persons that have crossed over the predetermined line, the number-of-persons output unit 19 identifies particles that have crossed over the predetermined line in the course of moving from the last location to the present location, and calculates the total sum of weight values of these particles. For example, when among particles that cross over the predetermined line, only the particles that have moved in a specific direction are targeted, the number-of-persons output unit 19 may target particles whose moving directions are each expressed by a vector that makes an inner product equal to or larger than zero, with a normal vector of the predetermined line. Concretely, assuming that the predetermined line mentioned here is a horizontal straight line, particles crossing over this straight line include particles moving from locations above this straight line to locations below this straight line, and particles moving from locations below this straight line to locations above this straight line. When among these particles, particles moving from above to below are targeted, the number-of-persons output unit 19 calculates an inner product of a normal vector (a vector directed just below in this case) to the horizontal straight line and a vector expressing a moving direction of each particle, and targets particles for which the inner product is equal to or larger than zero.
  • The number-of-persons output unit 19 can output the number of persons that have crossed over the predetermined line in a predetermined period, by integrating the total sums of weight values in the predetermined period. The number-of-persons output unit 19 can use current weight values, past weight values, or both thereof, as weight values of particles used in calculating the total sums of weight values, in the same manner as in the case of calculating the number of persons that have moved in the specific direction.
  • Particles newly generated at the current time point do not have past states, and for this reason, are not included in targets for calculating these numbers of persons.
  • FIG. 7 is an illustration depicting an example of a process of calculating a quantity of monitoring targets that have crossed over a predetermined location. The circles of the broken lines exemplified in FIG. 7 represent past particles, the circles of the solid lines represent current particles. The arrows exemplified in FIG. 7 represent a movement state of particles from the past time to the current time.
  • The case of calculating the number of persons that have crossed over the location indicated by the solid line 52 from above to below is assumed as one example. At this time, the number-of-persons output unit 19 identifies particles that have crossed over the solid line 52 from above to below, and calculates the total sum of weights of the identified particles. In the example illustrated in FIG. 7, the particle 50 b, the particle 50 c, and the particle 50 d cross over the solid line 52 from above to below. For this reason, the number-of-persons output unit 19 calculates the total sum of particle weights of the particle 50 b, the particle 50 c, and the particle 50 d, and outputs the total sum as the passing number of persons. For example, when the passing number of persons in a fixed period is to be output, the number-of-persons output unit 19 may integrate the total sums of weights of particles in the fixed period.
  • The image input unit 11, the number-of-persons estimation unit 12, the flow calculation unit 13, the state prediction unit 15, the staying information calculation unit 16, the individual person detection unit 17, the state update unit 18, and the number-of-persons output unit 19 can be implemented by programs. These respective units can be implemented by a processor of a computer operating in accordance with the programs.
  • FIG. 10 is a block diagram exemplifying a hardware configuration of a computer device 200 implementing a movement state estimation device 100. The computer device 200 includes a central processing unit (CPU) 201, a read-only memory (ROM) 202, a random access memory (RAM) 203, a storage device 204, a drive device 205, a communication interface 206, and input-output interface 207. The movement state estimation device 100 can be implemented by the configuration (or a part thereof) illustrated in FIG. 10.
  • The CPU 201 executes a program 208 by using the RAM 203. The program 208 may be stored in the ROM 202. The program 208 may be recorded in a recording medium 209 such as a flash memory, and may be read out by the drive device 205, or may be transmitted from an external device via a network 210. The communication interface 206 exchanges data with the external device via the network 210. The input-output interface 207 exchanges data with peripheral devices (an input device, a display device, and the like). The communication interface 206 and the input-output interface 207 can function as means for acquiring or outputting data.
  • The movement state estimation device 100 may be configured by a single piece of circuitry (a processor or the like), or may be configured by a combination of a plurality of pieces of circuitry. The circuitry mentioned here may be dedicated or multipurpose circuitry.
  • The CPU 201 may function as the image input unit 11, the number-of-persons estimation unit 12, the flow calculation unit 13, the state prediction unit 15, the staying information calculation unit 16, the individual person detection unit 17, the state update unit 18, and the number-of-persons output unit 19, in accordance with the program 208.
  • Each of the image input unit 11, the number-of-persons estimation unit 12, the flow calculation unit 13, the state prediction unit 15, the staying information calculation unit 16, the individual person detection unit 17, the state update unit 18, and the number-of-persons output unit 19 may be implemented by dedicated hardware. The state storage device 14 may be implemented by the storage device 204, or may be an external device connected via the communication interface 206.
  • Next, operation of the movement state estimation device 100 according to the present example embodiment is described. FIG. 8 is a flowchart illustrating an operation example of the movement state estimation device 100 according to the present example embodiment. The number-of-persons estimation unit 12 estimates the number of persons for each local region in each image using a plurality of temporally sequential images (step S11). Then, the number-of-persons estimation unit 12 estimates a movement state of persons, from time-series change of the quantity estimated for each local region (step S12).
  • Concretely, the number-of-persons estimation unit 12 predicts future locations of persons, from the movement state of persons at the time point that the estimation is made, and predicts the future number of persons in each local region. The movement state of persons can be determined from states of particles expressing persons, for example. The number-of-persons estimation unit 12 can predict the future number of persons from future locations of particles predicted by the state prediction unit 15, for example. Further, for each local region, the number-of-persons estimation unit 12 places more importance on the local region where a difference between the predicted number of persons and the estimated number of persons is small.
  • As described above, in the present example embodiment, the number-of-persons estimation unit 12 estimates the number of persons for each local region in each image by using a plurality of the temporally sequential images, and estimates a movement state of persons from the time-series change in the number of persons estimated for each local region. For this reason, according to the present example embodiment, even in a crowded environment where it is difficult to track individual persons, a movement state of persons can be accurately estimated.
  • In the present example embodiment, the state prediction unit 15 predicts future states of persons by using a plurality of particles expressing states of persons, and the number-of-persons output unit 19 calculates the number of persons by targeting particles having weights. In other words, according to the present example embodiment, states of monitoring targets are tracked by using particles having weight values so that the number of persons that have crossed over a specific place can be counted. According to the present example embodiment, not only the number of persons that have simply crossed over a specific place, but also the number of persons that move in a specific direction can be counted. Thereby, it becomes possible to measure not only a simple crowdedness degree in a certain location, but also a flow quantity (a person flow) at this location.
  • [Outline]
  • Next, an outline of the example embodiment according to the present invention is described. FIG. 9 is a block diagram illustrating the outline of a movement state estimation device in the present example embodiment. The movement state estimation device illustrated in FIG. 9 includes a quantity estimation unit 81 estimating a quantity of monitoring targets for each local region in each image by using a plurality of temporally sequential images, and a movement state estimation unit 82 estimating a movement state of the monitoring targets from time-series change in the estimated quantity for each local region. The quantity estimation unit 81 and the movement state estimation unit 82 correspond to the number-of-persons estimation unit 12 in the above-described example embodiment.
  • The thus-configured movement state estimation device can accurately estimate a movement state of monitoring targets even in a crowded environment.
  • The movement state estimation unit 82 may predict future locations of monitoring targets, from a movement state of the monitoring targets at the time point that the estimation is made, and may predict a future quantity of the monitoring targets in each local region. The movement state estimation unit 82 may compare the future quantity of the monitoring targets predicted for each local region, with the quantity of the monitoring targets estimated for each local region from a target image at the future time point, and may estimate a movement state of the monitoring targets, placing importance on the local region where the difference in the quantity is smaller.
  • The movement state estimation unit 82 may determine whether a shot monitoring target in a target image is a new monitoring target or a monitoring target that has moved within the target image, to estimate a movement state of the monitoring targets.
  • The movement state estimation device may include a prediction unit (e.g., the state prediction unit 15) predicting future states of monitoring targets by using a plurality of particles expressing states of the monitoring targets. The movement state estimation device may include an update unit (e.g., the state update unit 18) updating, depending on an estimated quantity of monitoring targets in each local region, weight values of the particles predicted to exist in each corresponding local region.
  • The movement state estimation unit 82 may estimate a movement state of monitoring targets from time-series change in the total sum of weight values of particles included in each local region. According to such a configuration, it becomes possible to deal with various states of movement performed by monitoring targets.
  • Concretely, the update unit may update weight values to be set for particles, such that the total sum of weight values to be set for particles predicted to be included in a local region approaches an estimated quantity of monitoring targets in the corresponding local region.
  • For particles expressing a state of one monitoring target, weight values are set so as to amount to one in total. For particles expressing a state of a monitoring target, at least locations and velocities for the monitoring target are set. Based on the locations and the velocities set for the particles, the prediction unit predicts a future location of the monitoring target.
  • The movement state estimation device may include a quantity output unit (e.g., the number-of-persons output unit 19) outputting a quantity of monitoring targets depending on weights set for particles.
  • Concretely, the quantity output unit may output a quantity of monitoring targets that have crossed over a predetermined location, based on weights of particles that have crossed over this location.
  • The movement state estimation device may include a flow calculation unit (e.g., the flow calculation unit 13) calculating an optical flow of target images. The update unit may perform update so as to increase weight values of particles whose movement is close to the optical flow. According to such a configuration, it becomes possible to estimate a movement state, placing importance on particles whose movement is close to movement estimated from images.
  • The movement state estimation device may include a staying information calculation unit (e.g., the staying information calculation unit 16) extracting a region determined as a staying object from target images. The prediction unit may predict future behavior states of monitoring targets, and as the particles whose behavior states have been predicted to be a stationary state are closer to a region determined as a staying object, the update unit may update weight values of these particles so as to more increase. According to such a configuration, states of monitoring targets that are not moving can be determined appropriately.
  • The movement state estimation device may include a monitoring target detection unit detecting a monitoring target from a target image (concretely, an individual detection region set, as an area where a monitoring target can be individually detected, in the target image). For the area from which the monitoring target detection means has detected a monitoring target, the update unit may update weight values set for particles predicted to be included in each local region thereof, by a method other than a method using an estimated quantity of monitoring targets in each local region. Examples of such a method include a method in which a general PHD particle filter updates weight values.
  • In this manner, the movement state estimation method according to the present invention is combined with a method actually detecting a monitoring target so that accuracy in estimating a movement state of monitoring targets can be improved.
  • In the above, the present invention is described by citing as a typical example the above-described example embodiment. However, the present invention is not limited to the above-described example embodiment. In other words, various configurations that can be understood by those skilled in the art can be adopted in the present invention within the scope of the present invention.
  • This patent application claims priority based on Japanese Patent Application No. 2015-004963 filed on Jan. 14, 2015, entire disclosure of which is incorporated herein.
  • INDUSTRIAL APPLICABILITY
  • The present invention is applicable to a movement state estimation device estimating a quantity of moving objects. For example, the present invention is applicable to a device estimating a flow of objects such as persons or vehicles, or the number of objects crossing over a specific spot, from a video taken with a camera in a monitoring system or the like using a fixed camera or the like.
  • REFERENCE SIGNS LIST
    • 11 Image input unit
    • 12 Number-of-persons estimation unit
    • 13 Flow calculation unit
    • 14 State storage device
    • 15 State prediction unit
    • 16 Staying information calculation unit
    • 17 Individual person detection unit
    • 18 State update unit
    • 19 Number-of-persons output unit
    • 20 Target image
    • 21, 41 a to 41 c, 42 a to 42 c, 43 Local region
    • 50 a to 50 f, 51 a to 51 f Particle

Claims (12)

1. A movement state estimation device comprising:
at least one memory storing instructions; and
at least one processor executing the instructions to perform:
estimating a number of people in a plurality of regions of an image by using machine learning; and
estimating a direction of flow for the image based on the estimated number of the people in the plurality of regions.
2. The movement state estimation device according to claim 1, wherein the at least one processor further executes the instructions to perform:
estimating the direction of flow for the image based on a movement of the people in the plurality of regions.
3. The movement state estimation device according to claim 2,
wherein the movement of the people is based on a motion vector of pixels among a plurality of images including the image.
4. The movement state estimation device according to claim 1,
wherein using the machine learning to estimate the number of the people comprises using a classifier which is trained based on crowd patches, and
wherein the crowd-patches are local images including a crowd state of an overlap of the people.
5. A movement state estimation method comprising:
estimating a number of people in a plurality of regions of an image by using machine learning; and
estimating a direction of flow for the image based on the estimated number of the people in the plurality of regions.
6. The movement state estimation method according to claim 5, further comprising:
estimating the direction of flow for the image is based on a movement of the people in the plurality of regions.
7. The movement state estimation method according to claim 6,
wherein the movement of the people is based on a motion vector of pixels among a plurality of images including the image.
8. The movement state estimation method according to claim 5,
wherein using the machine learning to estimate the number of the people comprises using a classifier which is trained based on crowd patches, and
wherein the crowd-patches are local images including a crowd state of an overlap of the people.
9. A non-transitory program recording medium recording a program for causing a computer to execute:
estimating a number of people in a plurality of regions of an image by using machine learning; and
estimating a direction of flow for the image based on the estimated number of the people in the plurality of regions.
10. The non-transitory program recording medium according to claim 9, wherein the program causes the computer to further execute:
estimating the direction of flow for the image is based on a movement of the people in the plurality of regions.
11. The non-transitory program recording medium according to claim 10,
wherein the movement of the people is based on a motion vector of pixels among a plurality of images including the image.
12. The non-transitory program recording medium according to claim 9,
wherein the program causes the computer to further execute using the machine learning to estimate the number of the people by using a classifier which is trained based on crowd patches, and
wherein the crowd-patches are local images including a crowd state of an overlap of the people.
US16/921,447 2015-01-14 2020-07-06 Movement state estimation device, movement state estimation method and program recording medium Abandoned US20200334472A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/921,447 US20200334472A1 (en) 2015-01-14 2020-07-06 Movement state estimation device, movement state estimation method and program recording medium

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
JP2015004963 2015-01-14
JP2015-004963 2015-01-14
PCT/JP2016/000146 WO2016114134A1 (en) 2015-01-14 2016-01-13 Motion condition estimation device, motion condition estimation method and program recording medium
US201715543408A 2017-07-13 2017-07-13
US16/296,516 US10755108B2 (en) 2015-01-14 2019-03-08 Movement state estimation device, movement state estimation method and program recording medium
US16/921,447 US20200334472A1 (en) 2015-01-14 2020-07-06 Movement state estimation device, movement state estimation method and program recording medium

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US16/296,516 Continuation US10755108B2 (en) 2015-01-14 2019-03-08 Movement state estimation device, movement state estimation method and program recording medium

Publications (1)

Publication Number Publication Date
US20200334472A1 true US20200334472A1 (en) 2020-10-22

Family

ID=56405682

Family Applications (5)

Application Number Title Priority Date Filing Date
US15/543,408 Active US10325160B2 (en) 2015-01-14 2016-01-13 Movement state estimation device, movement state estimation method and program recording medium
US16/296,468 Active US10657386B2 (en) 2015-01-14 2019-03-08 Movement state estimation device, movement state estimation method and program recording medium
US16/296,516 Active US10755108B2 (en) 2015-01-14 2019-03-08 Movement state estimation device, movement state estimation method and program recording medium
US16/921,447 Abandoned US20200334472A1 (en) 2015-01-14 2020-07-06 Movement state estimation device, movement state estimation method and program recording medium
US17/849,211 Pending US20220327839A1 (en) 2015-01-14 2022-06-24 Movement state estimation device, movement state estimation method and program recording medium

Family Applications Before (3)

Application Number Title Priority Date Filing Date
US15/543,408 Active US10325160B2 (en) 2015-01-14 2016-01-13 Movement state estimation device, movement state estimation method and program recording medium
US16/296,468 Active US10657386B2 (en) 2015-01-14 2019-03-08 Movement state estimation device, movement state estimation method and program recording medium
US16/296,516 Active US10755108B2 (en) 2015-01-14 2019-03-08 Movement state estimation device, movement state estimation method and program recording medium

Family Applications After (1)

Application Number Title Priority Date Filing Date
US17/849,211 Pending US20220327839A1 (en) 2015-01-14 2022-06-24 Movement state estimation device, movement state estimation method and program recording medium

Country Status (3)

Country Link
US (5) US10325160B2 (en)
JP (4) JP6969871B2 (en)
WO (1) WO2016114134A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11373425B2 (en) * 2020-06-02 2022-06-28 The Nielsen Company (U.S.), Llc Methods and apparatus for monitoring an audience of media based on thermal imaging
US11553247B2 (en) 2020-08-20 2023-01-10 The Nielsen Company (Us), Llc Methods and apparatus to determine an audience composition based on thermal imaging and facial recognition
US11595723B2 (en) 2020-08-20 2023-02-28 The Nielsen Company (Us), Llc Methods and apparatus to determine an audience composition based on voice recognition
US11763591B2 (en) 2020-08-20 2023-09-19 The Nielsen Company (Us), Llc Methods and apparatus to determine an audience composition based on voice recognition, thermal imaging, and facial recognition

Families Citing this family (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6988790B2 (en) 2016-03-07 2022-01-05 日本電気株式会社 Crowd type identification system, crowd type identification method and crowd type identification program
CN109479117A (en) * 2016-07-14 2019-03-15 三菱电机株式会社 Cluster monitoring arrangement and cluster monitoring system
JP6950692B2 (en) * 2016-08-04 2021-10-13 日本電気株式会社 People flow estimation device, people flow estimation method and program
US10810442B2 (en) 2016-09-13 2020-10-20 Nec Corporation People flow estimation device, people flow estimation method, and recording medium
JP6820204B2 (en) * 2017-01-18 2021-01-27 日本放送協会 State estimator and program
JP6914699B2 (en) * 2017-04-04 2021-08-04 キヤノン株式会社 Information processing equipment, information processing methods and programs
US10839552B2 (en) * 2017-06-01 2020-11-17 Nec Corporation Image processing apparatus, tracking method, and program
JP6841774B2 (en) * 2018-01-19 2021-03-10 日本電信電話株式会社 Predictors, prediction methods and computer programs
SG10201802673VA (en) * 2018-03-29 2019-10-30 Nec Asia Pacific Pte Ltd Method and system for integration and automatic switching of crowd estimation techniques
WO2019229979A1 (en) * 2018-06-01 2019-12-05 日本電気株式会社 Information processing device, control method, and program
JP7216487B2 (en) * 2018-06-21 2023-02-01 キヤノン株式会社 Image processing device and its control method
JP2019220087A (en) * 2018-06-22 2019-12-26 日本電信電話株式会社 Estimation method, estimation device, and estimation program
JP7443002B2 (en) * 2019-09-13 2024-03-05 キヤノン株式会社 Image analysis device, image analysis method, and program
JP7327645B2 (en) * 2020-03-12 2023-08-16 日本電気株式会社 Image processing device, image processing system, image processing method, and image processing program
US20230112675A1 (en) * 2020-03-27 2023-04-13 Nec Corporation Person flow prediction system, person flow prediction method, and programrecording medium

Family Cites Families (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0460880A (en) * 1990-06-29 1992-02-26 Shimizu Corp Moving body discrimination and analysis controlling system
JP2855157B2 (en) * 1990-07-17 1999-02-10 清水建設株式会社 Crowd walking simulation system
JP2001076291A (en) 1999-09-02 2001-03-23 Nri & Ncc Co Ltd Traffic measurement system
US6633232B2 (en) * 2001-05-14 2003-10-14 Koninklijke Philips Electronics N.V. Method and apparatus for routing persons through one or more destinations based on a least-cost criterion
US7123918B1 (en) 2001-08-20 2006-10-17 Verizon Services Corp. Methods and apparatus for extrapolating person and device counts
JP2006031645A (en) 2004-07-12 2006-02-02 Nariyuki Mitachi Real-time estimation method for dynamic crowd density and crowd accident prevention system
JP2006270865A (en) 2005-03-25 2006-10-05 Victor Co Of Japan Ltd Image-monitoring apparatus
JP2007243342A (en) * 2006-03-06 2007-09-20 Yokogawa Electric Corp Image-monitoring apparatus and image-monitoring system
JP4624396B2 (en) 2007-10-26 2011-02-02 パナソニック株式会社 Situation judging device, situation judging method, situation judging program, abnormality judging device, abnormality judging method and abnormality judging program
US20090158309A1 (en) 2007-12-12 2009-06-18 Hankyu Moon Method and system for media audience measurement and spatial extrapolation based on site, display, crowd, and viewership characterization
CA2719504A1 (en) * 2008-04-14 2009-10-22 Thomson Licensing Technique for automatically tracking an object
JP2009294887A (en) * 2008-06-05 2009-12-17 Vector Research Institute Inc Construction facility control system and program
JP2010198566A (en) 2009-02-27 2010-09-09 Nec Corp Device, method and program for measuring number of people
WO2012111138A1 (en) 2011-02-18 2012-08-23 株式会社日立製作所 Pedestrian movement information detection device
JP5776312B2 (en) * 2011-04-28 2015-09-09 大日本印刷株式会社 Image analysis apparatus, image analysis method, image analysis program, and recording medium
JP5680524B2 (en) * 2011-12-09 2015-03-04 株式会社日立国際電気 Image processing device
US9165190B2 (en) 2012-09-12 2015-10-20 Avigilon Fortress Corporation 3D human pose and shape modeling
JP2014106879A (en) 2012-11-29 2014-06-09 Railway Technical Research Institute Human distribution status estimation system
WO2014112407A1 (en) * 2013-01-16 2014-07-24 日本電気株式会社 Information processing system, information processing method, and program
JP6236448B2 (en) 2013-06-21 2017-11-22 株式会社日立製作所 Sensor arrangement determination device and sensor arrangement determination method
US9875431B2 (en) * 2013-06-28 2018-01-23 Nec Corporation Training data generating device, method, and program, and crowd state recognition device, method, and program
JP5613815B1 (en) 2013-10-29 2014-10-29 パナソニック株式会社 Residence status analysis apparatus, residence status analysis system, and residence status analysis method
JP6331785B2 (en) 2014-07-08 2018-05-30 日本電気株式会社 Object tracking device, object tracking method, and object tracking program
JP5854098B2 (en) 2014-08-08 2016-02-09 大日本印刷株式会社 Information display device and information display program

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11373425B2 (en) * 2020-06-02 2022-06-28 The Nielsen Company (U.S.), Llc Methods and apparatus for monitoring an audience of media based on thermal imaging
US11553247B2 (en) 2020-08-20 2023-01-10 The Nielsen Company (Us), Llc Methods and apparatus to determine an audience composition based on thermal imaging and facial recognition
US11595723B2 (en) 2020-08-20 2023-02-28 The Nielsen Company (Us), Llc Methods and apparatus to determine an audience composition based on voice recognition
US11763591B2 (en) 2020-08-20 2023-09-19 The Nielsen Company (Us), Llc Methods and apparatus to determine an audience composition based on voice recognition, thermal imaging, and facial recognition
US11962851B2 (en) 2020-08-20 2024-04-16 The Nielsen Company (Us), Llc Methods and apparatus to determine an audience composition based on thermal imaging and facial recognition

Also Published As

Publication number Publication date
WO2016114134A1 (en) 2016-07-21
JP2022166067A (en) 2022-11-01
US20220327839A1 (en) 2022-10-13
US10325160B2 (en) 2019-06-18
US20190205660A1 (en) 2019-07-04
JPWO2016114134A1 (en) 2017-10-26
JP6969871B2 (en) 2021-11-24
US10755108B2 (en) 2020-08-25
JP7428213B2 (en) 2024-02-06
US20190220672A1 (en) 2019-07-18
JP7163945B2 (en) 2022-11-01
US20180005046A1 (en) 2018-01-04
US10657386B2 (en) 2020-05-19
JP2021036437A (en) 2021-03-04
JP2024041997A (en) 2024-03-27

Similar Documents

Publication Publication Date Title
US20220327839A1 (en) Movement state estimation device, movement state estimation method and program recording medium
US10943204B2 (en) Realtime video monitoring applied to reduce customer wait times
US9767570B2 (en) Systems and methods for computer vision background estimation using foreground-aware statistical models
US7940957B2 (en) Object tracker for visually tracking object motion
JP5102410B2 (en) Moving body detection apparatus and moving body detection method
US20200050873A1 (en) People flow estimation device, display control device, people flow estimation method, and recording medium
AU2014240669B2 (en) Object monitoring system, object monitoring method, and monitoring target extraction project
JP5459674B2 (en) Moving object tracking system and moving object tracking method
JP2007128513A (en) Scene analysis
KR20190128500A (en) Method and apparatus for detecting garbage dumping action in real time on video surveillance system
KR101720781B1 (en) Apparatus and method for prediction of abnormal behavior of object
US9977970B2 (en) Method and system for detecting the occurrence of an interaction event via trajectory-based analysis
CN111126153B (en) Safety monitoring method, system, server and storage medium based on deep learning
KR20190023389A (en) Multi-Class Multi-Object Tracking Method using Changing Point Detection
CN104219488A (en) Method and device of generating target image as well as video monitoring system
US10929688B2 (en) System and method of video content filtering
Melo et al. Viewpoint independent detection of vehicle trajectories and lane geometry from uncalibrated traffic surveillance cameras
Płaczek A real time vehicle detection algorithm for vision-based sensors
CN115719347A (en) Behavior recognition method and device, electronic equipment and vehicle
Campo et al. Modeling and classification of trajectories based on a gaussian process decomposition into discrete components
KR20220154473A (en) System of peventing external intrusion using virtulal detection line in image
Foresti et al. Vehicle detection and tracking for traffic monitoring
WO2021157133A1 (en) Re-identification device, re-identification program, and re-identification method
TW201937404A (en) The intelligent camera which can identify event by itself
Ariyani et al. Multi-Person Key Points Detection for Abnormal Human Behavior Analysis Using The ConvLSTM-AE Method

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: ADVISORY ACTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION