WO2010113417A1 - Moving object tracking device, moving object tracking method, and moving object tracking program - Google Patents

Moving object tracking device, moving object tracking method, and moving object tracking program Download PDF

Info

Publication number
WO2010113417A1
WO2010113417A1 PCT/JP2010/002015 JP2010002015W WO2010113417A1 WO 2010113417 A1 WO2010113417 A1 WO 2010113417A1 JP 2010002015 W JP2010002015 W JP 2010002015W WO 2010113417 A1 WO2010113417 A1 WO 2010113417A1
Authority
WO
WIPO (PCT)
Prior art keywords
moving object
unit
image
contour
time
Prior art date
Application number
PCT/JP2010/002015
Other languages
French (fr)
Japanese (ja)
Inventor
上條俊介
Original Assignee
国立大学法人東京大学
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 国立大学法人東京大学 filed Critical 国立大学法人東京大学
Publication of WO2010113417A1 publication Critical patent/WO2010113417A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/215Motion-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30236Traffic on road, railway or crossing

Definitions

  • the present invention relates to a moving object tracking device, a moving object tracking method, and a moving object tracking program for tracking a moving object (for example, a car, a bicycle, and an animal) in an image by processing a time-series image.
  • a moving object for example, a car, a bicycle, and an animal
  • This application claims priority based on Japanese Patent Application No. 2009-090489 filed in Japan on April 2, 2009, the contents of which are incorporated herein by reference.
  • the vehicles M1 and M2 are identified, and the motion vectors of these vehicles M1 and M2 are obtained.
  • the vehicles M1 and M2 can be tracked with a single camera.
  • JP 2002-133421 A Japanese Patent Laid-Open No. 2004-207786
  • an image captured with the camera fixed that is, an image with a fixed background is subjected to image processing to accurately detect a moving object in the image.
  • the background image that has been fixed up to that time also varies depending on the camera panning or zooming.
  • image processing is performed on an image whose background changes in this way, there is a problem in that the boundary between the image of the moving object and the background image is not clear, and the moving object in the image cannot be accurately detected.
  • the present invention has been made in view of the above circumstances, and an object of the present invention is to provide a moving object tracking device, a moving object tracking method, and a moving object tracking program that can accurately detect a moving object in an image whose background changes. .
  • the moving object tracking device of the present invention is a moving object tracking device that detects a moving object in a time-series image by image processing, and an identification code corresponding to the moving object is assigned to each frame of the time-series image. Is assigned to each block divided into a plurality of blocks, and is stored; an outline extraction unit that extracts an outline of the moving object from the time-series image; and the blocks assigned to the blocks based on the outlines A correction unit that corrects the identification code.
  • the contour extraction unit calculates the contour of the moving object based on an image area corresponding to the moving object stored in the object map storage unit.
  • a target region setting unit that sets a target region to be extracted; and a contour extraction processing unit that extracts a contour of the moving object from the time-series image with respect to the target region.
  • the number of pixels corresponding to an edge in the target region is determined by the contour extraction unit with respect to an image obtained by performing edge extraction processing on the time-series image.
  • a target area correction unit that corrects the target area for each coordinate axis based on the histogram, and the contour extraction processing unit applies the corrected target area to the corrected target area.
  • the contour of the moving object is extracted from the time-series image; a configuration may be adopted.
  • the contour extraction unit sets a plurality of moving objects in which occlusion has occurred as an integrated moving object, and extracts the integrated moving object from the time-series image. An outline may be extracted.
  • the correction unit is based on identification codes corresponding to a plurality of moving objects in which the occlusion occurs, which is stored in the object map storage unit. Based on the information indicating the boundaries of the plurality of moving objects and the contour of the moving object that is extracted by integrating the plurality of moving objects in which the occlusion is generated by the contour extraction unit, the object map storage unit stores The stored identification code may be corrected.
  • a determination unit that determines whether or not the background image is fluctuating; and when the determination unit determines that the background image is fluctuating, A first control unit that causes the contour extraction unit to extract a contour and causes the correction unit to correct the identification code stored in the object map storage unit.
  • the moving object tracking device detects the size of the moving object or the amount of change in the movement amount stored in the object map storage unit for each unit time based on the identification code.
  • a moving object fluctuation amount detecting unit that detects the moving object size or movement amount of the moving object per unit time detected by the moving object fluctuation amount detection unit is larger than a predetermined size or movement amount fluctuation amount
  • a second control unit that causes the contour extraction unit to extract a contour and causes the correction unit to correct the identification code stored in the object map storage unit.
  • the moving object tracking device stores the identification code and the object map storage unit based on the result of image processing of the time-series image. And a moving object tracking unit that updates the motion vector of the moving object, and the moving object tracking unit is adjacent to each of the consecutive N images (N ⁇ 2) of the time-series images.
  • An update process for updating the identification code and the motion vector may be employed.
  • the moving object tracking method of the present invention is a moving object tracking method for detecting a moving object in a time-series image by image processing, and an identification code corresponding to the moving object is assigned to each frame of the time-series image. Assigning to a block divided into a plurality of blocks, storing the blocks, extracting a contour of the moving object from the time-series image, correcting the identification code assigned to the block based on the contours And comprising.
  • the moving object tracking program of the present invention provides a computer as a moving object tracking device that detects a moving object in a time-series image by image processing, and assigns an identification code corresponding to the moving object to each of the time-series images. Allocating and storing a frame into a plurality of divided blocks; extracting a contour of the moving object from the time-series image; and correcting the identification code allocated to the block based on the contour And a process is executed.
  • the moving object in the image can be detected even when the background of the image fluctuates. It can be detected accurately.
  • FIG. 1 is a schematic diagram of a moving object tracking system using a moving object tracking device according to an embodiment of the present invention. It is a block diagram which shows the structure of the moving object tracking apparatus which concerns on the same embodiment. It is a figure which shows ID of the moving object provided to the slit and block which were each set to the four entrances to an intersection and the four exits from an intersection in a frame image. It is a figure which shows typically the image in the time t-1 with a block boundary line. It is a figure which shows the image in the time t typically with a block boundary line. It is a figure which shows typically the image in the time t-1 with a pixel boundary line. It is a figure which shows the image in the time t typically with a pixel boundary line.
  • FIG. 10 is a diagram schematically showing motion vectors and object boundaries assigned to an object map at time t ⁇ 1 when the same estimation method is applied. It is a figure which shows typically the motion vector and object boundary provided to the object map at the same time t.
  • FIG. 6 is a state transition diagram showing state transition of modes in the moving object tracking device 20.
  • 12 is a flowchart illustrating an operation of the moving object tracking device 20 in the correction mode of FIG. 11. It is a figure which shows area term Earea. It is a figure which shows the 1st step of an example of the processing result of Snakes. It is a figure which shows the 2nd step. It is a figure which shows the said 3rd step.
  • FIG. 10 is a diagram showing a detection result in the same frame 80. It is a figure which shows the detection result in the same frame 95.
  • FIG. 10 is a diagram showing a detection result in the same frame 80. It is a figure which shows the detection result in the same frame 95.
  • FIG. 95 It is a figure which shows the detection result in the same frame 95.
  • FIG. It is a figure which shows the detection result in the same frame 107.
  • FIG. It is a figure which shows the detection result in the same frame 107.
  • FIG. It is a figure which shows the detection result in the same frame 115.
  • FIG. It is a figure which shows the detection result in the same frame 115.
  • FIG. It is a figure which shows the detection result in the same frame 142.
  • FIG. It is a figure which shows the detection result in the same frame 153.
  • FIG. It is a figure which shows the detection result in the same frame 153.
  • FIG. 10 is a diagram showing a detection result in the same frame 80. It is a figure which shows the detection result in the same frame 95. FIG. It is a figure which shows the detection result in the same frame 95. FIG. It is a figure which shows the detection result in the same frame 107. FIG. It is a figure which shows the detection result in the same frame 107. FIG. It is a figure which shows the detection result in the same frame 115. FIG. It is a figure which shows the detection result in the same frame 115. FIG. It is a figure which shows the detection result in the same frame 142.
  • FIG. 10 is a diagram showing a detection result in the same frame 80. It is a figure which shows the detection result in the same frame 95. FIG. It is a figure which shows the detection result in the same frame 95. FIG. It is a figure which shows the detection result in the same frame 107. FIG. It is a figure which shows the detection result in the same frame 107. FIG. It is a figure which shows the detection result in the same
  • FIG. It is a figure which shows the detection result in the same frame 142.
  • FIG. It is a figure which shows the detection result in the same frame 153.
  • FIG. It is a figure which shows the detection result in the same frame 153.
  • FIG. It is a figure which shows the detection result of the moving object in case the occlusion has generate
  • It is a figure which shows typically the image in the time t 1 imaged with the camera installed above the center line of the highway.
  • It is a figure which shows typically the image in the same time t 2.
  • It is a figure which shows typically the image in the same time t 3.
  • It is a figure which shows typically the image in the same time t 4.
  • FIG. 1 is a schematic diagram showing a configuration of a moving object tracking system using a moving object tracking device 20 according to an embodiment of the present invention.
  • the moving object tracking system includes an electronic camera 10 that captures an intersection and outputs an image signal, and a moving object tracking device 20 that processes the image and tracks the moving object. .
  • the time-series images captured by the electronic camera 10 are stored in an image memory 21 (described later) included in the moving object tracking device 20 at a rate of 12 frames / second.
  • image memory 21 included in the moving object tracking device 20
  • the oldest frame is rewritten with a new frame image.
  • the electronic camera 10 can change an image area to be photographed by panning or zooming.
  • the panning or zooming with respect to the electronic camera 10 may be controlled by the moving object tracking device 20 or may be controlled by a host control device that controls the moving object tracking system.
  • the moving object tracking device 20 performs image processing on a time series image (a time series image stored in an image memory 21 described later) taken by the electronic camera 10 to detect a moving object in the image.
  • the image conversion unit 22 copies each frame image in the image memory 21 to the frame buffer memory 23, and converts the corresponding frame image in the image memory 21 into a spatial difference frame image using the copied image data. . This conversion is performed in two stages.
  • the pixel value (luminance value) of the i-th row and j-th column of the original frame image is G (i, j)
  • the pixel value H (i, j) of the i-th row and j-th column after the conversion in the first stage Is represented by the following formula (1).
  • H (i, j) ⁇ neighberpixcels
  • c 1
  • ⁇ neighberpixcels is the sum of eight pixels adjacent to the pixel in the i-th row and j-th column.
  • the pixel value G (i, j) and the neighboring pixel value G (i + di, j + dj) change similarly. For this reason, the image of H (i, j) is invariant to the change in illuminance.
  • the absolute value of the difference between adjacent pixels is generally larger as the pixel value is larger.
  • H (i, j) is normalized as in the following formula (2).
  • H (i, j) ⁇ neighberpixcels
  • this H (i, j) is converted to I (i, j) by the following equation (3) using a sigmoid function.
  • the threshold value ⁇ is set to the most frequent value of the frequency distribution of H having edge information, for example, 80.
  • the image conversion unit 22 converts the image having the pixel value G (i, j) into a spatial difference frame image having the pixel value I (i, j). Store in the image memory 21.
  • the background image generation unit 24, the ID generation / annihilation unit 25, and the moving object tracking unit 27 perform processing based on the spatial difference frame image in the image memory 21.
  • the spatial difference frame image is referred to as a frame image.
  • the background image generation unit 24 includes a storage unit and a processing unit.
  • the processing unit accesses the image memory 21 and creates a histogram of pixel values for the corresponding pixels of all the frame images for the past 10 minutes.
  • the processing unit generates an image having the mode value (mode) as the pixel value of the pixel as a background image in which no moving object is present, and stores this in the storage unit.
  • the background image is updated by performing this process periodically.
  • the ID generation / annihilation unit 25 includes the positions of the slits EN1 to EN4 and EX1 to EX4 arranged at the four entrances to the intersection and the four exits from the intersection in the frame image. Size data is preset.
  • the ID generation / annihilation unit 25 reads the image data in the entrance slits EN1 to EN4 from the image memory 21, and determines whether there is a moving object in these entrance slits in units of blocks.
  • the meshes in FIG. 3 represent blocks, and one block is composed of 8 ⁇ 8 pixels. When one frame is composed of 480 ⁇ 640 pixels, one frame is divided into 60 ⁇ 80 blocks.
  • Whether or not there is a moving object in a block is determined by whether or not the sum of the absolute values of the differences between the pixels in the block and the corresponding pixels in the background image is greater than or equal to a predetermined value. This determination is similarly performed in the moving object tracking unit 27.
  • the ID generation / annihilation unit 25 determines that there is a moving object in the block, it gives a new object identification code (hereinafter referred to as ID) to this block. If the ID generation / annihilation unit 25 determines that a moving object is present in a block adjacent to the ID-assigned block, the ID generation / annihilation unit 25 assigns the same ID as the assigned block to this adjacent block.
  • ID is assigned to the corresponding block in the object map storage unit 26.
  • the moving object tracking unit 27 assigns and moves IDs for the blocks in the moving direction and disappears IDs for the blocks in the opposite direction with respect to the cluster that has passed through the entrance slit (that is, tracking process of the cluster). The tracking process by the moving object tracking unit 27 is performed up to the exit slit for each cluster.
  • the moving object tracking unit 27 Based on the object map at time t ⁇ 1 stored in the object map storage unit 26 and the frame images at times t ⁇ 1 and t stored in the image memory 21, the moving object tracking unit 27 performs time t ⁇ 1. Are created in the object map storage unit 26. Hereinafter, this operation will be described.
  • FIGS. 4A to 7B schematically show images at times t-1 and t.
  • the dotted lines in FIGS. 4A, 4B, and 6A to 7B are block boundaries.
  • the dotted lines in FIGS. 5A and 5B are pixel boundaries.
  • the block in the i-th row and the j-th column is denoted as B (i, j), and the block in the i-th row and the j-th column at the time t is denoted as B (t: i, j).
  • B (i, j) the block in the i-th row and the j-th column at the time t is denoted as B (t: i, j).
  • MV be the motion vector of block B (t ⁇ 1: 1, 4).
  • the block at time t that most corresponds to the area where block B (t ⁇ 1: 1, 4) has been moved by MV is found. In the case of FIG. 4B, this block is B (t: 1, 5).
  • the degree of correlation between the image of the block B (t: 1, 5) and the image of the block size area AX at time t ⁇ 1 is 1 in the area AX within the predetermined range AM. It is obtained every time the pixel is moved (that is, block matching).
  • the range AM is larger than the block, and one side thereof is 1.5 times the number of pixels on one side of the block.
  • the center of the range AM is a pixel at a position where the center of the block B (t: 1, 5) is moved by MV.
  • the correlation degree is a spatio-temporal texture correlation degree, and the larger the evaluation value UD that is the sum of absolute values of the difference between the block B (t: 1, 5) and the corresponding pixel value in the region AX, the larger the correlation degree.
  • an area AX having a maximum correlation within the range AM is obtained.
  • a vector starting from the center of this area and ending at the center of block B (1, 5) is determined as the motion vector of block B (t: 1, 5).
  • the ID of the block at time t ⁇ 1 that is closest to the region AX where the degree of correlation is maximum is determined as the ID of the block B (t: 1, 5).
  • the moving object tracking unit 27 assigns the same ID to blocks whose absolute values of motion vector differences between adjacent blocks are equal to or less than a predetermined value. Thereby, one cluster is divided into a plurality of objects (moving objects) having different IDs. In FIG. 6A and FIG. 6B, the boundary between objects is shown by the thick line.
  • FIG. 6A and FIG. 6B show object boundaries in bold lines on the object map, and correspond to FIG. 6A and FIG. 6B.
  • each object is traced back in time after one cluster is separated into a plurality of clusters.
  • the motion vector of the block in the cluster is obtained has been described.
  • FIG. 9A when there is a block for which the motion vector cannot be obtained, depending on the position of the block, It may be unknown if it belongs. If the color of each pixel in a block belonging to a certain moving object is almost the same, the motion vector cannot be determined by the block matching described above. For example, if an image (spatial difference frame image) is converted into a binary image and the number of “1” pixels in the block is equal to or smaller than a predetermined value, the block is not suitable for obtaining a motion vector by the above method. It is determined.
  • the motion vector of such a block is estimated by the method shown in FIG.
  • Step S1 If there is a block whose motion vector is undetermined, the process proceeds to step S2, and if it does not exist, the undetermined motion vector estimation process is terminated.
  • Step S2 The determined motion vectors MV1 to MVn are extracted from the eight blocks around the block B (i, j) whose motion vectors are undetermined.
  • Step S3 If the motion vector determined in step S2 exists, the process proceeds to step S4, and if not, the process proceeds to step S6.
  • Step S4 The motion vectors MV1 to MVn are divided into groups in which the absolute value of the difference between the vectors is within a predetermined value.
  • Step S5 The average value of the motion vectors of the group having the largest number of motion vectors is estimated as the motion vector of the block B (i, j).
  • the average value of the motion vectors of any one group is estimated as the motion vector of the block B (i, j).
  • the process returns to step S1.
  • any one of the motion vectors of the same group may be estimated as the motion vector of the block B (i, j).
  • Step S6 The motion vector estimated in step S5 is set as the determined motion vector, and the process returns to step S1.
  • This process makes it possible to uniquely estimate an undetermined motion vector.
  • the motion vector of the block B (i, j) in the i-th row and j-th column is denoted as MV (i, j).
  • the motion vectors of blocks B (2,2), B (2,4) and B (3,3) are undetermined.
  • FIG. 9B an object map as shown in FIG. 9B is generated.
  • the boundary of the object is indicated by a thick line.
  • step S6 the estimated motion vector is regarded as the determined motion vector, and steps S1 to S5 are executed again. Thereby, the motion vector of the block B (3, 4) is uniquely estimated as shown in FIG. 10C.
  • step S6 by assigning the same ID to blocks whose absolute values of motion vector differences between adjacent blocks are equal to or less than a predetermined value, one cluster is divided into a plurality of objects having different IDs.
  • the moving object tracking unit 27 stores the time series of the object map stored in the object map storage unit 26 in a hard disk (not shown) as a tracking result.
  • each image of the time-series image is divided into a plurality of blocks in the object map storage unit 26, and an identification code indicating the moving object in the image is attached to the block corresponding to the moving object.
  • the motion vector of the moving object corresponding to the block is attached to the block and stored.
  • the moving object tracking unit 27 updates the block identification code and the motion vector stored in the object map storage unit 26 based on the result of image processing of the time-series image as described above. Specifically, the moving object tracking unit 27 updates the block identification code and the motion vector stored in the object map storage unit 26 according to the following procedures (1) to (4).
  • the contour extraction unit 30 extracts the contour of the moving object from the time-series image.
  • the contour extracting unit 30 extracts a plurality of moving objects having occlusions as a single moving object, and extracts a contour of the moving object as a single unit from a time-series image.
  • the contour extraction unit 30 includes a target region setting unit 301, a target region correction unit 302, and a contour extraction processing unit 303.
  • the target area setting unit 301 sets a target area for extracting the outline of the moving object based on the image area corresponding to the block of the moving object stored in the object map storage unit 26.
  • the contour extraction processing unit 303 extracts the contour of the moving object from the time-series image for the target region set by the target region setting unit 301.
  • the target area correction unit 302 projects, for each coordinate axis, a histogram of the number of pixels corresponding to the edges in the target area set by the target area setting unit 301 with respect to an image obtained by performing edge extraction processing on a time-series image. And generate. Based on the histogram generated for each coordinate axis, the target area set by the target area setting unit 301 is corrected for each coordinate axis.
  • the contour extraction processing unit 303 described above may extract the contour of the moving object from the time-series image with respect to the target region corrected by the target region correction unit 302.
  • the correction unit 31 corrects the block identification code and the motion vector stored in the object map storage unit 26 based on the contour extracted by the contour extraction processing unit 303 of the contour extraction unit 30.
  • the correction unit 31 also includes information indicating the boundaries of the plurality of moving objects based on the identification codes corresponding to the plurality of moving objects in which occlusion occurs and stored in the object map storage unit 26, and contour extraction.
  • the block identification code and the motion vector stored in the object map storage unit 26 are corrected based on the extracted outline of the moving object obtained by integrating the plurality of moving objects in which occlusion has occurred.
  • the determination unit 32 determines whether or not the background image is changing. The determination unit 32 determines whether or not the background image has changed based on an input signal indicating that the camera 10 has been panned or zoomed.
  • the determination unit 32 may compare the background image generated by the background image generation unit 24 with the image input from the camera 10 and determine whether or not the background image is fluctuating. Whether or not the background image is fluctuated by detecting a change in the position of the marker in the image included in the image input from the camera 10 by the marker being embedded in the background region in advance. It may be determined whether or not.
  • the control unit 34 controls the contour extraction unit 30 every predetermined period or every predetermined frame. Then, the contour is extracted, and the correction unit 31 is controlled to correct the block identification code and the motion vector stored in the object map storage unit 26.
  • the moving object fluctuation amount detection unit 33 will be described later.
  • an object map may be created by regarding the background image as an object.
  • the object map generation method is different from the background image only in determining whether or not a moving object exists in the block. Since the background image is also regarded as an object, block matching is performed on all the blocks to assign IDs and determine MVs.
  • an ID predetermined for the background image may be given to the background image.
  • the background ID and the moving object can be easily identified by this predetermined ID.
  • the block belongs between the background image and the moving object as shown in FIGS. 4A to 7B.
  • the image can be determined.
  • the moving object tracking device 20 is in the camera fixed mode.
  • the background image is not given an ID as an object, and the image generated by the background image generation unit 24 is used as the background image. This is because the background image does not change because the camera is fixed.
  • the moving object tracking unit 27 identifies the moving object.
  • Step S1 In accordance with the panning or zooming of the camera, the determination unit 32 determines that the background image is fluctuating. In response to this determination, the control unit 34 shifts from the camera fixing mode to the camera variation mode, registers the background image in the object, and assigns an ID.
  • Step S2 In the camera variation mode, the control unit 34 shifts to the correction mode every predetermined period or every predetermined frame. In this correction mode, the control unit 34 controls the contour extraction unit 30 to extract a contour, and controls the correction unit 31 to correct the block identification code and motion vector stored in the object map storage unit 26. .
  • Step S3 When the correction is completed in the correction mode, the control unit 34 transitions from the correction mode to the camera fluctuation mode.
  • control unit 34 alternately switches between the camera change mode and the correction mode.
  • Step S4 Thereafter, when the determination unit 32 determines that the fluctuation of the camera 10 has stopped, the control unit 34 deletes the ID assigned to the background image, and uses the image generated by the background image generation unit 24 as the background image. Use.
  • a predetermined period such as 10 minutes is required for the background image generation unit 24 to generate the background image. Therefore, it is desirable to use the background image as an object until the background image is generated by the background image generation unit 24.
  • control unit 34 controls the contour extraction unit 30 to extract a contour, controls the correction unit 31, and is stored in the object map storage unit 26. The operation for correcting the block identification code and motion vector will be described.
  • Step S1201 First, the occlusion detection unit 35 determines whether or not occlusion has occurred based on the object map stored in the object map storage unit 26.
  • Step S1202 If it is determined in step S1201 that no occlusion has occurred, the target area setting unit 301 sets a target area for a moving object without occlusion.
  • the target area correction unit 302 generates the above-described histogram, and corrects the target area set by the target area setting unit 301 for each coordinate axis based on the generated histogram.
  • Step S1204 the contour extraction processing unit 303 extracts the contour of the moving object from the time-series image for the target region corrected by the target region correction unit 302.
  • Step S1205 the correction unit 31 corrects the block identification code and the motion vector stored in the object map storage unit 26 based on the contour extracted by the contour extraction unit 30.
  • Step S1212 On the other hand, if it is determined in step S1201 that occlusion has occurred, the target area setting unit 301 of the contour extraction unit 30 sets a plurality of moving objects in which occlusion has occurred as an integrated moving object.
  • Step S1213 the target area setting unit 301 sets a target area for an integrated moving object.
  • the target area correction unit 302 generates the above-described histogram for the integrated moving object, and corrects the target area set by the target area setting unit 301 for each coordinate axis based on the generated histogram.
  • Step S1215) the contour extraction processing unit 303 extracts the contour of the moving object integrated with the target region corrected by the target region correction unit 302 from the image of the time series image.
  • Step S1216 Next, as described with reference to FIG. 4A to FIG. 7B, the correction unit 31 performs object detection based on the boundary of the moving object in which the detected occlusion occurs and the contour extracted by the contour extraction unit 30. The block identification code and motion vector stored in the map storage unit 26 are corrected.
  • control unit 34 transitions from the correction mode to the camera variation mode.
  • the moving object tracking device 20 can track a moving object even when the camera is fixed or fluctuates.
  • the camera fixing mode since the background image is not registered in the object, it is not necessary to execute the processing as illustrated in FIGS. 4A to 7B on the object of the background image. The amount of processing or load can be reduced.
  • the identification code and motion vector for each block stored in the object map storage unit 26 will be described as “space-time MRF (Markov Random Field)”.
  • space-time MRF Markov Random Field
  • Non-Patent Document 1 Kass et.al “Snakes: Active contour models”, Proc. Of 1st ICCV, pp.259-268, 1987
  • v (s) (x (s), y (s)) (0 ⁇ s ⁇ 1) expressed as a parameter on the image plane (x, y)
  • This is a contour extraction model that is deformed so as to minimize the energy function defined by the following equation (4) and whose shape is determined as a minimum state of energy.
  • the first term Eint in this equation (4) is internal energy.
  • the Snakes spline has a property of smoothly contracting into a convex shape.
  • the theoretical definition is expressed by the following formula (5).
  • the first term in the following formula (5) has the property that the spline becomes smooth in a convex shape, and the spline contracts by the second term.
  • the second term Eimage of the above formula (4) is image energy.
  • This image energy has a property that the value of the image energy becomes smaller as the ratio of the image energy existing on the edge (a portion such as a contour having a large luminance gradient) increases.
  • This image energy is defined by the following equation (6) by the luminance I (v (s)) of the image. This time, in order to stably extract the contour edge regardless of the illuminance, the illuminance invariant filter image developed so far by the present inventors was used as the image energy.
  • the illuminance invariant filter image is an image converted by the image conversion unit 22 according to the above formulas (1) to (3).
  • the third term Econ in the equation (4) is external energy. This external energy is used when force is forcedly applied to Snakes from the outside. This energy is defined as needed. This time, an area term (see Non-Patent Document 2 below) proposed for extracting a concave contour that was difficult to extract due to the influence of Eint used for internal energy was defined as external energy.
  • the area term Area is derived by the following equation (7) (see FIG. 13).
  • Non-patent literature 2 Shoichi Araki, Naokazu Yokoya, Hidehiko Sasaiwa, Haruo Sasatakemura: “Dynamic contour model splitting by intersection judgment for the purpose of extracting multiple objects”, Journal of the Institute of Electronics, Information and Communication Engineers (D-II) Vol.J79-DII, No.10, pp1704-1711 (Oct, 1996)
  • FIG. 14A to 14C show the processing results by Snakes.
  • initial control points are arranged around the object whose contour is to be extracted (FIG. 14A).
  • Snakes begins to contract (FIG. 14B).
  • the contraction stops near the contour line (FIG. 14C).
  • the initial control points shown in FIG. 14A correspond to the target area set by the target area setting unit 301.
  • the edge distribution (binary distribution of the illuminance invariant filter image) in the surrounding rectangle of the human object obtained by the object map is analyzed, and the human area is estimated. Thereby, the accuracy of contour extraction is improved by arranging the initial control points around it.
  • the analysis of the edge distribution is performed by projecting onto the horizontal axis and the vertical axis and generating a histogram (see FIGS. 15A to 15C).
  • the edge distribution analysis is executed by the following steps STEP1 to STEP3.
  • STEP 2 Horizontal axis histogram analysis
  • the obtained edge distribution is projected on the horizontal axis to generate a histogram. Then, the window is scanned in the obtained horizontal axis histogram. Thus, by obtaining a continuously distributed region, the distribution region of the human edge in the horizontal component is estimated (FIG. 15A).
  • STEP3 Vertical axis histogram analysis
  • the edge distribution is projected onto the vertical axis, and a histogram is generated. Then, the window is scanned in the obtained horizontal axis histogram. Thus, the distribution area of the person edge in the horizontal component is estimated by obtaining the continuously distributed area (FIG. 15B).
  • a vertical axis histogram is generated from the edge distribution obtained in STEP 1.
  • the vertical axis histogram may be generated from the edge distribution in the local region narrowed in STEP2.
  • the histogram threshold (initial control point) may be set as follows. First, the histogram frequency values are clustered into two groups.
  • the clustering method may be any method such as a k-mean method (one-dimensional). Thereby, the frequency value of a histogram is divided into a high frequency cluster and a low frequency cluster.
  • the boundary between the previous low-frequency position and this high-frequency position is set as a boundary.
  • Inter-layer cooperation algorithm> a processing step of tracking by cooperation between layers of the spatio-temporal MRF and Snakes will be described.
  • correcting the object map using the spatio-temporal MRF and Snakes will be referred to as inter-layer cooperation.
  • STEP2 For each object, edge distribution analysis is performed in the local region obtained in STEP 2, and Snakes initial control points are arranged around the contour of the object.
  • STEP3 Execute Snakes on each object. Compared to the size of the circumscribed rectangle obtained in STEP 1, the object map is not corrected for an object whose spline has contracted too much. For other objects, the processing result of Snakes is reflected and the object map is corrected.
  • STEP2 Edge distribution analysis is performed in the local region obtained in STEP 2, and Snakes initial control points are arranged around the contour of the object.
  • FIG. 16A to FIG. 17B show examples of object map correction by Snakes.
  • 16A and 16B are examples in the case where there is no occlusion
  • FIGS. 17A and 17B are examples in the case of occlusion.
  • the persons with ID numbers 6 and 7 are grouped together, and the outline of the group (the boundary between the background and the person) is obtained by Snakes.
  • the region division within the group reflects the output information by the spatiotemporal MRF.
  • FIGS. 18A to 29B effects of the inter-layer cooperation algorithm when the camera 10 changes will be described.
  • FIGS. 18A to 29B the processing results of moving object detection in the case of no inter-layer cooperation and inter-layer cooperation for the same frame of the same scene will be described.
  • FIG. 18A to FIG. 23A show the processing results when inter-layer cooperation is not performed, and FIG. 18B to FIG. 23B are object maps thereof.
  • FIG. 24A to FIG. 29A show the processing results when inter-layer cooperation is performed, and FIG. 24B to FIG. 29B are object maps thereof.
  • the boundary between objects can be corrected by referring to the output of the spatio-temporal MRF, and tracking can be performed for a long time.
  • the cooperation algorithm between layers the right two columns of the frame 142 and the frame 153.
  • FIG. 30A and FIG. 30B are results of detecting a moving object when the camera 10 fluctuates when occlusion occurs.
  • FIG. 30A no occlusion has occurred.
  • FIG. 30B occlusion has occurred.
  • the moving object tracking device 20 can track the moving object.
  • the moving object tracking device 20 changes the background by correcting the block identification code and the motion vector stored in the object map storage unit based on the extracted contour. Even from an image, a moving object in the image can be accurately detected.
  • the motion vector is corrected together with the block identification code stored in the object map storage unit.
  • the identification code may be corrected. Even in this way, similarly, a moving object in an image can be accurately detected from an image whose background changes.
  • the moving object fluctuation amount detection unit 33 described above detects the size of the moving object or the fluctuation amount of the movement amount stored in the object map storage unit 26 based on the identification code or the motion vector for each unit time.
  • control unit 34 (second control unit) is configured such that the size of the moving object or the amount of movement of the moving object per unit time detected by the moving object fluctuation amount detection unit 33 is a predetermined size or movement amount. Is larger than the fluctuation amount, the contour extraction unit 30 is controlled to extract the contour, and the correction unit 31 is controlled to correct the block identification code and motion vector stored in the object map storage unit 26.
  • the block identification code and the motion vector stored in the object map storage unit 26 may be corrected in accordance with the size of the moving object stored in the object map storage unit 26 or the fluctuation amount of the movement amount. Good.
  • control unit 34 compares the object map at a timing at which it is likely to fail to detect a moving object, as compared with the case where the control unit 34 simply corrects the object map every predetermined period or every predetermined frame. Can be corrected. Therefore, a moving object can be detected and tracked more accurately.
  • the storage unit such as the frame buffer memory 3, the image memory 21, or the object map storage unit 26 in FIG. 2 is a non-volatile memory such as a hard disk device, a magneto-optical disk device, a flash memory, or a CD-ROM. May be configured by a storage medium that can only be stored in the memory, a volatile memory such as a RAM (Random Access Memory), or a combination thereof.
  • a non-volatile memory such as a hard disk device, a magneto-optical disk device, a flash memory, or a CD-ROM. May be configured by a storage medium that can only be stored in the memory, a volatile memory such as a RAM (Random Access Memory), or a combination thereof.
  • RAM Random Access Memory
  • the processing unit called the control unit 34 or the occlusion detection unit 35 may be realized by dedicated hardware.
  • the processing unit may be configured by a memory and a CPU (central processing unit), and the function may be realized by loading a program for realizing the function of the processing unit into the memory and executing the program.
  • the image conversion unit 22, the background generation unit 24, the ID generation / annihilation unit 25, the moving object tracking unit 27, the contour extraction unit 30, the correction unit 31, the determination unit 32, the moving object variation amount detection unit 33 By recording a program for realizing the function of the processing unit such as the control unit 34 or the occlusion detection unit 35 on a computer-readable recording medium, causing the computer system to read and execute the program recorded on the recording medium
  • the processing by this processing unit may be executed.
  • the “computer system” includes an OS and hardware such as peripheral devices.
  • the “computer system” includes a homepage providing environment (or display environment) if a WWW system is used.
  • the “computer-readable recording medium” is a portable medium such as a flexible disk, a magneto-optical disk, a ROM, a CD-ROM, or a storage device such as a hard disk built in a computer system.
  • “computer-readable recording medium” is a program that dynamically holds a program for a short time, such as a communication line when a program is transmitted via a network such as the Internet or a communication line such as a telephone line.
  • a volatile memory in a computer system that serves as a server or a client in this case includes a program that holds a program for a certain period of time.
  • the program may be a program for realizing a part of the functions described above, and may be a program capable of realizing the functions described above in combination with a program already recorded in a computer system. .
  • the moving object in the image can be detected even when the background of the image fluctuates. It can be detected accurately.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)
  • Closed-Circuit Television Systems (AREA)
  • Traffic Control Systems (AREA)

Abstract

A moving object tracking device (20) detects a moving object in a time series image by performing image processing. The moving object tracking device (20) comprises an object map storage unit (26), a contour extraction unit (30), and a correction unit (31). The object map storage unit (26) allocates an identification code corresponding to the moving object to each of a plurality of blocks obtained by dividing each frame in the time series image and stores the identification code. The contour extraction unit (30) extracts the contour of the moving object from the time series image. The correction unit (31) corrects the identification code allocated to the block on the basis of the contour.

Description

移動物体追跡装置、移動物体追跡方法および移動物体追跡プログラムMoving object tracking device, moving object tracking method, and moving object tracking program
 本発明は、時系列画像を画像処理して画像中の移動物体(例えば、車、自転車、及び動物等)を追跡する移動物体追跡装置、移動物体追跡方法および移動物体追跡プログラムに関する。
 本願は、2009年4月2日に、日本に出願された特願2009-090489号に基づき優先権を主張し、その内容をここに援用する。
The present invention relates to a moving object tracking device, a moving object tracking method, and a moving object tracking program for tracking a moving object (for example, a car, a bicycle, and an animal) in an image by processing a time-series image.
This application claims priority based on Japanese Patent Application No. 2009-090489 filed in Japan on April 2, 2009, the contents of which are incorporated herein by reference.
 近年、カメラで撮像された画像を画像処理して、画像中の移動物体を正確に検出する技術が求められている。例えば、交通事故の早期発見は、迅速な救助活動により人命救助の成功率を高めるだけでなく、警察の実地検分などを迅速にさせて事故渋滞を緩和することもできる。このため、様々な交通事故に対する認識の自動化が期待されている。交通事故の認識率を高めるためには、カメラで撮像された画像を画像処理して移動物体を正確に検出する必要がある。 In recent years, there has been a demand for a technique for accurately detecting a moving object in an image by processing an image captured by a camera. For example, early detection of traffic accidents can not only increase the success rate of lifesaving by quick rescue activities, but also speed up on-site inspection by the police to alleviate traffic jams. For this reason, automation of recognition for various traffic accidents is expected. In order to increase the recognition rate of traffic accidents, it is necessary to accurately detect a moving object by performing image processing on an image captured by a camera.
 図31A~図31Dは、高速道路の中央線上方に設置されたカメラで撮像された時刻t=1~4の画像を模式的に示す。 31A to 31D schematically show images at time t = 1 to 4 captured by a camera installed above the center line of the expressway.
 画像上で車両同士が頻繁に重なる場合、画像処理により各車両を追跡するのが困難になるという問題がある。この問題を解決するには、道路に沿って複数台のカメラを設置し、それらのカメラによって撮影された画像を総合的に画像処理する方法がある。 When vehicles frequently overlap on the image, there is a problem that it becomes difficult to track each vehicle by image processing. In order to solve this problem, there is a method in which a plurality of cameras are installed along the road and the images photographed by these cameras are comprehensively processed.
 しかし、この方法は、カメラ及び画像処理装置を複数台備える必要があるので、コスト高になってしまう。また、この場合、各カメラの撮影画像を関係付けて総合的に画像処理しなければならないので、画像処理が複雑になってしまう。 However, since this method needs to include a plurality of cameras and image processing apparatuses, the cost becomes high. Further, in this case, the image processing becomes complicated because the image processing of each camera must be related and comprehensively processed.
 そこで、本願発明者らは、下記特許文献1および下記特許文献2に記載のように、時間を遡って移動物体を検出する方法で上記問題を解決した。 Therefore, the inventors of the present application solved the above problem by a method of detecting a moving object retroactively as described in Patent Document 1 and Patent Document 2 below.
 まず、時刻t=1~4の時系列画像を一時的に記憶する。時刻t=4で、車両M1とM2とを識別し、これら車両M1及びM2の動きベクトルを求める。これらの動きベクトルに基づいて、時刻t=4の画像中の車両M1及びM2を移動させることにより、車両M1とM2とが識別されている時刻t=3の画像を想定する。この想定された画像と実際の画像との相関関係から、時刻t=3の画像中の車両M1とM2とを識別する。 First, the time series images at time t = 1 to 4 are temporarily stored. At time t = 4, the vehicles M1 and M2 are identified, and the motion vectors of these vehicles M1 and M2 are obtained. An image at time t = 3 in which the vehicles M1 and M2 are identified is assumed by moving the vehicles M1 and M2 in the image at time t = 4 based on these motion vectors. From the correlation between the assumed image and the actual image, the vehicles M1 and M2 in the image at time t = 3 are identified.
 次に、時刻t=3及び時刻t=2の画像についても同様の画像処理により、時刻t=2の画像中の車両M1とM2とを識別する。次に、時刻t=2及び時刻t=1の画像についても同様の画像処理により、時刻t=1の画像中の車両M1とM2とを識別する。 Next, for the images at time t = 3 and time t = 2, the vehicles M1 and M2 in the image at time t = 2 are identified by similar image processing. Next, for the images at time t = 2 and time t = 1, the vehicles M1 and M2 in the image at time t = 1 are identified by similar image processing.
 このような画像処理を行うことにより、1台のカメラで車両M1とM2とを追跡できる。 By performing such image processing, the vehicles M1 and M2 can be tracked with a single camera.
特開2002-133421号公報JP 2002-133421 A 特開2004-207786号公報Japanese Patent Laid-Open No. 2004-207786
 上記の従来技術においては、カメラが固定されている状態で撮像された画像、すなわち、背景が固定されている画像を画像処理して、画像中の移動物体を正確に検出していた。 In the above-described prior art, an image captured with the camera fixed, that is, an image with a fixed background is subjected to image processing to accurately detect a moving object in the image.
 これに対して、カメラをパンニングまたはズーミングした場合には、それまで固定されていた背景画像も、カメラのパンニングまたはズーミングに応じて変動してしまう。このように背景が変動する画像を画像処理した場合には、移動物体の画像と背景画像との境界が明確でなくなり、画像中の移動物体を正確に検出できないという問題があった。 On the other hand, when the camera is panned or zoomed, the background image that has been fixed up to that time also varies depending on the camera panning or zooming. When image processing is performed on an image whose background changes in this way, there is a problem in that the boundary between the image of the moving object and the background image is not clear, and the moving object in the image cannot be accurately detected.
 本発明は、上記事情を鑑みてなされたものであって、背景が変動する画像中の移動物体を正確に検出できる移動物体追跡装置、移動物体追跡方法および移動物体追跡プログラムの提供を目的とする。 The present invention has been made in view of the above circumstances, and an object of the present invention is to provide a moving object tracking device, a moving object tracking method, and a moving object tracking program that can accurately detect a moving object in an image whose background changes. .
 本発明は、上記課題を解決して係る目的を達成するために以下の手段を採用した。
(1)本発明の移動物体追跡装置は、画像処理により時系列画像中の移動物体を検出する移動物体追跡装置であって、前記移動物体に対応する識別符号を、前記時系列画像の各フレームが複数に分割された各ブロックに割付けて記憶するオブジェクトマップ記憶部と;前記時系列画像から前記移動物体の輪郭を抽出する輪郭抽出部と;前記輪郭に基づいて、前記ブロックに割付けられた前記識別符号を補正する補正部と;を備える。
The present invention employs the following means in order to solve the above problems and achieve the object.
(1) The moving object tracking device of the present invention is a moving object tracking device that detects a moving object in a time-series image by image processing, and an identification code corresponding to the moving object is assigned to each frame of the time-series image. Is assigned to each block divided into a plurality of blocks, and is stored; an outline extraction unit that extracts an outline of the moving object from the time-series image; and the blocks assigned to the blocks based on the outlines A correction unit that corrects the identification code.
(2)上記(1)に記載の移動物体追跡装置では、前記輪郭抽出部が、前記オブジェクトマップ記憶部に記憶されている前記移動物体に対応する画像領域に基づいて、前記移動物体の輪郭を抽出する対象領域を設定する対象領域設定部と;前記対象領域に対して、前記時系列画像から前記移動物体の輪郭を抽出する輪郭抽出処理部と;を備えていてもよい。 (2) In the moving object tracking device according to (1), the contour extraction unit calculates the contour of the moving object based on an image area corresponding to the moving object stored in the object map storage unit. A target region setting unit that sets a target region to be extracted; and a contour extraction processing unit that extracts a contour of the moving object from the time-series image with respect to the target region.
(3)上記(2)に記載の移動物体追跡装置では、前記輪郭抽出部が、前記時系列画像をエッジ抽出処理した画像に対して、前記対象領域内でエッジに対応する画素数の個数についてのヒストグラムを座標軸毎に射影して生成し、このヒストグラムに基づいて前記対象領域を前記座標軸毎に補正する対象領域補正部をさらに備え;前記輪郭抽出処理部が、補正された前記対象領域に対して、前記時系列画像から前記移動物体の輪郭を抽出する;構成を採用してもよい。 (3) In the moving object tracking device according to (2) above, the number of pixels corresponding to an edge in the target region is determined by the contour extraction unit with respect to an image obtained by performing edge extraction processing on the time-series image. And a target area correction unit that corrects the target area for each coordinate axis based on the histogram, and the contour extraction processing unit applies the corrected target area to the corrected target area. Then, the contour of the moving object is extracted from the time-series image; a configuration may be adopted.
(4)上記(1)に記載の移動物体追跡装置では、前記輪郭抽出部が、オクルージョンが発生している複数の移動物体を一体の移動物体とし、前記時系列画像から前記一体の移動物体の輪郭を抽出してもよい。 (4) In the moving object tracking device according to (1), the contour extraction unit sets a plurality of moving objects in which occlusion has occurred as an integrated moving object, and extracts the integrated moving object from the time-series image. An outline may be extracted.
(5)上記(4)に記載の移動物体追跡装置では、前記補正部が、前記オブジェクトマップ記憶部に記憶されている前記オクルージョンが発生している複数の移動物体に対応する識別符号に基づいたこれら複数の移動物体の境界を示す情報と、前記輪郭抽出部により前記オクルージョンが発生している複数の移動物体を一体として抽出された前記移動物体の輪郭とに基づいて、前記オブジェクトマップ記憶部に記憶されている前記識別符号を補正してもよい。 (5) In the moving object tracking device according to (4), the correction unit is based on identification codes corresponding to a plurality of moving objects in which the occlusion occurs, which is stored in the object map storage unit. Based on the information indicating the boundaries of the plurality of moving objects and the contour of the moving object that is extracted by integrating the plurality of moving objects in which the occlusion is generated by the contour extraction unit, the object map storage unit stores The stored identification code may be corrected.
(6)上記(1)に記載の移動物体追跡装置が、背景画像が変動しているか否かを判定する判定部と;前記判定部が前記背景画像が変動していると判定した場合に、前記輪郭抽出部に輪郭を抽出させ、前記補正部に前記オブジェクトマップ記憶部に記憶されている前記識別符号を補正させる第1の制御部と;をさらに備えていてもよい。 (6) The moving object tracking device according to (1) above, a determination unit that determines whether or not the background image is fluctuating; and when the determination unit determines that the background image is fluctuating, A first control unit that causes the contour extraction unit to extract a contour and causes the correction unit to correct the identification code stored in the object map storage unit.
(7)上記(1)に記載の移動物体追跡装置が、前記オブジェクトマップ記憶部に記憶されている前記移動物体のサイズまたは移動量の変動量を、前記識別符号に基づいて単位時間毎に検出する移動物体変動量検出部と;前記移動物体変動量検出部により検出された単位時間毎の前記移動物体のサイズまたは移動量の前記変動量が所定のサイズまたは移動量の変動量よりも大きい場合に、前記輪郭抽出部に輪郭を抽出させ、前記補正部に前記オブジェクトマップ記憶部に記憶されている前記識別符号を補正させる第2の制御部と;をさらに備えていてもよい。 (7) The moving object tracking device according to (1) detects the size of the moving object or the amount of change in the movement amount stored in the object map storage unit for each unit time based on the identification code. A moving object fluctuation amount detecting unit that detects the moving object size or movement amount of the moving object per unit time detected by the moving object fluctuation amount detection unit is larger than a predetermined size or movement amount fluctuation amount And a second control unit that causes the contour extraction unit to extract a contour and causes the correction unit to correct the identification code stored in the object map storage unit.
(8)上記(1)に記載の移動物体追跡装置では、移動物体追跡装置が、前記時系列画像を画像処理した結果に基づいて、前記識別符号と、前記オブジェクトマップ記憶部に記憶されている前記移動物体の動きベクトルと、を更新する移動物体追跡部をさらに備え、前記移動物体追跡部が、前記時系列画像のうちの連続するN画像(N≧2)の各々に対して、隣り合うブロックの動きベクトルの差の絶対値が所定値以内のブロックに同一の識別符号を付けることにより、画像上で互いに重なった前記移動物体に互いに異なる識別符号を付ける識別符号付与工程と;前記N画像の各々において、第1識別符号が付けられたブロック群である第1オブジェクトと第2識別符号が付けられたブロック群である第2オブジェクトとが接し、かつ、前記N画像について時間的に隣り合う画像の前記第1オブジェクト間の相関度が所定値以上であるか否かを判定する判定工程と;前記判定工程で肯定と判定された場合に、時間を遡って前記第1オブジェクトと前記第2オブジェクトとを追跡する追跡工程と;前記追跡工程により時間を遡って追跡された前記第1オブジェクトと前記第2オブジェクトとに基づいて、前記オブジェクトマップ記憶部に記憶されている前記識別符号と前記動きベクトルとを更新する更新工程と;を実行する構成を採用してもよい。 (8) In the moving object tracking device according to (1), the moving object tracking device stores the identification code and the object map storage unit based on the result of image processing of the time-series image. And a moving object tracking unit that updates the motion vector of the moving object, and the moving object tracking unit is adjacent to each of the consecutive N images (N ≧ 2) of the time-series images. An identification code providing step of attaching different identification codes to the moving objects that overlap each other on an image by attaching the same identification code to blocks whose absolute values of motion vector differences of the blocks are within a predetermined value; Each of the first object, which is a block group to which a first identification code is attached, and the second object, which is a block group to which a second identification code is attached, and A determination step of determining whether or not the degree of correlation between the first objects of temporally adjacent images with respect to an image is greater than or equal to a predetermined value; and when the determination step determines affirmative, A tracking step of tracking the first object and the second object; and a data stored in the object map storage unit based on the first object and the second object tracked back in time by the tracking step. An update process for updating the identification code and the motion vector may be employed.
(9)本発明の移動物体追跡方法は、画像処理により時系列画像中の移動物体を検出する移動物体追跡方法であって、前記移動物体に対応する識別符号を、前記時系列画像の各フレームが複数に分割されたブロックに割付けて記憶する工程と;前記時系列画像から前記移動物体の輪郭を抽出する工程と;前記輪郭に基づいて、前記ブロックに割付けられた前記識別符号を補正する工程と;を備える。 (9) The moving object tracking method of the present invention is a moving object tracking method for detecting a moving object in a time-series image by image processing, and an identification code corresponding to the moving object is assigned to each frame of the time-series image. Assigning to a block divided into a plurality of blocks, storing the blocks, extracting a contour of the moving object from the time-series image, correcting the identification code assigned to the block based on the contours And comprising.
(10)本発明の移動物体追跡プログラムは、画像処理により時系列画像中の移動物体を検出する移動物体追跡装置としてのコンピュータに、前記移動物体に対応する識別符号を、前記時系列画像の各フレームが複数に分割されたブロックに割付けて記憶する工程と;前記時系列画像から前記移動物体の輪郭を抽出する工程と;前記輪郭に基づいて、前記ブロックに割付けられた前記識別符号を補正する工程と;を実行させる。 (10) The moving object tracking program of the present invention provides a computer as a moving object tracking device that detects a moving object in a time-series image by image processing, and assigns an identification code corresponding to the moving object to each of the time-series images. Allocating and storing a frame into a plurality of divided blocks; extracting a contour of the moving object from the time-series image; and correcting the identification code allocated to the block based on the contour And a process is executed.
 本発明によれば、抽出した移動物体の輪郭に基づいて、オブジェクトマップ記憶部に記憶されているブロックの識別符号を補正することにより、画像の背景が変動する場合でも、画像中の移動物体を正確に検出できる。 According to the present invention, by correcting the block identification code stored in the object map storage unit based on the extracted contour of the moving object, the moving object in the image can be detected even when the background of the image fluctuates. It can be detected accurately.
本発明の一実施形態に係る移動物体追跡装置を用いた移動物体追跡システムの概略図である。1 is a schematic diagram of a moving object tracking system using a moving object tracking device according to an embodiment of the present invention. 同実施形態に係る移動物体追跡装置の構成を示すブロック図である。It is a block diagram which shows the structure of the moving object tracking apparatus which concerns on the same embodiment. フレーム画像内の、交差点への4つの入口及び交差点からの4つの出口にそれぞれ設定されたスリット及びブロックに付与された移動物体のIDを示す図である。It is a figure which shows ID of the moving object provided to the slit and block which were each set to the four entrances to an intersection and the four exits from an intersection in a frame image. 時刻t-1における画像を、ブロック境界線とともに模式的に示す図である。It is a figure which shows typically the image in the time t-1 with a block boundary line. 時刻tにおける画像を、ブロック境界線とともに模式的に示す図である。It is a figure which shows the image in the time t typically with a block boundary line. 時刻t-1における画像を、画素境界線とともに模式的に示す図である。It is a figure which shows typically the image in the time t-1 with a pixel boundary line. 時刻tにおける画像を、画素境界線とともに模式的に示す図である。It is a figure which shows the image in the time t typically with a pixel boundary line. 時刻t-1における画像を、ブロックに付与された動きベクトルとともに模式的に示す図である。It is a figure which shows typically the image in the time t-1 with the motion vector provided to the block. 時刻tにおける画像を、ブロックに付与された動きベクトルとともに模式的に示す図である。It is a figure which shows typically the image in the time t with the motion vector provided to the block. 時刻t-1における、オブジェクトマップに付与された動きベクトル及びオブジェクト境界を模式的に示す図である。It is a figure which shows typically the motion vector and object boundary which were provided to the object map at the time t-1. 時刻tにおける、オブジェクトマップに付与された動きベクトル及びオブジェクト境界を模式的に示す図である。It is a figure which shows typically the motion vector and object boundary provided to the object map at the time t. 未定動きベクトルの推定方法を示すフローチャートである。It is a flowchart which shows the estimation method of an undetermined motion vector. 同推定方法を適用した場合の、時刻t-1における、オブジェクトマップに付与された動きベクトル及びオブジェクト境界を模式的に示す図である。FIG. 10 is a diagram schematically showing motion vectors and object boundaries assigned to an object map at time t−1 when the same estimation method is applied. 同時刻tにおける、オブジェクトマップに付与された動きベクトル及びオブジェクト境界を模式的に示す図である。It is a figure which shows typically the motion vector and object boundary provided to the object map at the same time t. 同推定方法を適用した場合の、オブジェクトマップに付与された動きベクトル及びオブジェクト境界の第1段階を模式的に示す図である。It is a figure which shows typically the 1st step of the motion vector and object boundary provided to the object map at the time of applying the estimation method. 同第2段階を模式的に示す図である。It is a figure which shows the said 2nd step typically. 同第3段階を模式的に示す図である。It is a figure which shows the said 3rd step typically. 移動物体追跡装置20におけるモードの状態遷移を示す状態遷移図である。FIG. 6 is a state transition diagram showing state transition of modes in the moving object tracking device 20. 図11の補正モードにおける移動物体追跡装置20の動作を示すフローチャートである。12 is a flowchart illustrating an operation of the moving object tracking device 20 in the correction mode of FIG. 11. 面積項Eareaを示す図である。It is a figure which shows area term Earea. Snakesの処理結果の一例の第1段階を示す図である。It is a figure which shows the 1st step of an example of the processing result of Snakes. 同第2段階を示す図である。It is a figure which shows the 2nd step. 同第3段階を示す図である。It is a figure which shows the said 3rd step. エッジ分布およびヒストグラムの生成の一例の第1段階を示す図である。It is a figure which shows the 1st step of an example of generation of edge distribution and a histogram. 同第2段階を示す図である。It is a figure which shows the 2nd step. 同第3段階を示す図である。It is a figure which shows the said 3rd step. Snakesによるオブジェクトマップの修正例を示す図である。It is a figure which shows the example of correction of the object map by Snakes. 同修正例を示す図である。It is a figure which shows the modification example. 同修正例を示す図である。It is a figure which shows the modification example. 同修正例を示す図である。It is a figure which shows the modification example. カメラ10が変動する場合において、階層間協調アルゴリズムを適用しない場合における移動物体のフレーム80での検出結果を示す図である。It is a figure which shows the detection result in the flame | frame 80 in the case where the camera 10 fluctuates, when not applying an inter-layer cooperation algorithm. 同フレーム80での検出結果を示す図である。FIG. 10 is a diagram showing a detection result in the same frame 80. 同フレーム95での検出結果を示す図である。It is a figure which shows the detection result in the same frame 95. FIG. 同フレーム95での検出結果を示す図である。It is a figure which shows the detection result in the same frame 95. FIG. 同フレーム107での検出結果を示す図である。It is a figure which shows the detection result in the same frame 107. FIG. 同フレーム107での検出結果を示す図である。It is a figure which shows the detection result in the same frame 107. FIG. 同フレーム115での検出結果を示す図である。It is a figure which shows the detection result in the same frame 115. FIG. 同フレーム115での検出結果を示す図である。It is a figure which shows the detection result in the same frame 115. FIG. 同フレーム142での検出結果を示す図である。It is a figure which shows the detection result in the same frame 142. FIG. 同フレーム142での検出結果を示す図である。It is a figure which shows the detection result in the same frame 142. FIG. 同フレーム153での検出結果を示す図である。It is a figure which shows the detection result in the same frame 153. FIG. 同フレーム153での検出結果を示す図である。It is a figure which shows the detection result in the same frame 153. FIG. カメラ10が変動する場合において、階層間協調アルゴリズムを適用する場合における移動物体のフレーム80での検出結果を示す図である。It is a figure which shows the detection result in the flame | frame 80 in the case of applying the cooperation algorithm between layers, when the camera 10 fluctuates. 同フレーム80での検出結果を示す図である。FIG. 10 is a diagram showing a detection result in the same frame 80. 同フレーム95での検出結果を示す図である。It is a figure which shows the detection result in the same frame 95. FIG. 同フレーム95での検出結果を示す図である。It is a figure which shows the detection result in the same frame 95. FIG. 同フレーム107での検出結果を示す図である。It is a figure which shows the detection result in the same frame 107. FIG. 同フレーム107での検出結果を示す図である。It is a figure which shows the detection result in the same frame 107. FIG. 同フレーム115での検出結果を示す図である。It is a figure which shows the detection result in the same frame 115. FIG. 同フレーム115での検出結果を示す図である。It is a figure which shows the detection result in the same frame 115. FIG. 同フレーム142での検出結果を示す図である。It is a figure which shows the detection result in the same frame 142. FIG. 同フレーム142での検出結果を示す図である。It is a figure which shows the detection result in the same frame 142. FIG. 同フレーム153での検出結果を示す図である。It is a figure which shows the detection result in the same frame 153. FIG. 同フレーム153での検出結果を示す図である。It is a figure which shows the detection result in the same frame 153. FIG. カメラ10が変動する場合において、オクルージョンが発生している場合における移動物体の検出結果を示す図である。It is a figure which shows the detection result of the moving object in case the occlusion has generate | occur | produced when the camera 10 fluctuates. 同検出結果を示す図である。It is a figure which shows the same detection result. 高速道路の中央線上方に設置されたカメラで撮像された時刻t=1での画像を模式的に示す図である。It is a figure which shows typically the image in the time t = 1 imaged with the camera installed above the center line of the highway. 同時刻t=2での画像を模式的に示す図である。It is a figure which shows typically the image in the same time t = 2. 同時刻t=3での画像を模式的に示す図である。It is a figure which shows typically the image in the same time t = 3. 同時刻t=4での画像を模式的に示す図である。It is a figure which shows typically the image in the same time t = 4.
 本発明の一実施形態を、図面を参照しながら以下に説明する。図1は、本発明の一実施形態に係る移動物体追跡装置20を用いた移動物体追跡システムの構成を示す概略図である。この移動物体追跡システムは、図1に示すように、交差点を撮像して画像信号を出力する電子カメラ10と、その画像を処理して移動物体を追跡する移動物体追跡装置20とを備えている。 An embodiment of the present invention will be described below with reference to the drawings. FIG. 1 is a schematic diagram showing a configuration of a moving object tracking system using a moving object tracking device 20 according to an embodiment of the present invention. As shown in FIG. 1, the moving object tracking system includes an electronic camera 10 that captures an intersection and outputs an image signal, and a moving object tracking device 20 that processes the image and tracks the moving object. .
 電子カメラ10で撮影された時系列画像は、12フレーム/秒のレートで、移動物体追跡装置20が有する後述する画像メモリ21に格納される。この画像メモリ21では、最も古いフレームが新しいフレーム画像に書き換えられる。この電子カメラ10は、パンニングまたはズーミングにより、撮影する画像領域を変更できる。この電子カメラ10に対するパンニングまたはズーミングは、移動物体追跡装置20が制御してもよく、この移動物体追跡システムを制御する上位制御装置が制御してもよい。 The time-series images captured by the electronic camera 10 are stored in an image memory 21 (described later) included in the moving object tracking device 20 at a rate of 12 frames / second. In this image memory 21, the oldest frame is rewritten with a new frame image. The electronic camera 10 can change an image area to be photographed by panning or zooming. The panning or zooming with respect to the electronic camera 10 may be controlled by the moving object tracking device 20 or may be controlled by a host control device that controls the moving object tracking system.
 移動物体追跡装置20は、電子カメラ10で撮影された時系列画像(後述する画像メモリ21に格納された時系列画像)を画像処理して、画像中の移動物体を検出する。 The moving object tracking device 20 performs image processing on a time series image (a time series image stored in an image memory 21 described later) taken by the electronic camera 10 to detect a moving object in the image.
 次に図2を用いて、移動物体追跡装置20の構成を説明する。 Next, the configuration of the moving object tracking device 20 will be described with reference to FIG.
 画像変換部22は、画像メモリ21内の各フレーム画像をフレームバッファメモリ23にコピーし、コピーされた画像のデータを用いて画像メモリ21内の対応するフレーム画像を空間的差分フレーム画像に変換する。この変換は、2段階で行われる。 The image conversion unit 22 copies each frame image in the image memory 21 to the frame buffer memory 23, and converts the corresponding frame image in the image memory 21 into a spatial difference frame image using the copied image data. . This conversion is performed in two stages.
 元のフレーム画像の第i行第j列の画素値(輝度値)をG(i,j)とすると、第1段階における変換後の第i行第j列の画素値H(i,j)は、下記式(1)で表される。 If the pixel value (luminance value) of the i-th row and j-th column of the original frame image is G (i, j), the pixel value H (i, j) of the i-th row and j-th column after the conversion in the first stage. Is represented by the following formula (1).
 H(i,j)=Σneighberpixcels|G(i+di,j+dj)-G(i,j)| ・・・(1) H (i, j) = Σneighberpixcels | G (i + di, j + dj) −G (i, j) | (1)
 ここで、Σneighberpixcelsは、cを自然数とすると、di=-c~c及びdj=-c~cにわたっての総和を意味する。例えば、c=1のとき、Σneighberpixcelsは、第i行第j列の画素と隣り合う8画素にわたる総和である。照度が変化すると、画素値G(i,j)とその付近の画素値G(i+di,j+dj)が同様に変化する。このため、H(i,j)の画像は、照度の変化に対し不変である。 Here, Σneighberpixcels means the sum over di = −c to c and dj = −c to c, where c is a natural number. For example, when c = 1, Σneighberpixcels is the sum of eight pixels adjacent to the pixel in the i-th row and j-th column. When the illuminance changes, the pixel value G (i, j) and the neighboring pixel value G (i + di, j + dj) change similarly. For this reason, the image of H (i, j) is invariant to the change in illuminance.
 ここで、隣り合う画素の差分の絶対値は、一般に画素値が大きいほど大きい。移動物体追跡の成功率を高めるには、画素値が小さくて差分が小さい場合も、画素値及び差分が大きい場合とほぼ等価にエッジ情報を取得することが好ましい。そこで、H(i,j)を下記式(2)ように規格化する。 Here, the absolute value of the difference between adjacent pixels is generally larger as the pixel value is larger. In order to increase the success rate of moving object tracking, it is preferable to acquire edge information almost equivalently to the case where the pixel value and the difference are large even when the pixel value is small and the difference is small. Therefore, H (i, j) is normalized as in the following formula (2).
 H(i,j)=Σneighberpixcels|G(i+di,j+dj)-G(i,j)|/(Gi,j,max/Gmax) ・・・(2) H (i, j) = Σneighberpixcels | G (i + di, j + dj) −G (i, j) | / (Gi, j, max / Gmax) (2)
 ここで、Gi,j,maxは、H(i,j)の計算に用いた元の画素の値の最大値である。例えば、c=1の場合、Gi,j,maxは、第i行第j列の画素を中心とする3×3画素の値の最大値である。Gmaxは、画素値G(i,j)の取りうる最大値、例えば画素値が8ビットで表される場合、255である。以下、c=1、Gmax=255である場合について説明する。 Here, Gi, j, max is the maximum value of the original pixel value used in the calculation of H (i, j). For example, when c = 1, Gi, j, max is the maximum value of 3 × 3 pixels centered on the pixel in the i-th row and j-th column. Gmax is 255 when the pixel value G (i, j) can take a maximum value, for example, when the pixel value is represented by 8 bits. Hereinafter, a case where c = 1 and Gmax = 255 will be described.
 H(i,j)の取りうる最大値は、移動物体毎に異なる。例えば、G(i,j)=Gmaxで第i行第j列の画素に隣り合う8画素の値がいずれも0で有る場合、H(i,j)=8Gmaxとなるため、H(i,j)を8ビットで表せない。 The maximum value that H (i, j) can take differs for each moving object. For example, if G (i, j) = Gmax and the values of 8 pixels adjacent to the pixel in the i-th row and j-th column are all 0, H (i, j) = 8 Gmax, so H (i, j, j) cannot be represented by 8 bits.
 一方、移動物体のエッジ部のH(i,j)の値のヒストグラムを作成すると、頻度の大部分がH=50~110の範囲に含まれることがわかった。すなわち、Hの値が約110より大きくなるほど、移動物体追跡のためのエッジ情報の数が少ないので、その重要度が低い。 On the other hand, when a histogram of the values of H (i, j) at the edge portion of the moving object was created, it was found that most of the frequency was included in the range of H = 50 to 110. That is, as the value of H is larger than about 110, the number of pieces of edge information for tracking a moving object is small, so the importance is low.
 したがって、Hの値の大きい部分を抑圧して変換画素のビット長を短くすることにより、画像処理を高速に行なうのが好ましい。そこで、第2段階として、このH(i,j)を、シグモイド関数を用いた下記式(3)により、I(i,j)に変換する。 Therefore, it is preferable to perform image processing at high speed by suppressing the portion where the value of H is large and shortening the bit length of the converted pixel. Therefore, as a second stage, this H (i, j) is converted to I (i, j) by the following equation (3) using a sigmoid function.
 I=Gmax/{1+exp〔-β(H-α)〕} ・・・(3) I = Gmax / {1 + exp [−β (H−α)]}〕 (3)
 シグモイド関数は、H=αの付近で線形性が良い。そこで、閾値αの値を、エッジ情報を持つHの度数分布の最頻値、例えば80にする。 Sigmoid function has good linearity around H = α. Therefore, the threshold value α is set to the most frequent value of the frequency distribution of H having edge information, for example, 80.
 画像変換部22は、上式(2)及び(3)に基づいて、画素値G(i,j)の画像を画素値I(i,j)の空間的差分フレーム画像に変換し、これを画像メモリ21に格納する。 Based on the above equations (2) and (3), the image conversion unit 22 converts the image having the pixel value G (i, j) into a spatial difference frame image having the pixel value I (i, j). Store in the image memory 21.
 背景画像生成部24、ID生成/消滅部25及び移動物体追跡部27は、画像メモリ21中の空間的差分フレーム画像に基づいて処理を行う。以下、空間的差分フレーム画像をフレーム画像と称す。 The background image generation unit 24, the ID generation / annihilation unit 25, and the moving object tracking unit 27 perform processing based on the spatial difference frame image in the image memory 21. Hereinafter, the spatial difference frame image is referred to as a frame image.
 背景画像生成部24は、記憶部と処理部とを備える。処理部は、画像メモリ21にアクセスし、過去10分間の全てのフレーム画像の対応する画素について画素値のヒストグラムを作成する。この処理部は、その最頻値(モード)をその画素の画素値とする画像を、移動物体が存在しない背景画像として生成し、これをこの記憶部に格納する。背景画像は、この処理が定期的に行われて更新される。 The background image generation unit 24 includes a storage unit and a processing unit. The processing unit accesses the image memory 21 and creates a histogram of pixel values for the corresponding pixels of all the frame images for the past 10 minutes. The processing unit generates an image having the mode value (mode) as the pixel value of the pixel as a background image in which no moving object is present, and stores this in the storage unit. The background image is updated by performing this process periodically.
 ID生成/消滅部25には、図3に示すように、フレーム画像内の、交差点への4つの入口及び交差点からの4つの出口にそれぞれ配置されるスリットEN1~EN4及びEX1~EX4の位置及びサイズのデータが予め設定されている。ID生成/消滅部25は、画像メモリ21から入口スリットEN1~EN4内の画像データを読み込み、これら入口スリット内に移動物体が存在するかどうかをブロック単位で判定する。図3中のメッシュの升目はブロックを表し、1ブロックは8×8画素からなる。1フレームが480×640画素からなる場合、1フレームは、60×80ブロックに分割される。あるブロックに移動物体が存在するかどうかは、このブロック内の各画素と背景画像の対応する画素との差の絶対値の総和が、所定値以上であるかどうかにより判定する。この判定は、移動物体追跡部27においても同様に行われる。 As shown in FIG. 3, the ID generation / annihilation unit 25 includes the positions of the slits EN1 to EN4 and EX1 to EX4 arranged at the four entrances to the intersection and the four exits from the intersection in the frame image. Size data is preset. The ID generation / annihilation unit 25 reads the image data in the entrance slits EN1 to EN4 from the image memory 21, and determines whether there is a moving object in these entrance slits in units of blocks. The meshes in FIG. 3 represent blocks, and one block is composed of 8 × 8 pixels. When one frame is composed of 480 × 640 pixels, one frame is divided into 60 × 80 blocks. Whether or not there is a moving object in a block is determined by whether or not the sum of the absolute values of the differences between the pixels in the block and the corresponding pixels in the background image is greater than or equal to a predetermined value. This determination is similarly performed in the moving object tracking unit 27.
 ID生成/消滅部25は、ブロック内に移動物体が存在すると判定すると、このブロックに新たなオブジェクト識別符号(以下、ID)を付与する。ID生成/消滅部25は、ID付与済ブロックと隣接しているブロックに移動物体が存在すると判定すると、この隣接ブロックに付与済ブロックと同一のIDを付与する。このID付与済ブロックは、入口スリットに隣接しているブロックも含まれる。例えば、図3中の入口スリットEN1内のブロックにはID=1が付与される。 When the ID generation / annihilation unit 25 determines that there is a moving object in the block, it gives a new object identification code (hereinafter referred to as ID) to this block. If the ID generation / annihilation unit 25 determines that a moving object is present in a block adjacent to the ID-assigned block, the ID generation / annihilation unit 25 assigns the same ID as the assigned block to this adjacent block. This ID-added block includes a block adjacent to the entrance slit. For example, ID = 1 is assigned to the block in the entrance slit EN1 in FIG.
 オブジェクトマップ記憶部26内の対応するブロックに対して、IDが付与される。オブジェクトマップ記憶部26は、上述の場合、60×80ブロックのオブジェクトマップを記憶する。各ブロックには、IDが付与されているかどうかのフラグと、IDが付与されている場合には、その番号と後述のブロックの動きベクトルとがブロック情報として付与される。なお、このフラグを用いずに、ID=0の場合、IDが付与されていないと判定してもよい。また、IDの最上位ビットをフラグとしてもよい。 ID is assigned to the corresponding block in the object map storage unit 26. In the above-described case, the object map storage unit 26 stores an object map of 60 × 80 blocks. Each block is provided with a flag indicating whether an ID is assigned, and when an ID is assigned, the number and a motion vector of a block described later are assigned as block information. Note that without using this flag, if ID = 0, it may be determined that no ID is assigned. The most significant bit of the ID may be used as a flag.
 移動物体追跡部27は、入口スリットを通過したクラスタに対して、移動方向のブロックに対するIDの付与及び移動と、反対方向のブロックに対するIDの消滅と、を行う(すなわち、クラスタの追跡処理)。移動物体追跡部27による追跡処理は、各クラスタに対して、出口スリット内まで行われる。 The moving object tracking unit 27 assigns and moves IDs for the blocks in the moving direction and disappears IDs for the blocks in the opposite direction with respect to the cluster that has passed through the entrance slit (that is, tracking process of the cluster). The tracking process by the moving object tracking unit 27 is performed up to the exit slit for each cluster.
 ID生成/消滅部25は、さらに、オブジェクトマップ記憶部26の内容に基づき、出口スリットEX1~EX4内のブロックにIDが付与されているかどうかを調べる。IDが付与されていれば、出口スリットをクラスタが通過したときに、そのIDを消滅させる。例えば、図3中の出口スリットEX1内のブロックにID=3が付されている状態から、IDが付されない状態に変化したときに、ID=3を消滅させる。消滅したIDは、次回に生成するIDとして用いることができる。 The ID generation / annihilation unit 25 further checks whether or not IDs are assigned to the blocks in the exit slits EX1 to EX4 based on the contents of the object map storage unit 26. If an ID is assigned, the ID is extinguished when the cluster passes through the exit slit. For example, when the block in the exit slit EX1 in FIG. 3 is changed from the state in which ID = 3 is assigned to the state in which no ID is assigned, ID = 3 is extinguished. The disappeared ID can be used as an ID to be generated next time.
 移動物体追跡部27は、オブジェクトマップ記憶部26に格納されている時刻t-1のオブジェクトマップと、画像メモリ21に格納されている時刻t-1及びtのフレーム画像とに基づいて、時刻tのオブジェクトマップを、オブジェクトマップ記憶部26内に作成する。以下、この動作を説明する。 Based on the object map at time t−1 stored in the object map storage unit 26 and the frame images at times t−1 and t stored in the image memory 21, the moving object tracking unit 27 performs time t−1. Are created in the object map storage unit 26. Hereinafter, this operation will be described.
 図4A~図7Bは、いずれも、時刻t-1及びtの画像を模式的に示す。図4A、図4B、および図6A~図7B中の点線は、ブロックの境界線である。図5Aおよび図5B中の点線は、画素の境界線である。 4A to 7B schematically show images at times t-1 and t. The dotted lines in FIGS. 4A, 4B, and 6A to 7B are block boundaries. The dotted lines in FIGS. 5A and 5B are pixel boundaries.
 第i行第j列のブロックをB(i,j)と表記し、時刻tでの第i行第j列のブロックをB(t:i,j)と表記する。ブロックB(t-1:1,4)の動きベクトルをMVとする。まず、ブロックB(t-1:1,4)をMVによって移動させた領域に最も対応する、時刻tのブロックを見つける。図4Bの場合、このブロックは、B(t:1,5)である。図5Aおよび図5Bに示すように、ブロックB(t:1,5)の画像と、時刻t-1のブロックサイズの領域AXの画像との相関度が、所定範囲AM内で領域AXを1画素移動させる毎に求められる(すなわち、ブロックマッチング)。 The block in the i-th row and the j-th column is denoted as B (i, j), and the block in the i-th row and the j-th column at the time t is denoted as B (t: i, j). Let MV be the motion vector of block B (t−1: 1, 4). First, the block at time t that most corresponds to the area where block B (t−1: 1, 4) has been moved by MV is found. In the case of FIG. 4B, this block is B (t: 1, 5). As shown in FIGS. 5A and 5B, the degree of correlation between the image of the block B (t: 1, 5) and the image of the block size area AX at time t−1 is 1 in the area AX within the predetermined range AM. It is obtained every time the pixel is moved (that is, block matching).
 ここで、範囲AMはブロックよりも大きく、その一辺は、ブロックの一辺の画素数の1.5倍である。範囲AMの中心は、ブロックB(t:1,5)の中心をMVによって移動させた位置にある画素である。 Here, the range AM is larger than the block, and one side thereof is 1.5 times the number of pixels on one side of the block. The center of the range AM is a pixel at a position where the center of the block B (t: 1, 5) is moved by MV.
 相関度は、時空的テクスチャ相関度であり、ブロックB(t:1,5)と領域AXの対応する画素値との差の絶対値の総和である評価値UDが小さいほど、大きい。 The correlation degree is a spatio-temporal texture correlation degree, and the larger the evaluation value UD that is the sum of absolute values of the difference between the block B (t: 1, 5) and the corresponding pixel value in the region AX, the larger the correlation degree.
 次に、範囲AM内で相関度が最大になる領域AXが求められる。この領域の中心を始点としブロックB(1,5)の中心を終点とするベクトルが、ブロックB(t:1,5)の動きベクトルと決定される。また、相関度が最大になる領域AXに最も近い、時刻t-1のブロックのIDを、ブロックB(t:1,5)のIDと決定する。 Next, an area AX having a maximum correlation within the range AM is obtained. A vector starting from the center of this area and ending at the center of block B (1, 5) is determined as the motion vector of block B (t: 1, 5). Further, the ID of the block at time t−1 that is closest to the region AX where the degree of correlation is maximum is determined as the ID of the block B (t: 1, 5).
 移動物体追跡部27は、隣り合うブロックの動きベクトルの差の絶対値が所定値以下のブロックに同一のIDを付与する。これにより、1つのクラスタが、互いに異なるIDをもつ複数のオブジェクト(移動物体)に分割される。図6Aおよび図6Bでは、オブジェクト間の境界を太線で示している。 The moving object tracking unit 27 assigns the same ID to blocks whose absolute values of motion vector differences between adjacent blocks are equal to or less than a predetermined value. Thereby, one cluster is divided into a plurality of objects (moving objects) having different IDs. In FIG. 6A and FIG. 6B, the boundary between objects is shown by the thick line.
 オブジェクトマップ上には移動物体の画像が存在しないが、図6Aおよび図6Bでは、理解を容易にするために、オブジェクトマップ上に移動物体が模式的に描かれている。図7Aおよび図7Bは、オブジェクトマップにオブジェクトの境界を太線で示したものであり、図6Aおよび図6Bに対応している。 Although the image of the moving object does not exist on the object map, in FIG. 6A and FIG. 6B, the moving object is schematically drawn on the object map for easy understanding. FIG. 7A and FIG. 7B show object boundaries in bold lines on the object map, and correspond to FIG. 6A and FIG. 6B.
 図3の入口スリットEN1で1つのクラスタが検出され、複数のオブジェクトに分割されず、その後、時刻t1に上記のようにして複数のオブジェクトに分割された場合、時刻t1から時間を遡って、時間が正方向の場合と同様に、オブジェクトマップを求める。これにより、時刻t1よりも前のオブジェクトマップに対し、オブジェクトを複数のオブジェクトに分割する処理を行う。これにより、分割できなかったオブジェクトを分割して認識でき、個々のオブジェクトを追跡できる。 When one cluster is detected at the entrance slit EN1 of FIG. 3 and is not divided into a plurality of objects, and then divided into a plurality of objects as described above at time t1, the time is traced back from time t1. The object map is obtained in the same way as when is in the positive direction. Thereby, the process which divides | segments an object into a some object is performed with respect to the object map before time t1. Thereby, an object that could not be divided can be divided and recognized, and individual objects can be tracked.
 上記特許文献1では、1つのクラスタが複数のクラスタに分離してから時間を遡って個々のオブジェクトを追跡していたが、本実施形態によれば、複数のクラスタに分離する前から、例えば図31A~図31Dのt=4より前のt=2から、時間を遡って個々のオブジェクトを追跡できる。そのため、画像メモリ21の記憶容量を低減でき、また、画像処理量を低減してCPUの負担を軽くできる。 In the above-mentioned Patent Document 1, each object is traced back in time after one cluster is separated into a plurality of clusters. However, according to the present embodiment, before separation into a plurality of clusters, for example, FIG. From t = 2 in FIG. 31A to FIG. 31D before t = 4, individual objects can be traced back in time. Therefore, the storage capacity of the image memory 21 can be reduced, and the amount of image processing can be reduced to reduce the load on the CPU.
 上記説明では、クラスタ内のブロックの動きベクトルが求まる場合について説明したが、図9Aに示すように、動きベクトルが求まらないブロックが存在する場合、その位置によってはこのブロックがどちらのオブジェクトに属するのかが不明である場合がある。ある移動物体に属するブロック内の各画素の色がほぼ同一であった場合、上述のブロックマッチングにより動きベクトルを決定できない。例えば、画像(空間的差分フレーム画像)を2値画像に変換し、ブロック内に‘1’の画素の数が所定値以下であれば、上記方法により動きベクトルを求めるのに適しないブロックであると判定される。 In the above description, the case where the motion vector of the block in the cluster is obtained has been described. However, as shown in FIG. 9A, when there is a block for which the motion vector cannot be obtained, depending on the position of the block, It may be unknown if it belongs. If the color of each pixel in a block belonging to a certain moving object is almost the same, the motion vector cannot be determined by the block matching described above. For example, if an image (spatial difference frame image) is converted into a binary image and the number of “1” pixels in the block is equal to or smaller than a predetermined value, the block is not suitable for obtaining a motion vector by the above method. It is determined.
 このようなブロックの動きベクトルを、図8に示す方法で推定する。 The motion vector of such a block is estimated by the method shown in FIG.
 (ステップS1)
 動きベクトルが未定であるブロックが、存在すればステップS2へ進み、存在しなければ未定動きベクトル推定処理を終了する。
(Step S1)
If there is a block whose motion vector is undetermined, the process proceeds to step S2, and if it does not exist, the undetermined motion vector estimation process is terminated.
 (ステップS2)
 動きベクトルが未定であるブロックB(i,j)の回りの8個のブロックのうちから、決定されている動きベクトルMV1~MVnを取り出す。
(Step S2)
The determined motion vectors MV1 to MVn are extracted from the eight blocks around the block B (i, j) whose motion vectors are undetermined.
 (ステップS3)
 ステップS2で決定済の動きベクトルが、存在すればステップS4へ進み、存在しなければステップS6へ進む。
(Step S3)
If the motion vector determined in step S2 exists, the process proceeds to step S4, and if not, the process proceeds to step S6.
 (ステップS4)
 動きベクトルMV1~MVnを、ベクトル間の差の絶対値が所定値以内のグループに分ける。
(Step S4)
The motion vectors MV1 to MVn are divided into groups in which the absolute value of the difference between the vectors is within a predetermined value.
 (ステップS5)
 動きベクトル数が最大のグループの動きベクトルの平均値を、ブロックB(i,j)の動きベクトルと推定する。動きベクトル数が最大のグループが複数存在する場合、任意の1つのグループの動きベクトルの平均値を、ブロックB(i,j)の動きベクトルと推定する。次にステップS1へ戻る。
(Step S5)
The average value of the motion vectors of the group having the largest number of motion vectors is estimated as the motion vector of the block B (i, j). When there are a plurality of groups having the largest number of motion vectors, the average value of the motion vectors of any one group is estimated as the motion vector of the block B (i, j). Next, the process returns to step S1.
 なお、同一グループの動きベクトルは互いに略等しいので、この同一グループの動きベクトルの任意の1つをブロックB(i,j)の動きベクトルと推定してもよい。 Note that since the motion vectors of the same group are substantially equal to each other, any one of the motion vectors of the same group may be estimated as the motion vector of the block B (i, j).
 (ステップS6)
 ステップS5で推定された動きベクトルを、決定された動きベクトルとし、ステップS1へ戻る。
(Step S6)
The motion vector estimated in step S5 is set as the determined motion vector, and the process returns to step S1.
 このような処理により、未定の動きベクトルを一意的に推定できる。 This process makes it possible to uniquely estimate an undetermined motion vector.
 次に、この推定方法の具体例を説明する。図9Aにおいて、第i行第j列のブロックB(i,j)の動きベクトルをMV(i,j)と表記する。図9Aでは、ブロックB(2,2)、B(2,4)及びB(3,3)の動きベクトルが未定である。 Next, a specific example of this estimation method will be described. In FIG. 9A, the motion vector of the block B (i, j) in the i-th row and j-th column is denoted as MV (i, j). In FIG. 9A, the motion vectors of blocks B (2,2), B (2,4) and B (3,3) are undetermined.
 ブロックB(2,2)の回りのブロックの動きベクトルは、MV(2,1)、MV(3,1)、MV(3,2)及びMV(2,3)のグループと、MV(1,2)及びMV(1,3)のグループに分けられる。このため、前者のグループを選択し、
 MV(2,2)=(MV(2,1)+MV(3,1)+MV(3,2)+MV(2,3))/4
と推定する。
The motion vectors of the blocks around the block B (2, 2) are the group of MV (2, 1), MV (3, 1), MV (3, 2) and MV (2, 3), and MV (1 , 2) and MV (1, 3). Therefore, select the former group,
MV (2,2) = (MV (2,1) + MV (3,1) + MV (3,2) + MV (2,3)) / 4
Estimated.
 ブロックB(2,4)の回りのブロックの動きベクトルは、MV(2,3)、MV(3,4)及びMV(3,5)のグループと、MV(1,3)、MV(1,4)、MV(1,5)及びMV(2,5)のグループに分けられる。このため、後者のグループを選択し、
 MV(2,4)=(MV(1,3)+MV(1,4)+MV(1,5)+MV(2,5))/4
と推定する。
The motion vectors of the blocks around the block B (2, 4) are a group of MV (2, 3), MV (3, 4) and MV (3, 5), and MV (1, 3), MV (1 4), MV (1, 5) and MV (2, 5). So select the latter group,
MV (2,4) = (MV (1,3) + MV (1,4) + MV (1,5) + MV (2,5)) / 4
Estimated.
 ブロックB(3,3)の回りのブロックの動きベクトルは、MV(2,3)、MV(3,2)、MV(4,2)、MV(4,4)及びMV(3,4)の1グループであるので、
 MV(3,3)=(MV(2,3)+MV(3,2)+MV(4,2)+MV(4,4)+MV(3,4))/5
と推定する。
The block motion vectors around block B (3, 3) are MV (2, 3), MV (3, 2), MV (4, 2), MV (4, 4) and MV (3,4). Because it is one group of
MV (3,3) = (MV (2,3) + MV (3,2) + MV (4,2) + MV (4,4) + MV (3,4))) / 5
Estimated.
 このようにして、図9Bに示すようなオブジェクトマップが生成される。図9Bでは、オブジェクトの境界を太線で示している。 In this way, an object map as shown in FIG. 9B is generated. In FIG. 9B, the boundary of the object is indicated by a thick line.
 図10Aのように未定動きベクトルの数が多い場合であっても、ステップS3で否定判定されるまで、ステップS1~S5を繰り返す。その結果、一意的に動きベクトルが推定されて、図10Bのようになる。次に、ステップS6で推定動きベクトルを、決定された動きベクトルとみなして、再度ステップS1~S5を実行する。これにより、ブロックB(3,4)の動きベクトルが一意的に推定されて、図10Cのようになる。次に、隣り合うブロックの動きベクトルの差の絶対値が所定値以下のブロックに同一のIDを付与することにより、1つのクラスタが、互いに異なるIDをもつ複数のオブジェクトに分割される。 Even if the number of undetermined motion vectors is large as shown in FIG. 10A, steps S1 to S5 are repeated until a negative determination is made in step S3. As a result, the motion vector is uniquely estimated as shown in FIG. 10B. Next, in step S6, the estimated motion vector is regarded as the determined motion vector, and steps S1 to S5 are executed again. Thereby, the motion vector of the block B (3, 4) is uniquely estimated as shown in FIG. 10C. Next, by assigning the same ID to blocks whose absolute values of motion vector differences between adjacent blocks are equal to or less than a predetermined value, one cluster is divided into a plurality of objects having different IDs.
 なお、移動物体追跡部27は、オブジェクトマップ記憶部26に格納されているオブジェクトマップの時系列を、追跡結果としてハードディスク(不図示)に格納する。 The moving object tracking unit 27 stores the time series of the object map stored in the object map storage unit 26 in a hard disk (not shown) as a tracking result.
 上述したような処理により、オブジェクトマップ記憶部26には、時系列画像の各画像が複数のブロックに分割され、当該画像中の移動物体を示す識別符号が移動物体に対応するブロックに付けられているとともに、ブロックに対応する移動物体の動きベクトルがブロックに付けられて記憶されている。 Through the processing described above, each image of the time-series image is divided into a plurality of blocks in the object map storage unit 26, and an identification code indicating the moving object in the image is attached to the block corresponding to the moving object. In addition, the motion vector of the moving object corresponding to the block is attached to the block and stored.
 そして、移動物体追跡部27は、上述したように時系列画像を画像処理した結果に基づいて、オブジェクトマップ記憶部26に記憶されているブロックの識別符号と動きベクトルとを更新する。具体的には、移動物体追跡部27は、次の(1)から(4)の手順により、オブジェクトマップ記憶部26に記憶されているブロックの識別符号と動きベクトルとを更新する。 Then, the moving object tracking unit 27 updates the block identification code and the motion vector stored in the object map storage unit 26 based on the result of image processing of the time-series image as described above. Specifically, the moving object tracking unit 27 updates the block identification code and the motion vector stored in the object map storage unit 26 according to the following procedures (1) to (4).
(1)時系列画像のうちの連続するN画像(N≧2)の各々について、隣り合うブロックの動きベクトルの差の絶対値が所定値以内のブロックに同一の識別符号を付けることにより、画像上で互いに重なった移動物体に互いに異なる識別符号を付ける識別符号手順。 (1) For each of consecutive N images (N ≧ 2) of time-series images, the same identification code is attached to a block in which the absolute value of the motion vector difference between adjacent blocks is within a predetermined value. An identification code procedure for attaching different identification codes to moving objects that overlap each other.
(2)N画像の各々において、第1識別符号が付けられたブロック群である第1オブジェクトと第2識別符号が付けられたブロック群である第2オブジェクトとが接し、かつ、N画像について時間的に隣り合う画像の第1オブジェクト間の相関度が所定値以上であるか否かを判定する判定手順。 (2) In each of the N images, a first object that is a block group to which a first identification code is attached is in contact with a second object that is a block group to which a second identification code is attached, and the time for the N image A determination procedure for determining whether or not the degree of correlation between first objects of adjacent images is equal to or greater than a predetermined value.
(3)判定手順で肯定と判定された後に、時間を遡って第1オブジェクトと第2オブジェクトとを追跡する追跡手順。 (3) A tracking procedure in which the first object and the second object are traced back in time after it is determined affirmative in the determination procedure.
(4)追跡手順により時間を遡って追跡された第1オブジェクトと第2オブジェクトとに基づいて、オブジェクトマップ記憶部26に記憶されているブロックの識別符号と動きベクトルとを更新する更新手順。 (4) An update procedure for updating the block identification code and the motion vector stored in the object map storage unit 26 based on the first object and the second object tracked back in time by the tracking procedure.
 輪郭抽出部30は、時系列画像の画像から移動物体の輪郭を抽出する。また、この輪郭抽出部30は、オクルージョンが発生している複数の移動物体を一体の移動物体とし、時系列画像の画像から一体とした移動物体の輪郭を抽出する。 The contour extraction unit 30 extracts the contour of the moving object from the time-series image. The contour extracting unit 30 extracts a plurality of moving objects having occlusions as a single moving object, and extracts a contour of the moving object as a single unit from a time-series image.
 この輪郭抽出部30は、対象領域設定部301と、対象領域補正部302と、輪郭抽出処理部303とを、有している。 The contour extraction unit 30 includes a target region setting unit 301, a target region correction unit 302, and a contour extraction processing unit 303.
 対象領域設定部301は、オブジェクトマップ記憶部26に記憶されている移動物体のブロックに対応する画像領域に基づいて、移動物体の輪郭を抽出する対象領域を設定する。 The target area setting unit 301 sets a target area for extracting the outline of the moving object based on the image area corresponding to the block of the moving object stored in the object map storage unit 26.
 輪郭抽出処理部303は、対象領域設定部301が設定した対象領域に対して、時系列画像の画像から移動物体の輪郭を抽出する。 The contour extraction processing unit 303 extracts the contour of the moving object from the time-series image for the target region set by the target region setting unit 301.
 対象領域補正部302は、時系列画像の画像をエッジ抽出処理した画像に対して、対象領域設定部301が設定した対象領域内でエッジに対応する画素数の個数についてのヒストグラムを座標軸毎に射影して生成する。この座標軸毎に生成したヒストグラムに基づいて、対象領域設定部301が設定した対象領域が座標軸毎に補正される。 The target area correction unit 302 projects, for each coordinate axis, a histogram of the number of pixels corresponding to the edges in the target area set by the target area setting unit 301 with respect to an image obtained by performing edge extraction processing on a time-series image. And generate. Based on the histogram generated for each coordinate axis, the target area set by the target area setting unit 301 is corrected for each coordinate axis.
 そして、上述した輪郭抽出処理部303は、対象領域補正部302が補正した対象領域に対して、時系列画像の画像から移動物体の輪郭を抽出してもよい。 Then, the contour extraction processing unit 303 described above may extract the contour of the moving object from the time-series image with respect to the target region corrected by the target region correction unit 302.
 補正部31は、輪郭抽出部30の輪郭抽出処理部303が抽出した輪郭に基づいて、オブジェクトマップ記憶部26に記憶されているブロックの識別符号および動きベクトルを補正する。 The correction unit 31 corrects the block identification code and the motion vector stored in the object map storage unit 26 based on the contour extracted by the contour extraction processing unit 303 of the contour extraction unit 30.
 また、この補正部31は、オブジェクトマップ記憶部26に記憶されているオクルージョンが発生している複数の移動物体に対応する識別符号に基づいたこの複数の移動物体の境界を示す情報と、輪郭抽出部30によりオクルージョンが発生している複数の移動物体を一体とした抽出された移動物体の輪郭とに基づいて、オブジェクトマップ記憶部26に記憶されているブロックの識別符号および動きベクトルを補正する。 The correction unit 31 also includes information indicating the boundaries of the plurality of moving objects based on the identification codes corresponding to the plurality of moving objects in which occlusion occurs and stored in the object map storage unit 26, and contour extraction. The block identification code and the motion vector stored in the object map storage unit 26 are corrected based on the extracted outline of the moving object obtained by integrating the plurality of moving objects in which occlusion has occurred.
 判定部32は、背景画像が変動しているか否かを判定する。判定部32は、カメラ10がパンニングまたはズーミングされたことを示す入力信号に基づいて、背景画像が変動しているか否かを判定する。 The determination unit 32 determines whether or not the background image is changing. The determination unit 32 determines whether or not the background image has changed based on an input signal indicating that the camera 10 has been panned or zoomed.
 判定部32は、背景画像生成部24が生成した背景画像と、カメラ10から入力される画像とを比較して、背景画像が変動しているか否かを判定してもよい。背景となる領域に予めマーカが埋め込まれており、判定部32は、カメラ10から入力される画像に含まれているマーカの画像における位置の変動を検出することにより、背景画像が変動しているか否かを判定してもよい。 The determination unit 32 may compare the background image generated by the background image generation unit 24 with the image input from the camera 10 and determine whether or not the background image is fluctuating. Whether or not the background image is fluctuated by detecting a change in the position of the marker in the image included in the image input from the camera 10 by the marker being embedded in the background region in advance. It may be determined whether or not.
 制御部34(第1の制御部)は、判定部32が背景画像が変動していると判定した場合に、予め定められている所定期間毎にまたは所定フレーム毎に、輪郭抽出部30を制御して輪郭を抽出させ、補正部31を制御してオブジェクトマップ記憶部26に記憶されているブロックの識別符号および動きベクトルを補正させる。 When the determination unit 32 determines that the background image is fluctuating, the control unit 34 (first control unit) controls the contour extraction unit 30 every predetermined period or every predetermined frame. Then, the contour is extracted, and the correction unit 31 is controlled to correct the block identification code and the motion vector stored in the object map storage unit 26.
 移動物体変動量検出部33については、後述する。 The moving object fluctuation amount detection unit 33 will be described later.
<背景ブロック>
 上記の説明においては、ブロック単位で入力画像を背景画像と比較することにより、オブジェクトが存在するかどうかを調べている。このため、背景画像とオブジェクトとの処理方法が異なる。また、例えば、過去10分間の撮影画像に基づいて背景画像を生成しているので、カメラが揺れた場合には、この揺れを背景画像に反映できない。
<Background block>
In the above description, whether or not an object exists is checked by comparing the input image with the background image in units of blocks. For this reason, the processing method of a background image and an object differs. For example, since the background image is generated based on the captured images for the past 10 minutes, when the camera shakes, the shake cannot be reflected in the background image.
 そこで、背景画像をオブジェクトとみなして、オブジェクトマップを作成してもよい。オブジェクトマップ生成方法は、背景画像と比較してブロックに移動物体が存在するかどうかを判定する点のみが異なる。背景画像もオブジェクトとみなすので、全てのブロックについて、ブロックマッチングを行うことによりIDを付与しMVを決定する。 Therefore, an object map may be created by regarding the background image as an object. The object map generation method is different from the background image only in determining whether or not a moving object exists in the block. Since the background image is also regarded as an object, block matching is performed on all the blocks to assign IDs and determine MVs.
 なお、この背景画像には、背景画像に対して予め定められているIDを付与してもよい。この予め定められているIDにより背景画像と、移動物体とが識別しやすくなる。 Note that an ID predetermined for the background image may be given to the background image. The background ID and the moving object can be easily identified by this predetermined ID.
 このように、背景画像を1つのブロックとして、背景画像に対してIDを付与するようにしても、背景画像と移動物体との間で、図4Aから図7Bに示したように、ブロックの属する画像を判定できる。 As described above, even if the background image is set as one block and an ID is assigned to the background image, the block belongs between the background image and the moving object as shown in FIGS. 4A to 7B. The image can be determined.
 このように、背景画像を1つのブロックとすることにより、カメラがパンニングまたはズーミングされたことに応じて背景画像が変動する場合においても、背景画像が固定されていた場合と同様に処理できる。 As described above, by setting the background image as one block, even when the background image fluctuates in response to panning or zooming of the camera, processing can be performed in the same manner as when the background image is fixed.
<移動物体追跡装置20の動作>
 次に、図11と図12とを用いて、移動物体追跡装置20の動作を説明する。図11を用いて、移動物体追跡装置20の動作モードを説明する。
<Operation of Moving Object Tracking Device 20>
Next, the operation of the moving object tracking device 20 will be described with reference to FIGS. 11 and 12. The operation mode of the moving object tracking device 20 will be described with reference to FIG.
 まず、移動物体追跡装置20は、カメラ固定モードの状態にある。このカメラ固定モードにおいては、背景画像はオブジェクトとしてのIDが付与されることなく、背景画像生成部24により生成されている画像を背景画像として用いている。これは、カメラが固定されているため、背景画像に変動がないためである。 First, the moving object tracking device 20 is in the camera fixed mode. In this camera fixing mode, the background image is not given an ID as an object, and the image generated by the background image generation unit 24 is used as the background image. This is because the background image does not change because the camera is fixed.
 このカメラ固定モードの状態にある場合には、背景画像が固定されているため、移動物体追跡部27により、移動物体が識別されている。 In this camera fixing mode, since the background image is fixed, the moving object tracking unit 27 identifies the moving object.
 (ステップS1)
 次に、カメラがパンニングまたはズーミングされたことに応じて、判定部32が、背景画像が変動していることを判定する。この判定に応じて制御部34が、カメラ固定モードから、カメラ変動モードに遷移させるとともに、背景画像をオブジェクトに登録してIDを付与する。
(Step S1)
Next, in accordance with the panning or zooming of the camera, the determination unit 32 determines that the background image is fluctuating. In response to this determination, the control unit 34 shifts from the camera fixing mode to the camera variation mode, registers the background image in the object, and assigns an ID.
 (ステップS2)
 カメラ変動モードにおいて、予め定められている所定期間毎にまたは所定フレーム毎に、制御部34は、補正モードに遷移させる。この補正モードにおいて、制御部34は、輪郭抽出部30を制御して輪郭を抽出させ、補正部31を制御してオブジェクトマップ記憶部26に記憶されているブロックの識別符号および動きベクトルを補正させる。
(Step S2)
In the camera variation mode, the control unit 34 shifts to the correction mode every predetermined period or every predetermined frame. In this correction mode, the control unit 34 controls the contour extraction unit 30 to extract a contour, and controls the correction unit 31 to correct the block identification code and motion vector stored in the object map storage unit 26. .
 (ステップS3)
 制御部34は、補正モードにおいて補正が完了すると、補正モードからカメラ変動モードに遷移させる。
(Step S3)
When the correction is completed in the correction mode, the control unit 34 transitions from the correction mode to the camera fluctuation mode.
 その後、カメラ10が変動している期間においては、制御部34は、カメラ変動モードと補正モードとを交互に遷移させる。 Thereafter, during a period in which the camera 10 is changing, the control unit 34 alternately switches between the camera change mode and the correction mode.
 (ステップS4)
 その後、カメラ10の変動が停止したことを判定部32が判定すると、制御部34は、背景画像に付与されているIDを削除し、背景画像生成部24により生成されている画像を背景画像として用いる。
(Step S4)
Thereafter, when the determination unit 32 determines that the fluctuation of the camera 10 has stopped, the control unit 34 deletes the ID assigned to the background image, and uses the image generated by the background image generation unit 24 as the background image. Use.
 なお、背景画像生成部24により背景画像が生成されるには、たとえば10分などの所定の期間を要する。そのため、背景画像生成部24により背景画像が生成されるまでは、背景画像をオブジェクトとして用いることが望ましい。 It should be noted that a predetermined period such as 10 minutes is required for the background image generation unit 24 to generate the background image. Therefore, it is desirable to use the background image as an object until the background image is generated by the background image generation unit 24.
 次に図12を用いて、図11の補正モードにおいて、制御部34が、輪郭抽出部30を制御して輪郭を抽出させ、補正部31を制御してオブジェクトマップ記憶部26に記憶されているブロックの識別符号および動きベクトルを補正させる動作について説明する。 Next, referring to FIG. 12, in the correction mode of FIG. 11, the control unit 34 controls the contour extraction unit 30 to extract a contour, controls the correction unit 31, and is stored in the object map storage unit 26. The operation for correcting the block identification code and motion vector will be described.
 (ステップS1201)
 まず、オクルージョン検出部35が、オブジェクトマップ記憶部26に記憶されているオブジェクトマップに基づいて、オクルージョンの発生の有無を判定する。
(Step S1201)
First, the occlusion detection unit 35 determines whether or not occlusion has occurred based on the object map stored in the object map storage unit 26.
 (ステップS1202)
 このステップS1201でオクルージョンの発生が無いと判定された場合には、対象領域設定部301が、オクルージョンが無い移動物体に対して対象領域を設定する。
(Step S1202)
If it is determined in step S1201 that no occlusion has occurred, the target area setting unit 301 sets a target area for a moving object without occlusion.
 (ステップS1203)
 次に、対象領域補正部302が、上述したヒストグラムを生成し、生成したヒストグラムに基づいて対象領域設定部301が設定した対象領域を座標軸毎に補正する。
(Step S1203)
Next, the target area correction unit 302 generates the above-described histogram, and corrects the target area set by the target area setting unit 301 for each coordinate axis based on the generated histogram.
 (ステップS1204)
 次に輪郭抽出処理部303は、対象領域補正部302が補正した対象領域に対して、時系列画像の画像から移動物体の輪郭を抽出する。
(Step S1204)
Next, the contour extraction processing unit 303 extracts the contour of the moving object from the time-series image for the target region corrected by the target region correction unit 302.
 (ステップS1205)
 次に、補正部31は、輪郭抽出部30が抽出した輪郭に基づいて、オブジェクトマップ記憶部26に記憶されているブロックの識別符号および動きベクトルを補正する。
(Step S1205)
Next, the correction unit 31 corrects the block identification code and the motion vector stored in the object map storage unit 26 based on the contour extracted by the contour extraction unit 30.
 (ステップS1212)
 一方、ステップS1201でオクルージョンの発生があると判定された場合には、輪郭抽出部30の対象領域設定部301が、オクルージョンが発生している複数の移動物体を一体の移動物体とする。
(Step S1212)
On the other hand, if it is determined in step S1201 that occlusion has occurred, the target area setting unit 301 of the contour extraction unit 30 sets a plurality of moving objects in which occlusion has occurred as an integrated moving object.
 (ステップS1213)
 次に、対象領域設定部301が、一体とした移動物体に対して対象領域を設定する。
(Step S1213)
Next, the target area setting unit 301 sets a target area for an integrated moving object.
 (ステップS1214)
 次に、対象領域補正部302が、一体とした移動物体に対して上述したヒストグラムを生成し、生成したヒストグラムに基づいて対象領域設定部301が設定した対象領域を座標軸毎に補正する。
(Step S1214)
Next, the target area correction unit 302 generates the above-described histogram for the integrated moving object, and corrects the target area set by the target area setting unit 301 for each coordinate axis based on the generated histogram.
 (ステップS1215)
 次に、輪郭抽出処理部303は、対象領域補正部302が補正した対象領域に対して、時系列画像の画像から、一体とした移動物体移動物体の輪郭を抽出する。
(Step S1215)
Next, the contour extraction processing unit 303 extracts the contour of the moving object integrated with the target region corrected by the target region correction unit 302 from the image of the time series image.
 (ステップS1216)
 次に、補正部31は、図4Aから図7Bを用いて説明したように、検出されているオクルージョンが生じている移動物体の境界と、輪郭抽出部30が抽出した輪郭とに基づいて、オブジェクトマップ記憶部26に記憶されているブロックの識別符号および動きベクトルを補正する。
(Step S1216)
Next, as described with reference to FIG. 4A to FIG. 7B, the correction unit 31 performs object detection based on the boundary of the moving object in which the detected occlusion occurs and the contour extracted by the contour extraction unit 30. The block identification code and motion vector stored in the map storage unit 26 are corrected.
 上記に説明した処理が、オブジェクトマップ記憶部26に記憶されている全ての移動物体に対して終了した後、制御部34は、補正モードからカメラ変動モードへと遷移させる。 After the processing described above is completed for all moving objects stored in the object map storage unit 26, the control unit 34 transitions from the correction mode to the camera variation mode.
 以上、図11及び図12を用いて説明したように、移動物体追跡装置20は、カメラが固定している場合や変動している場合でも、移動物体を追跡できる。なお、カメラ固定モードでは背景画像をオブジェクトに登録していないことにより、図4Aから図7Bに示すような処理を背景画像のオブジェクトに対して実行する必要がないために、移動物体追跡装置20における処理量または負荷を軽減できる。 As described above with reference to FIGS. 11 and 12, the moving object tracking device 20 can track a moving object even when the camera is fixed or fluctuates. In the camera fixing mode, since the background image is not registered in the object, it is not necessary to execute the processing as illustrated in FIGS. 4A to 7B on the object of the background image. The amount of processing or load can be reduced.
 以降、オブジェクトマップ記憶部26に記憶されているブロック毎の識別符号および動きベクトルを、「時空間MRF(Markov Random Field)」と称して説明する。次に、図12を用いて説明した動作及びその結果について、図13から図30Bを用いて説明する。 Hereinafter, the identification code and motion vector for each block stored in the object map storage unit 26 will be described as “space-time MRF (Markov Random Field)”. Next, the operation and results described with reference to FIG. 12 will be described with reference to FIGS. 13 to 30B.
<Snakes>
 まず、輪郭抽出部30による画像から移動物体の輪郭の抽出について説明する。輪郭抽出部30の構成のうち、まず、輪郭抽出処理部303による移動物体の輪郭を抽出する技術の一例について詳述する。ここでは、輪郭を抽出する技術としてSnakes(下記非特許文献1参照)を用いる場合について説明する。
<Snakes>
First, extraction of the contour of the moving object from the image by the contour extraction unit 30 will be described. Of the configuration of the contour extraction unit 30, first, an example of a technique for extracting the contour of a moving object by the contour extraction processing unit 303 will be described in detail. Here, the case where Snakes (refer nonpatent literature 1 below) is used as a technique which extracts an outline is demonstrated.
(非特許文献1)Kass et.al “Snakes: Active contour models”, Proc. of 1st ICCV, pp.259-268, 1987 (Non-Patent Document 1) Kass et.al “Snakes: Active contour models”, Proc. Of 1st ICCV, pp.259-268, 1987
 最初に、このSnakesの概要について説明する。一般にSnakesは、画像平面(x,y)上で媒介変数表現されたスプライン(制御点の集合)v(s)=(x(s),y(s))(0≦s≦1)を、下記式(4)で定義されるエネルギー関数を最小化するように変形し、エネルギーの極小状態としてその形状が決まる輪郭抽出のモデルである。 First, the outline of Snakes will be described. In general, Snakes expresses a spline (set of control points) v (s) = (x (s), y (s)) (0 ≦ s ≦ 1) expressed as a parameter on the image plane (x, y), This is a contour extraction model that is deformed so as to minimize the energy function defined by the following equation (4) and whose shape is determined as a minimum state of energy.
Figure JPOXMLDOC01-appb-M000001
Figure JPOXMLDOC01-appb-M000001
 この式(4)の第一項Eintは、内部エネルギーである。これにより、Snakesのスプラインが、凸型に滑らかに収縮する性質をもつ。理論上の定義は、下記式(5)で表される。下記式(5)における第一項により、スプラインが凸型に滑らかになり、第二項によりスプラインが収縮する性質をもつ。 The first term Eint in this equation (4) is internal energy. Thus, the Snakes spline has a property of smoothly contracting into a convex shape. The theoretical definition is expressed by the following formula (5). The first term in the following formula (5) has the property that the spline becomes smooth in a convex shape, and the spline contracts by the second term.
Figure JPOXMLDOC01-appb-M000002
Figure JPOXMLDOC01-appb-M000002
 次に、上記式(4)の第二項Eimageは、画像エネルギーである。この画像エネルギーは、スプライン全体として、エッジ(輪郭などの輝度の勾配が大きい箇所)上に存在する割合が大きいほど、その値が小さくなるという性質をもつ。この画像エネルギーは、画像の輝度I(v(s))により下記式(6)で定義される。今回は、照度によらず安定して輪郭エッジを抽出するために、これまで本願発明者らが開発してきた照度不変フィルタ画像を画像エネルギーとして用いた。 Next, the second term Eimage of the above formula (4) is image energy. This image energy has a property that the value of the image energy becomes smaller as the ratio of the image energy existing on the edge (a portion such as a contour having a large luminance gradient) increases. This image energy is defined by the following equation (6) by the luminance I (v (s)) of the image. This time, in order to stably extract the contour edge regardless of the illuminance, the illuminance invariant filter image developed so far by the present inventors was used as the image energy.
Figure JPOXMLDOC01-appb-M000003
Figure JPOXMLDOC01-appb-M000003
 ここで、照度不変フィルタ画像とは、画像変換部22により上記式(1)~(3)により変換された画像である。 Here, the illuminance invariant filter image is an image converted by the image conversion unit 22 according to the above formulas (1) to (3).
 そして、式(4)の第三項Econは、外部エネルギーである。この外部エネルギーは、Snakesに外部から強制的に力を働かせる場合に用いる。このエネルギーは、必要に応じて定義される。今回は、内部エネルギーに用いられるEintの影響で抽出が困難であった凹形状の輪郭を抽出するために提案された面積項(下記非特許文献2参照)を、外部エネルギーとして定義した。面積項Eareaは、下記式(7)により導出される(図13参照)。 And the third term Econ in the equation (4) is external energy. This external energy is used when force is forcedly applied to Snakes from the outside. This energy is defined as needed. This time, an area term (see Non-Patent Document 2 below) proposed for extracting a concave contour that was difficult to extract due to the influence of Eint used for internal energy was defined as external energy. The area term Area is derived by the following equation (7) (see FIG. 13).
Figure JPOXMLDOC01-appb-M000004
Figure JPOXMLDOC01-appb-M000004
(非特許文献2)荒木昭一, 横矢直和, 岩佐英彦, 竹村治雄:“複数物体の抽出を目的とした交差判定により分裂する動的輪郭モデル”, 電子情報通信学会論文誌(D-II)Vol.J79-DII, No.10,pp1704-1711(Oct,1996) (Non-patent literature 2) Shoichi Araki, Naokazu Yokoya, Hidehiko Sasaiwa, Haruo Sasatakemura: “Dynamic contour model splitting by intersection judgment for the purpose of extracting multiple objects”, Journal of the Institute of Electronics, Information and Communication Engineers (D-II) Vol.J79-DII, No.10, pp1704-1711 (Oct, 1996)
 図14A~図14Cに、Snakesによる処理結果を示す。まず、輪郭を抽出したい対象の周囲に、初期制御点を配置する(図14A)。次に、Snakesは収縮を始める(図14B)。最後に、輪郭線付近で収縮が止まる(図14C)。 14A to 14C show the processing results by Snakes. First, initial control points are arranged around the object whose contour is to be extracted (FIG. 14A). Next, Snakes begins to contract (FIG. 14B). Finally, the contraction stops near the contour line (FIG. 14C).
 ここで、この図14Aに示す初期制御点が、対象領域設定部301が設定した対象領域に対応する。 Here, the initial control points shown in FIG. 14A correspond to the target area set by the target area setting unit 301.
<局所領域におけるエッジ分布の解析>
 次に、局所領域におけるエッジ分布の解析について説明する。Snakesは,エネルギー関数Esnakesを最小化するようにスプラインを変形し,極小状態に至ったときにその探索を終えるアルゴリズムである。背景エッジが多く含まれる場合に、初期制御点を輪郭から離れた場所に配置すると、スプラインがオブジェクトの輪郭線に収束する前に背景エッジに捉われ、エネルギーが極小状態になる。この結果、オブジェクトの輪郭線抽出に失敗してしまうことがある。
<Analysis of edge distribution in local region>
Next, analysis of edge distribution in the local region will be described. Snakes is an algorithm that finishes the search when the spline is deformed to minimize the energy function E_snakes and reaches a minimum state. If there are many background edges and the initial control points are placed away from the contour, the spline will be caught by the background edges before converging on the contour of the object, and the energy will be minimal. As a result, the outline extraction of the object may fail.
 そのため、初期制御点をある程度オブジェクトの輪郭付近に配置しなければならない。一方、時空間MRFで得られる人物の局所領域では、パンニング中に背景オブジェクトと人物オブジェクトとの境界が曖昧になる場合がある。 Therefore, some initial control points must be placed near the contour of the object. On the other hand, in the local region of the person obtained by the spatiotemporal MRF, the boundary between the background object and the person object may become ambiguous during panning.
 そこで、オブジェクトマップによって得られる人物オブジェクトの周辺矩形内におけるエッジ分布(照度不変フィルタ画像の2値分布)を解析し、人物領域を推定する。これにより、その周囲に初期制御点を配置することで、輪郭抽出の精度を向上させる。エッジ分布の解析は、水平軸、鉛直軸にそれぞれ射影し、ヒストグラムを生成することにより解析を行う(図15A~図15C参照)。 Therefore, the edge distribution (binary distribution of the illuminance invariant filter image) in the surrounding rectangle of the human object obtained by the object map is analyzed, and the human area is estimated. Thereby, the accuracy of contour extraction is improved by arranging the initial control points around it. The analysis of the edge distribution is performed by projecting onto the horizontal axis and the vertical axis and generating a histogram (see FIGS. 15A to 15C).
 次に、エッジ分布解析の手順の一例について説明する。エッジ分布解析は、次のSTEP1からSTEP3の手順により実行される。 Next, an example of an edge distribution analysis procedure will be described. The edge distribution analysis is executed by the following steps STEP1 to STEP3.
(STEP1:エッジ分布の前処理)
 エッジ画像をラベリングし、面積の小さいものをノイズとして除外する。
(STEP 1: Preprocessing of edge distribution)
The edge image is labeled, and a small area is excluded as noise.
(STEP2:水平軸ヒストグラム解析)
 STEP1で得られたエッジ分布から水平軸ヒストグラムを生成する。人物は縦に連続する長いエッジを持つので、縦方向の連続性の弱いエッジを除去した後、一度水平軸に正射影し、縦方向の強い分布から人物の水平領域を絞る。
(STEP 2: Horizontal axis histogram analysis)
A horizontal axis histogram is generated from the edge distribution obtained in STEP 1. Since a person has long vertical edges, after removing edges with low vertical continuity, the person is orthogonally projected once to narrow down the horizontal area of the person from a strong vertical distribution.
 その後、絞られた領域内において、得られたエッジ分布を水平軸に射影し、ヒストグラムを生成する。そして、得られた水平軸ヒストグラムにおいてウインドウをスキャンする。このように連続して分布する領域を求めることで、水平成分における人物エッジの分布領域の推定を行う(図15A)。 After that, within the narrowed area, the obtained edge distribution is projected on the horizontal axis to generate a histogram. Then, the window is scanned in the obtained horizontal axis histogram. Thus, by obtaining a continuously distributed region, the distribution region of the human edge in the horizontal component is estimated (FIG. 15A).
(STEP3:鉛直軸ヒストグラム解析)
 STEP1で得られたエッジ分布から鉛直軸ヒストグラムを生成する。人物は横方向にもある程度長い連続エッジを持つので、横方向の連続性の弱いエッジを除去した後、一度鉛直軸に射影し、横方向の分布から人物の鉛直領域を絞る。
(STEP3: Vertical axis histogram analysis)
A vertical axis histogram is generated from the edge distribution obtained in STEP 1. Since a person has a continuous edge that is somewhat long in the horizontal direction, after removing edges with low continuity in the horizontal direction, the person is projected once onto the vertical axis, and the vertical region of the person is narrowed down from the distribution in the horizontal direction.
 その後、絞られた領域内において、エッジ分布を鉛直軸に射影し、ヒストグラムを生成する。そして、得られた水平軸ヒストグラムにおいてウインドウをスキャンする。このように連続して分布する領域を求めることで、水平成分における人物エッジの分布領域の推定を行う(図15B)。 After that, within the narrowed area, the edge distribution is projected onto the vertical axis, and a histogram is generated. Then, the window is scanned in the obtained horizontal axis histogram. Thus, the distribution area of the person edge in the horizontal component is estimated by obtaining the continuously distributed area (FIG. 15B).
 なお、このSTEP3において、STEP1で得られたエッジ分布から鉛直軸ヒストグラムを生成している。これに代えて、鉛直軸ヒストグラムは、STEP2で狭められた局所領域内のエッジ分布から生成してもよい。 In STEP 3, a vertical axis histogram is generated from the edge distribution obtained in STEP 1. Alternatively, the vertical axis histogram may be generated from the edge distribution in the local region narrowed in STEP2.
 以上のステップにより、エッジの分布情報を参照することで、より正確な人物の外接矩形領域を得られる(図15C)。 By referring to the edge distribution information through the above steps, a more accurate circumscribed rectangular region of the person can be obtained (FIG. 15C).
 なお、ヒストグラム閾値(初期制御点)の設定方法として、次のようにしてもよい。
 まず、ヒストグラムの頻度値を2つのグループにクラスタリングする。クラスタリング方法は、k-mean法(1次元)など、いずれの方法でもよい。これにより、ヒストグラムの頻度値は、高頻度クラスタと低頻度クラスタとに分かれる。
Note that the histogram threshold (initial control point) may be set as follows.
First, the histogram frequency values are clustered into two groups. The clustering method may be any method such as a k-mean method (one-dimensional). Thereby, the frequency value of a histogram is divided into a high frequency cluster and a low frequency cluster.
 次に、画像の両端から内側へ探索して、初めて高頻度クラスタに属する頻度値に当たった場合、直前の低頻度の位置とこの高頻度の位置との間を境界とする。この場合、高頻度のさらに内側に低頻度のものがあってもよい。 Next, when searching from the both ends of the image to the inside and hitting the frequency value belonging to the high-frequency cluster for the first time, the boundary between the previous low-frequency position and this high-frequency position is set as a boundary. In this case, there may be a low frequency inside the high frequency.
<階層間協調アルゴリズム>
 ここで、時空間MRFとSnakesとの階層間協調によるトラッキングの処理ステップを述べる。以降、時空間MRFとSnakesとを相互に用いてオブジェクトマップを修正することを、階層間協調と称して説明する。
<Inter-layer cooperation algorithm>
Here, a processing step of tracking by cooperation between layers of the spatio-temporal MRF and Snakes will be described. Hereinafter, correcting the object map using the spatio-temporal MRF and Snakes will be referred to as inter-layer cooperation.
 まず、Snakesによるオブジェクトマップの修正(オクルージョンなし)の場合、すなわち、図12のステップS1202からS1205に対応する処理について説明する。 First, in the case of object map correction (without occlusion) by Snakes, that is, processing corresponding to steps S1202 to S1205 in FIG. 12 will be described.
 (STEP1)
 時空間MRFの出力として、オブジェクトマップを受け取り、各オブジェクトにおける外接矩形領域の情報を得る。
(STEP1)
As an output of the spatiotemporal MRF, an object map is received, and information on a circumscribed rectangular area in each object is obtained.
 (STEP2)
 各オブジェクトにおいて、STEP2で得られた局所領域においてエッジの分布解析を行い、オブジェクトの輪郭周辺にSnakesの初期制御点を配置する。
(STEP2)
For each object, edge distribution analysis is performed in the local region obtained in STEP 2, and Snakes initial control points are arranged around the contour of the object.
 (STEP3)
 各オブジェクトにおいて、Snakesを実行する。STEP1で得られた外接矩形の大きさと比較し、スプラインが収縮しすぎてしまったオブジェクトに対しては、オブジェクトマップを修正しない。それ以外のオブジェクトに対しては、Snakesの処理結果を反映し、オブジェクトマップを修正する。
(STEP3)
Execute Snakes on each object. Compared to the size of the circumscribed rectangle obtained in STEP 1, the object map is not corrected for an object whose spline has contracted too much. For other objects, the processing result of Snakes is reflected and the object map is corrected.
 次に、Snakesによるオブジェクトマップの修正(オクルージョンあり)の場合、すなわち、図12のステップS1212からS1215に対応する処理について説明する。 Next, in the case of object map correction by Snakes (with occlusion), that is, processing corresponding to steps S1212 to S1215 in FIG. 12 will be described.
 (STEP1)
 時空間MRFの出力として、オブジェクトマップを受け取り、各オブジェクトにおける外接矩形領域の情報を得る。オクルージョンが検知された(他のオブジェクトの外接矩形と重複領域をもつ)オブジェクトに対しては、オクルージョンし合っているものを一つのグループとして外接矩形領域を求める。
(STEP1)
As an output of the spatio-temporal MRF, an object map is received, and information on a circumscribed rectangular area in each object is obtained. For objects in which occlusion is detected (having overlapping areas with circumscribed rectangles of other objects), the circumscribed rectangular area is obtained with the objects that are occluded as one group.
 (STEP2)
 STEP2で得られた局所領域においてエッジの分布解析を行い、オブジェクトの輪郭周辺にSnakesの初期制御点を配置する。
(STEP2)
Edge distribution analysis is performed in the local region obtained in STEP 2, and Snakes initial control points are arranged around the contour of the object.
 (STEP3)
 Snakesを実行する。STEP1で得られた外接矩形の大きさと比較し、スプラインが収縮しすぎてしまったオブジェクトに対しては、オブジェクトマップを修正しない。それ以外のオブジェクトに対しては、Snakesの処理結果を反映し、オブジェクトマップを修正する。その際、Snakesで抽出された輪郭に囲まれた内部領域の各オブジェクトのID番号のラベリングにおいては、時空間MRFモデルの出力結果によりラベリングを行なう。ただし、この輪郭内部領域において、背景オブジェクトと認識されたブロックに対しては、ラベルを背景オブジェクトから不定へ変更する。これにより、次フレームにおいて時空間MRFモデルによるID割り当てが行われる。あるいは、現フレームにおいて、このブロックのみ時空間MRFモデルによるID割り当てを再度行なえる。
(STEP3)
Execute Snakes. Compared to the size of the circumscribed rectangle obtained in STEP 1, the object map is not corrected for an object whose spline has contracted too much. For other objects, the processing result of Snakes is reflected and the object map is corrected. At that time, in the labeling of the ID number of each object in the inner region surrounded by the contour extracted by Snakes, the labeling is performed based on the output result of the spatiotemporal MRF model. However, the label is changed from the background object to indefinite for the block recognized as the background object in this contour inner region. Thereby, ID allocation by the spatio-temporal MRF model is performed in the next frame. Or, in the current frame, only this block can be assigned ID again by the spatio-temporal MRF model.
 図16A~図17Bに、Snakesによるオブジェクトマップの修正例を示す。図16A及び図16Bはオクルージョンのない場合の例であり、図17A及び図17Bはオクルージョン時の例である。 FIG. 16A to FIG. 17B show examples of object map correction by Snakes. 16A and 16B are examples in the case where there is no occlusion, and FIGS. 17A and 17B are examples in the case of occlusion.
 図17A及び図17Bでは、まず、ID番号6及びID番号7の人物を一つのグループとし、Snakesによりグループの輪郭線(背景と人物との境界)を求める。そして、グループ内の領域分割は、時空間MRFによる出力情報を反映する。 17A and 17B, first, the persons with ID numbers 6 and 7 are grouped together, and the outline of the group (the boundary between the background and the person) is obtained by Snakes. The region division within the group reflects the output information by the spatiotemporal MRF.
<Snakesによるオブジェクトマップの修正>
 次に、図18A~図29Bを用いて、カメラ10が変動する場合における階層間協調アルゴリズムによる効果を示す。ここでは、図18A~図29Bに示すように、同じシーンの同じフレームに対して、階層間協調なしと階層間協調ありとの場合における、移動物体検出の処理結果を対比して説明する。
<Correction of object map by Snakes>
Next, with reference to FIGS. 18A to 29B, effects of the inter-layer cooperation algorithm when the camera 10 changes will be described. Here, as shown in FIGS. 18A to 29B, the processing results of moving object detection in the case of no inter-layer cooperation and inter-layer cooperation for the same frame of the same scene will be described.
 図18A~図23Aは階層間協調を行わない場合の処理結果であり、図18B~図23Bはそのオブジェクトマップである。図24A~図29Aは階層間協調を行う場合の処理結果であり、図24B~図29Bはそのオブジェクトマップである。 FIG. 18A to FIG. 23A show the processing results when inter-layer cooperation is not performed, and FIG. 18B to FIG. 23B are object maps thereof. FIG. 24A to FIG. 29A show the processing results when inter-layer cooperation is performed, and FIG. 24B to FIG. 29B are object maps thereof.
 カメラのパンニングは、フレーム番号80の直後から始まる。階層間協調アルゴリズムを行わない場合、20フレーム超のトラッキングに成功しているものの、人物オブジェクトと背景オブジェクトとの境界は、その徐々に曖昧になってしまっている(フレーム95およびフレーム107)。 カ メ ラ Camera panning starts immediately after frame number 80. When the inter-layer cooperation algorithm is not performed, although the tracking of more than 20 frames has been successful, the boundary between the person object and the background object has gradually become ambiguous (frame 95 and frame 107).
 一方、Snakesにより協調させた場合、時空間MRFの出力を参照することでオブジェクト間の境界を補正できており、長時間のトラッキングができる。また、ズーミングした場合においても、階層間協調アルゴリズムの効果がある(フレーム142およびフレーム153の右2列)。 On the other hand, when coordinated by Snakes, the boundary between objects can be corrected by referring to the output of the spatio-temporal MRF, and tracking can be performed for a long time. In addition, even when zooming, there is an effect of the cooperation algorithm between layers (the right two columns of the frame 142 and the frame 153).
 図30Aおよび図30Bは、オクルージョンが発生している場合に、カメラ10が変動した場合に移動物体を検出した結果である。図30Aでは、オクルージョンが発生していない。その後、図30Bでは、オクルージョンが発生している。いずれの場合においても、本実施形態による移動物体追跡装置20は、移動物体を追跡できている。 FIG. 30A and FIG. 30B are results of detecting a moving object when the camera 10 fluctuates when occlusion occurs. In FIG. 30A, no occlusion has occurred. Thereafter, in FIG. 30B, occlusion has occurred. In any case, the moving object tracking device 20 according to the present embodiment can track the moving object.
 以上説明したように、本実施形態による移動物体追跡装置20は、抽出した輪郭に基づいて、オブジェクトマップ記憶部に記憶されているブロックの識別符号および動きベクトルを補正することにより、背景が変動する画像からでも、画像中の移動物体を正確に検出できる。 As described above, the moving object tracking device 20 according to the present embodiment changes the background by correcting the block identification code and the motion vector stored in the object map storage unit based on the extracted contour. Even from an image, a moving object in the image can be accurately detected.
 なお、上記の実施形態の説明においては、オブジェクトマップ記憶部に記憶されているブロックの識別符号とともに動きベクトルを補正するものとして説明したが、識別符号のみを補正してもよい。このようにしても、同様に、背景が変動する画像からでも、画像中の移動物体を正確に検出できる。 In the description of the above embodiment, the motion vector is corrected together with the block identification code stored in the object map storage unit. However, only the identification code may be corrected. Even in this way, similarly, a moving object in an image can be accurately detected from an image whose background changes.
 なお、上述した移動物体変動量検出部33は、オブジェクトマップ記憶部26に記憶されている移動物体のサイズまたは移動量の変動量を、識別符号または動きベクトルに基づいて単位時間毎に検出する。 The moving object fluctuation amount detection unit 33 described above detects the size of the moving object or the fluctuation amount of the movement amount stored in the object map storage unit 26 based on the identification code or the motion vector for each unit time.
 そして、制御部34(第2の制御部)は、移動物体変動量検出部33により検出された単位時間毎の移動物体のサイズまたは移動量の変動量が、予め定められているサイズまたは移動量の変動量よりも大きい場合に、輪郭抽出部30を制御して輪郭を抽出させ、補正部31を制御してオブジェクトマップ記憶部26に記憶されているブロックの識別符号および動きベクトルを補正させる。 Then, the control unit 34 (second control unit) is configured such that the size of the moving object or the amount of movement of the moving object per unit time detected by the moving object fluctuation amount detection unit 33 is a predetermined size or movement amount. Is larger than the fluctuation amount, the contour extraction unit 30 is controlled to extract the contour, and the correction unit 31 is controlled to correct the block identification code and motion vector stored in the object map storage unit 26.
 このように、オブジェクトマップ記憶部26に記憶されている移動物体のサイズまたは移動量の変動量も応じて、オブジェクトマップ記憶部26に記憶されているブロックの識別符号および動きベクトルを補正させてもよい。 As described above, the block identification code and the motion vector stored in the object map storage unit 26 may be corrected in accordance with the size of the moving object stored in the object map storage unit 26 or the fluctuation amount of the movement amount. Good.
 これにより、制御部34が、単に予め定められている所定期間毎にまたは所定フレーム毎にオブジェクトマップを補正する場合に対比して、移動物体を検出することに失敗しそうなタイミングで、オブジェクトマップを補正できる。そのため、移動物体をより的確に検出して追跡できる。 As a result, the control unit 34 compares the object map at a timing at which it is likely to fail to detect a moving object, as compared with the case where the control unit 34 simply corrects the object map every predetermined period or every predetermined frame. Can be corrected. Therefore, a moving object can be detected and tracked more accurately.
 なお、図2における、フレームバッファメモリ3、画像メモリ21、またはオブジェクトマップ記憶部26などの記憶部は、ハードディスク装置、光磁気ディスク装置、フラッシュメモリ等の不揮発性のメモリ、CD-ROM等の読み出しのみが可能な記憶媒体、RAM(Random Access Memory)のような揮発性のメモリ、またはこれらの組み合わせにより構成されていてもよい。 Note that the storage unit such as the frame buffer memory 3, the image memory 21, or the object map storage unit 26 in FIG. 2 is a non-volatile memory such as a hard disk device, a magneto-optical disk device, a flash memory, or a CD-ROM. May be configured by a storage medium that can only be stored in the memory, a volatile memory such as a RAM (Random Access Memory), or a combination thereof.
 また、図2における、画像変換部22、背景生成部24、ID生成/消滅部25、移動物体追跡部27、輪郭抽出部30、補正部31、判定部32、移動物体変動量検出部33、制御部34、またはオクルージョン検出部35という処理部は、専用のハードウェアにより実現されてもよい。また、この処理部はメモリおよびCPU(中央演算装置)により構成され、処理部の機能を実現するためのプログラムをこのメモリにロードして実行することによりその機能を実現させてもよい。 2, the image conversion unit 22, the background generation unit 24, the ID generation / annihilation unit 25, the moving object tracking unit 27, the contour extraction unit 30, the correction unit 31, the determination unit 32, the moving object fluctuation amount detection unit 33, The processing unit called the control unit 34 or the occlusion detection unit 35 may be realized by dedicated hardware. The processing unit may be configured by a memory and a CPU (central processing unit), and the function may be realized by loading a program for realizing the function of the processing unit into the memory and executing the program.
 また、図1における、画像変換部22、背景生成部24、ID生成/消滅部25、移動物体追跡部27、輪郭抽出部30、補正部31、判定部32、移動物体変動量検出部33、制御部34またはオクルージョン検出部35という処理部の機能を実現するためのプログラムをコンピュータ読み取り可能な記録媒体に記録して、この記録媒体に記録されたプログラムをコンピュータシステムに読み込ませ、実行することにより、この処理部による処理を実行してもよい。なお、「コンピュータシステム」とは、OSや周辺機器等のハードウェアを含む。 Further, in FIG. 1, the image conversion unit 22, the background generation unit 24, the ID generation / annihilation unit 25, the moving object tracking unit 27, the contour extraction unit 30, the correction unit 31, the determination unit 32, the moving object variation amount detection unit 33, By recording a program for realizing the function of the processing unit such as the control unit 34 or the occlusion detection unit 35 on a computer-readable recording medium, causing the computer system to read and execute the program recorded on the recording medium The processing by this processing unit may be executed. The “computer system” includes an OS and hardware such as peripheral devices.
 また、「コンピュータシステム」は、WWWシステムを利用している場合であれば、ホームページ提供環境(あるいは表示環境)も含む。
 また、「コンピュータ読み取り可能な記録媒体」は、フレキシブルディスク、光磁気ディスク、ROM、CD-ROM等の可搬媒体、コンピュータシステムに内蔵されるハードディスク等の記憶装置である。
 さらに「コンピュータ読み取り可能な記録媒体」は、インターネット等のネットワークや電話回線等の通信回線を介してプログラムを送信する場合の通信線のように、短時間の間、動的にプログラムを保持するもの、その場合のサーバやクライアントとなるコンピュータシステム内部の揮発性メモリのように、一定時間プログラムを保持しているものも含む。
 また、上記プログラムは、前述した機能の一部を実現するためのものであっても良く、さらに前述した機能をコンピュータシステムにすでに記録されているプログラムとの組み合わせで実現できるものであっても良い。
Further, the “computer system” includes a homepage providing environment (or display environment) if a WWW system is used.
The “computer-readable recording medium” is a portable medium such as a flexible disk, a magneto-optical disk, a ROM, a CD-ROM, or a storage device such as a hard disk built in a computer system.
Furthermore, “computer-readable recording medium” is a program that dynamically holds a program for a short time, such as a communication line when a program is transmitted via a network such as the Internet or a communication line such as a telephone line. In this case, a volatile memory in a computer system that serves as a server or a client in this case includes a program that holds a program for a certain period of time.
Further, the program may be a program for realizing a part of the functions described above, and may be a program capable of realizing the functions described above in combination with a program already recorded in a computer system. .
 以上、本発明の一実施形態を、図面を参照しながら詳述してきたが、具体的な構成は本実施形態のみに限られるものではなく、本発明の要旨を逸脱しない範囲の設計等も含まれる。 As mentioned above, although one embodiment of the present invention has been described in detail with reference to the drawings, the specific configuration is not limited to this embodiment alone, and includes a design and the like within a scope not departing from the gist of the present invention. It is.
 本発明によれば、抽出した移動物体の輪郭に基づいて、オブジェクトマップ記憶部に記憶されているブロックの識別符号を補正することにより、画像の背景が変動する場合でも、画像中の移動物体を正確に検出できる。 According to the present invention, by correcting the block identification code stored in the object map storage unit based on the extracted contour of the moving object, the moving object in the image can be detected even when the background of the image fluctuates. It can be detected accurately.
 10  カメラ
 20  移動物体追跡装置
 21  画像メモリ
 22  画像変換部
 23  フレームバッファメモリ
 24  背景画像生成部
 25  ID生成/消滅部
 26  オブジェクトマップ記憶部
 27  移動物体追跡部
 30  輪郭抽出部
 31  補正部
 32  判定部
 33  移動物体変動量検出部
 34  制御部
 35  オクルージョン検出部
 301 対象領域設定部
 302 対象領域補正部
 303 輪郭抽出処理部
DESCRIPTION OF SYMBOLS 10 Camera 20 Moving object tracking device 21 Image memory 22 Image conversion part 23 Frame buffer memory 24 Background image generation part 25 ID production | generation / extinction part 26 Object map memory | storage part 27 Moving object tracking part 30 Contour extraction part 31 Correction | amendment part 32 Determination part 33 Moving object variation detection unit 34 Control unit 35 Occlusion detection unit 301 Target region setting unit 302 Target region correction unit 303 Contour extraction processing unit

Claims (10)

  1.  画像処理により時系列画像中の移動物体を検出する移動物体追跡装置であって、
     前記移動物体に対応する識別符号を、前記時系列画像の各フレームが複数に分割されたブロックに割付けて記憶するオブジェクトマップ記憶部と;
     前記時系列画像から前記移動物体の輪郭を抽出する輪郭抽出部と;
     前記輪郭に基づいて、前記各ブロックに割付けられた前記識別符号を補正する補正部と;
    を備えることを特徴とする移動物体追跡装置。
    A moving object tracking device for detecting a moving object in a time-series image by image processing,
    An object map storage unit for storing an identification code corresponding to the moving object by assigning and storing the identification code corresponding to a block obtained by dividing each frame of the time-series image into a plurality of blocks;
    A contour extraction unit that extracts a contour of the moving object from the time-series image;
    A correction unit for correcting the identification code assigned to each block based on the contour;
    A moving object tracking device comprising:
  2.  前記輪郭抽出部が、
     前記オブジェクトマップ記憶部に記憶されている前記移動物体に対応する画像領域に基づいて、前記移動物体の輪郭を抽出する対象領域を設定する対象領域設定部と;
     前記対象領域に対して、前記時系列画像から前記移動物体の輪郭を抽出する輪郭抽出処理部と;
    を備えることを特徴とする請求項1に記載の移動物体追跡装置。
    The contour extraction unit
    A target area setting unit that sets a target area for extracting an outline of the moving object based on an image area corresponding to the moving object stored in the object map storage unit;
    A contour extraction processing unit that extracts a contour of the moving object from the time-series image with respect to the target region;
    The moving object tracking device according to claim 1, comprising:
  3.  前記輪郭抽出部が、前記時系列画像をエッジ抽出処理した画像に対して、前記対象領域内でエッジに対応する画素数の個数についてのヒストグラムを座標軸毎に射影して生成し、このヒストグラムに基づいて前記対象領域を前記座標軸毎に補正する対象領域補正部をさらに備え;
     前記輪郭抽出処理部が、補正された前記対象領域に対して、前記時系列画像から前記移動物体の輪郭を抽出する;
    ことを特徴とする請求項2に記載の移動物体追跡装置。
    The contour extraction unit generates and generates a histogram for the number of pixels corresponding to an edge in the target region for an image obtained by performing edge extraction processing on the time-series image, based on the histogram. A target area correction unit that corrects the target area for each coordinate axis;
    The contour extraction processing unit extracts the contour of the moving object from the time-series image with respect to the corrected target region;
    The moving object tracking device according to claim 2.
  4.  前記輪郭抽出部が、オクルージョンが発生している複数の移動物体を一体の移動物体とし、前記時系列画像から前記一体の移動物体の輪郭を抽出する
    ことを特徴とする請求項1に記載の移動物体追跡装置。
    The movement according to claim 1, wherein the contour extraction unit extracts a contour of the integral moving object from the time-series image by using a plurality of moving objects in which occlusion has occurred as an integral moving object. Object tracking device.
  5.  前記補正部が、
    前記オブジェクトマップ記憶部に記憶されている前記オクルージョンが発生している複数の移動物体に対応する識別符号に基づいたこれら複数の移動物体の境界を示す情報と、前記輪郭抽出部により前記オクルージョンが発生している複数の移動物体を一体として抽出された前記移動物体の輪郭とに基づいて、
    前記オブジェクトマップ記憶部に記憶されている前記識別符号を補正する
    ことを特徴とする請求項4に記載の移動物体追跡装置。
    The correction unit is
    Information indicating boundaries of the plurality of moving objects based on the identification codes corresponding to the plurality of moving objects in which the occlusion occurs, which is stored in the object map storage unit, and the occlusion is generated by the contour extraction unit Based on the contour of the moving object extracted as a plurality of moving objects that are integrated,
    The moving object tracking device according to claim 4, wherein the identification code stored in the object map storage unit is corrected.
  6.  背景画像が変動しているか否かを判定する判定部と;
     前記判定部が前記背景画像が変動していると判定した場合に、前記輪郭抽出部に輪郭を抽出させ、前記補正部に前記オブジェクトマップ記憶部に記憶されている前記識別符号を補正させる第1の制御部と;
    をさらに備えることを特徴とする請求項1に記載の移動物体追跡装置。
    A determination unit for determining whether the background image is fluctuating;
    First, when the determination unit determines that the background image is fluctuating, the contour extraction unit extracts a contour, and the correction unit corrects the identification code stored in the object map storage unit. A control unit;
    The moving object tracking device according to claim 1, further comprising:
  7.  前記オブジェクトマップ記憶部に記憶されている前記移動物体のサイズまたは移動量の変動量を、前記識別符号に基づいて単位時間毎に検出する移動物体変動量検出部と;
     前記移動物体変動量検出部により検出された単位時間毎の前記移動物体のサイズまたは移動量の前記変動量が所定のサイズまたは移動量の変動量よりも大きい場合に、前記輪郭抽出部に輪郭を抽出させ、前記補正部に前記オブジェクトマップ記憶部に記憶されている前記識別符号を補正させる第2の制御部と;
    をさらに備えることを特徴とする請求項1に記載の移動物体追跡装置。
    A moving object fluctuation amount detection unit that detects a fluctuation amount of the size or movement amount of the moving object stored in the object map storage unit per unit time based on the identification code;
    If the moving object size or moving amount per unit time detected by the moving object changing amount detection unit is larger than a predetermined size or moving amount variation, a contour is added to the contour extracting unit. A second control unit that causes the correction unit to extract and correct the identification code stored in the object map storage unit;
    The moving object tracking device according to claim 1, further comprising:
  8.  前記時系列画像を画像処理した結果に基づいて、前記識別符号と、前記オブジェクトマップ記憶部に記憶されている前記移動物体の動きベクトルと、を更新する移動物体追跡部をさらに備え、
     前記移動物体追跡部が、
     前記時系列画像のうちの連続するN画像(N≧2)の各々に対して、隣り合うブロックの動きベクトルの差の絶対値が所定値以内のブロックに同一の識別符号を付けることにより、画像上で互いに重なった前記移動物体に互いに異なる識別符号を付ける識別符号付与工程と;
     前記N画像の各々において、第1識別符号が付けられたブロック群である第1オブジェクトと第2識別符号が付けられたブロック群である第2オブジェクトとが接し、かつ、前記N画像について時間的に隣り合う画像の前記第1オブジェクト間の相関度が所定値以上であるか否かを判定する判定工程と;
     前記判定工程で肯定と判定された場合に、時間を遡って前記第1オブジェクトと前記第2オブジェクトとを追跡する追跡工程と;
     前記追跡工程により時間を遡って追跡された前記第1オブジェクトと前記第2オブジェクトとに基づいて、前記オブジェクトマップ記憶部に記憶されている前記識別符号と前記動きベクトルとを更新する更新工程と;
    を実行することを特徴とする請求項1に記載の移動物体追跡装置。
    A moving object tracking unit that updates the identification code and a motion vector of the moving object stored in the object map storage unit based on a result of image processing of the time-series image;
    The moving object tracking unit is
    For each successive N images (N ≧ 2) of the time-series images, the same identification code is attached to the blocks whose absolute values of motion vector differences between adjacent blocks are within a predetermined value. An identification code providing step of attaching different identification codes to the moving objects that overlap each other;
    In each of the N images, a first object that is a block group to which a first identification code is attached is in contact with a second object that is a block group to which a second identification code is attached, and the N image is temporally related. A determination step of determining whether or not the degree of correlation between the first objects of images adjacent to each other is equal to or greater than a predetermined value;
    A tracking step of tracking the first object and the second object retroactively when it is determined affirmative in the determination step;
    An update step of updating the identification code and the motion vector stored in the object map storage unit based on the first object and the second object traced back in time by the tracking step;
    The moving object tracking device according to claim 1, wherein:
  9.  画像処理により時系列画像中の移動物体を検出する移動物体追跡方法であって、
     前記移動物体に対応する識別符号を、前記時系列画像の各フレームが複数に分割されたブロックに割付けて記憶する工程と;
     前記時系列画像から前記移動物体の輪郭を抽出する工程と;
     前記輪郭に基づいて、前記ブロックに割付けられた前記識別符号を補正する工程と;
    を備えることを特徴とする移動物体追跡方法。
    A moving object tracking method for detecting a moving object in a time-series image by image processing,
    Assigning and storing an identification code corresponding to the moving object to a block obtained by dividing each frame of the time-series image into a plurality of blocks;
    Extracting the contour of the moving object from the time-series image;
    Correcting the identification code assigned to the block based on the contour;
    A moving object tracking method comprising:
  10.  画像処理により時系列画像中の移動物体を検出する移動物体追跡装置としてのコンピュータに、
     前記移動物体に対応する識別符号を、前記時系列画像の各フレームが複数に分割されたブロックに割付けて記憶する工程と;
     前記時系列画像から前記移動物体の輪郭を抽出する工程と;
     前記輪郭に基づいて、前記ブロックに割付けられた前記識別符号を補正する工程と;
    を実行させるための移動物体追跡プログラム。
    To a computer as a moving object tracking device that detects moving objects in time-series images by image processing,
    Assigning and storing an identification code corresponding to the moving object to a block obtained by dividing each frame of the time-series image into a plurality of blocks;
    Extracting the contour of the moving object from the time-series image;
    Correcting the identification code assigned to the block based on the contour;
    Moving object tracking program to execute.
PCT/JP2010/002015 2009-04-02 2010-03-19 Moving object tracking device, moving object tracking method, and moving object tracking program WO2010113417A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2009-090489 2009-04-02
JP2009090489A JP5371040B2 (en) 2009-04-02 2009-04-02 Moving object tracking device, moving object tracking method, and moving object tracking program

Publications (1)

Publication Number Publication Date
WO2010113417A1 true WO2010113417A1 (en) 2010-10-07

Family

ID=42827733

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2010/002015 WO2010113417A1 (en) 2009-04-02 2010-03-19 Moving object tracking device, moving object tracking method, and moving object tracking program

Country Status (2)

Country Link
JP (1) JP5371040B2 (en)
WO (1) WO2010113417A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103310465A (en) * 2013-06-27 2013-09-18 东南大学 Vehicle occlusion treating method based on Markov random field

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101180259B1 (en) 2011-06-10 2012-09-06 엘아이지넥스원 주식회사 Target tracking method
JP5687657B2 (en) * 2012-05-25 2015-03-18 株式会社東芝 Object detection apparatus, object detection method, and object detection program
KR101284815B1 (en) 2012-06-18 2013-07-10 엘아이지넥스원 주식회사 Target tracking apparatus
JP6090786B2 (en) 2013-05-09 2017-03-08 国立大学法人 東京大学 Background difference extraction apparatus and background difference extraction method
KR101619953B1 (en) * 2015-08-03 2016-05-23 주식회사 영국전자 Method for image processing of surveilance camera by using auto multi-tracing
KR101619956B1 (en) * 2015-08-03 2016-05-11 주식회사 영국전자 Apparatus for image processing of surveilance camera by using auto multi-tracing
CN112085953B (en) * 2019-06-12 2022-08-05 杭州海康威视系统技术有限公司 Traffic command method, device and equipment

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001167282A (en) * 1999-12-10 2001-06-22 Toshiba Corp Device and method for extracting moving object
JP2006024148A (en) * 2004-07-09 2006-01-26 Foundation For The Promotion Of Industrial Science Moving object tracking method and device on image
JP2008217706A (en) * 2007-03-07 2008-09-18 Tokyo Institute Of Technology Labeling device, labeling method and program

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001167282A (en) * 1999-12-10 2001-06-22 Toshiba Corp Device and method for extracting moving object
JP2006024148A (en) * 2004-07-09 2006-01-26 Foundation For The Promotion Of Industrial Science Moving object tracking method and device on image
JP2008217706A (en) * 2007-03-07 2008-09-18 Tokyo Institute Of Technology Labeling device, labeling method and program

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
MASAAKI KOBAYASHI ET AL.: "Notanchi to Optical Flow o Heiyo shita Gazo Ryoiki Bunkatsu", IEICE TECHNICAL REPORT, vol. 94, no. 509, 17 February 1995 (1995-02-17), pages 47 - 54 *
SHUNSUKE KAMIJO ET AL.: "Jikukan Markov Random Field Model ni yoru Kakure ni Robust na Tracking Algorithm", THE TRANSACTIONS OF THE INSTITUTE OF ELECTRONICS, INFORMATION AND COMMUNICATION ENGINEERS, vol. J83-D-II, no. 12, 25 December 2000 (2000-12-25), pages 2597 - 2609 *
SHUNSUKE KAMIJO ET AL.: "Jikukan MRF Model to Kaisokan Kyocho Algorithm ni yoru Sharyo Bunkatsu", THE TRANSACTIONS OF THE INSTITUTE OF ELECTRONICS, INFORMATION AND COMMUNICATION ENGINEERS, vol. J9-1-A, no. 1, 1 January 2008 (2008-01-01), pages 55 - 67 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103310465A (en) * 2013-06-27 2013-09-18 东南大学 Vehicle occlusion treating method based on Markov random field
CN103310465B (en) * 2013-06-27 2016-02-03 东南大学 A kind of occlusion disposal route based on markov random file

Also Published As

Publication number Publication date
JP5371040B2 (en) 2013-12-18
JP2010244207A (en) 2010-10-28

Similar Documents

Publication Publication Date Title
WO2010113417A1 (en) Moving object tracking device, moving object tracking method, and moving object tracking program
JP4217876B2 (en) Method and apparatus for tracking moving object in image
JP5102410B2 (en) Moving body detection apparatus and moving body detection method
US7139411B2 (en) Pedestrian detection and tracking with night vision
TWI420906B (en) Tracking system and method for regions of interest and computer program product thereof
EP1542153A1 (en) Object detection
JP2000011133A (en) Device and method for detecting moving object
EP1542154A2 (en) Object detection
KR100994367B1 (en) Method for tracking a movement of a moving target of image tracking apparatus
KR101139930B1 (en) Apparatus and method for object recognition based on part-template matching
JP2002133421A (en) Moving body recognition method and device
US20220366570A1 (en) Object tracking device and object tracking method
CN109785357A (en) A method of the robot automtion panorama photoelectronic reconnaissance suitable for battlefield surroundings
Sincan et al. Moving object detection by a mounted moving camera
JP4283532B2 (en) Image recognition apparatus and method
KR100566629B1 (en) System for detecting moving objects and method thereof
JP4543106B2 (en) Method and apparatus for tracking moving object in image
CN115019241A (en) Pedestrian identification and tracking method and device, readable storage medium and equipment
JP6555940B2 (en) Subject tracking device, imaging device, and method for controlling subject tracking device
JP2005517257A (en) Segmentation apparatus and method
JP3763279B2 (en) Object extraction system, object extraction method, and object extraction program
JP2001307104A (en) Object extraction device for moving image
JP7488673B2 (en) MOVING OBJECT TRACKING DEVICE, MOVING OBJECT TRACKING METHOD, AND MOVING OBJECT TRACKING PROGRAM
JP2020181268A (en) Body homologizing device, body homologizing system, body homologizing method, and computer program
CN111563912B (en) Pedestrian tracking system and method

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 10758202

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 10758202

Country of ref document: EP

Kind code of ref document: A1