WO2017126187A1 - Video monitoring apparatus and video monitoring method - Google Patents

Video monitoring apparatus and video monitoring method Download PDF

Info

Publication number
WO2017126187A1
WO2017126187A1 PCT/JP2016/082598 JP2016082598W WO2017126187A1 WO 2017126187 A1 WO2017126187 A1 WO 2017126187A1 JP 2016082598 W JP2016082598 W JP 2016082598W WO 2017126187 A1 WO2017126187 A1 WO 2017126187A1
Authority
WO
WIPO (PCT)
Prior art keywords
person
suspicious
candidate
flow information
human flow
Prior art date
Application number
PCT/JP2016/082598
Other languages
French (fr)
Japanese (ja)
Inventor
雅志 神谷
芳知 中村
内藤 正博
Original Assignee
三菱電機株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 三菱電機株式会社 filed Critical 三菱電機株式会社
Publication of WO2017126187A1 publication Critical patent/WO2017126187A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/18Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
    • G08B13/189Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
    • G08B13/194Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
    • G08B13/196Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B25/00Alarm systems in which the location of the alarm condition is signalled to a central station, e.g. fire or police telegraphic systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast

Definitions

  • This invention relates to a video monitoring device for monitoring a suspicious person or the like using a camera.
  • a security guard or other monitoring person looks at the display on which the camera image is displayed, estimates the suspicious person from the person appearing on the image displayed on the display, and performs the action of the person who is suspected of the estimated suspicious person
  • the suspicious person is identified by checking the camera image or directly in the field.
  • this method has a problem that it takes time to identify a suspicious person if the number of people or the number of camera images is large.
  • Patent Document 1 if there are many people gathering, it is necessary to follow the movement trajectory of all the people shown in the camera image. If the number of people is large, the computer processing is enormous. there were.
  • the present invention has been made to solve the above-described problems, and an object of the present invention is to provide a video monitoring apparatus that suppresses an increase in processing when the number of persons shown in a camera video increases.
  • a person is detected from a frame image, a person tracking unit that tracks each detected person over a plurality of frames, and a tracking result of the person tracking unit,
  • the human flow information generating unit that recognizes the person estimated to be close in moving direction and moving speed as a group, and generates human flow information including the group moving direction and the group moving speed for each group,
  • a human flow information analysis unit for determining whether or not an abnormal event has occurred, and identifying a position when it is determined that an abnormal event has occurred; among the persons, the human flow information analysis unit is present in the vicinity of the position specified by the human flow information analysis unit
  • a suspicious person candidate detecting unit that detects the person to be detected as a suspicious person candidate, and a suspicious person candidate presenting unit that presents the detected suspicious person candidate.
  • this invention recognizes a plurality of detected persons as a group and detects suspicious candidates from the change in the flow of the group, even if the number of persons appearing in the camera video increases, it can be grasped collectively by the group. it can. For this reason, there is an effect that it is possible to suppress an increase in processing even if the number of persons shown in the camera video increases.
  • FIG. 1 is a block diagram showing a configuration of a video monitoring system according to a first exemplary embodiment. It is a figure which shows an example of the person tracking information D1 concerning Embodiment 1.
  • FIG. It is a figure which shows an example of the human flow information D2 concerning Embodiment 1.
  • FIG. It is a figure which shows an example of the suspicious person candidate presentation information D5 shown on the display.
  • FIG. 3 is a flowchart showing processing of the video monitoring apparatus 10 according to the first exemplary embodiment. It is a block diagram which shows the structure of the video monitoring system concerning Embodiment 2.
  • FIG. It is a block diagram which shows the structural modification of the video surveillance system concerning Embodiment 2.
  • FIG. 1 is a block diagram showing the configuration of the video monitoring system according to the present embodiment.
  • the video monitoring apparatus 10 presents the suspicious person candidate presentation information D5 on the display 30 with the video from the camera 20 as an input.
  • the video monitoring apparatus 10 includes a person tracking unit 101, a human flow information generation unit 102, a human flow information analysis unit 103, a suspicious person information detection unit 104, and a suspicious person information presentation unit 105.
  • the camera 20 holds captured images in units of frames. Further, the camera 20 outputs to the video monitoring apparatus 10 in units of frames.
  • the captured image is also referred to as a frame image.
  • the camera 20 is provided at a position where an area with a person's flow line can be imaged.
  • the “flow line” is a line that represents a path that a person is supposed to take when moving naturally.
  • the person tracking unit 101 detects a person from the frame image input from the camera 20. Further, when the detected person is recognized as the same person over a plurality of frames, the person tracking unit 101 generates person tracking information D1 indicating the movement information. The person tracking unit 101 outputs the generated person tracking information D1 to the person information generating unit 102. In addition, the person tracking unit 101 outputs the person detection result to the suspicious person candidate detecting unit 104.
  • a person detection method by the person tracking unit 101 there is a method in which a large number of person images are collected in advance and a person feature such as brightness gradient information and contour information is learned.
  • a person discriminator when a person discriminator is created and a frame image is acquired from the camera 20, it is determined whether or not the person discriminator includes an area having a feature similar to the learned human feature in the frame image. Determine. If an area is included, the coordinates of the area (starting point coordinates and vertical and horizontal lengths) are specified.
  • HOG Heistogram of orientated gradients
  • the person tracking unit 101 As a person tracking method by the person tracking unit 101, the person feature extracted when the person is detected by the person detection method described above is used, and the similarity between the person features detected in a plurality of frames is equal to or higher than a certain threshold. In such a case, there is a method for determining that these persons are the same. However, when there are a plurality of combinations having a similarity exceeding the threshold, the one with the highest similarity is determined as the same person.
  • the degree of similarity is high, it can be determined that they are not the same person if the specified coordinates differ greatly.
  • the time between frames that the person tracking unit 101 continuously acquires from the camera 20 is 33 ms, and the actual distance to one pixel on the frame image is 0.5 m.
  • the start point coordinates of a person specified by the above-described person detection method and whose similarity is equal to or greater than the threshold are 1000 pixels apart between consecutive frames. In this case, if this person is the same person, it will proceed 500 m in 33 ms, which is practically impossible, so it can be determined that they are not the same person.
  • the moving speed of a similar person can be estimated from the information on the positional relationship of the camera 20 installed in the monitoring system and the time interval between frame images, the moving speed is unlikely to be considered a human moving speed. It is determined that they are not the same person.
  • FIG. 2 is a diagram illustrating an example of the person tracking information D1 generated by the person tracking unit 101.
  • the person tracking information D1 includes a person number D1a indicating each detected person, a detection frame number D1b, a detection coordinate D1c, and person feature information D1d.
  • the person number D1a is a number that the person tracking unit 101 assigns separately to enable identification of different persons.
  • the detected frame number D1b indicates the frame number of the frame image in which the detected person is shown for each detected person.
  • the frame number may be a number assigned to each frame image output from the camera 20.
  • the frame number may be a number assigned by the person tracking unit 101 for each frame image.
  • the detected coordinates D1c are coordinates at which the person tracking unit 101 detects the person in the frame image.
  • the coordinates are expressed by a two-dimensional axis with the pixel position at the upper left of the frame image as (0, 0), for example, and the position where the center position of the detected person is estimated is detected coordinates D1c.
  • the position of a part of the detected person such as the center position of the head may be used as the detection coordinate D1c.
  • the person feature information D1d is a person feature extracted by the person tracking unit 101.
  • the person feature is, for example, a HOG feature amount, and indicates a feature amount obtained by histogramating the relationship between the gradient direction (edge direction) and the gradient strength in the image area obtained by dividing the frame image. For example, when the image size of a frame image or an image area in which a person is shown is 30 pixels ⁇ 60 pixels, the gradient intensity for each gradient direction in units of cells, with 1 unit cell of the divided image region being 5 pixels ⁇ 5 pixels To obtain a histogram. For example, when the histogram is formed, the gradient direction is divided into 9 directions of 20 [deg] from 0 [deg] to 180 [deg].
  • the person feature of the person 1 in the frame number 1625 has a gradient strength of 30 in the gradient direction 0 [deg], a gradient strength of 19 in the gradient direction 20 [deg], and a gradient direction of 40 [deg]. It shows that the gradient strength is 17.
  • the person feature of the person 2 in the frame number 1173 has a gradient strength of 10 in the gradient direction 0 [deg], a gradient strength of 2 in the gradient direction 20 [deg], and a gradient strength in the gradient direction of 40 [deg]. Is 57.
  • the human flow information generation unit 102 recognizes, as a group, detected people whose movement direction and movement speed are close, and generates human flow information D2.
  • the human flow means that when a person who moves at the same speed in the same direction from the person tracked by the person tracking unit 101 and recognizes a person existing within a certain range (neighborhood) as a group, it is one of each group.
  • the movement information in the above frame (group center coordinates, movement direction, movement speed) is represented.
  • the center coordinates of the group are average coordinates of the detection coordinates D1c of the persons included in the group.
  • FIG. 3 is a diagram illustrating an example of the human flow information D2 output from the human flow information generation unit 102 to the human flow information analysis unit 103.
  • D2a is a group number, which is a number assigned separately to enable identification for different groups.
  • FIG. 3 shows that persons corresponding to person numbers D1a, 1, 4, and 7 are recognized as one group (group number 1).
  • the center coordinate D2c of the group corresponding to the group number 1 is (275, 196).
  • the human flow information D2 includes the person number D2b of the included person, the center coordinates D2c of the group, the group movement direction D2d, and the group movement speed D2e for each group.
  • the collective movement direction D2d represents an angle of a direction moved between consecutive frames when a specific coordinate direction is set to 0 [deg].
  • the unit of the collective movement direction D2d is [deg].
  • the collective movement speed D2e represents the distance of the central coordinates D2c of the collective between successive frames.
  • the unit of the collective movement speed D2e is a moving pixel distance [pixel / frame] in one frame unit.
  • the collective moving speed D2e may be obtained by adding the square of the number of moving pixels in the horizontal direction and the square of the number of moving pixels in the vertical direction during one frame.
  • the collective moving speed D2e may be a value obtained by calculating the square root from the sum of the square of the number of moving pixels in the horizontal direction and the square of the number of moving pixels in the vertical direction during one frame.
  • the distance per pixel can be converted by setting in advance from the positional relationship in which the camera 20 is provided. For example, the distance conversion setting per pixel can be determined from the number of floor tile pixels in the frame image captured by the camera 20 and the actual floor tile distance.
  • the human flow information analysis unit 103 analyzes the human flow information D2 generated by the human flow information generation unit 102 and generates anomaly occurrence information D3. Also, the incident occurrence information D ⁇ b> 3 is output to the suspicious person candidate detection unit 104.
  • the anomaly occurrence information D3 is the anomaly occurrence frame number D3a when the anomaly occurs when it is determined that an anomaly has occurred in the person flow as a result of analyzing the person flow information D2 by the person flow information analysis unit 103.
  • the coordinates at the time of occurrence of an anomaly are included as the anomaly occurrence coordinates D3b.
  • a method will be described in which the human flow information analysis unit 103 determines that an anomaly has occurred in the human flow.
  • An anomaly in the human flow indicates a sudden change in the group movement speed D2e in the group being tracked or a change in which the number of persons belonging to the group being tracked decreases rapidly.
  • the former is a case where the time fluctuation amount of the collective movement speed D2e exceeds a set threshold value. For example, when the collective movement speed D2e is rapidly increased, an abnormality occurs in the vicinity of the collective movement speed D2e. it is conceivable that. As an example of the latter, if three people belong to a group and suddenly start moving in different directions, the conditions of “neighboring” or “same direction” for recognizing the group will not be met.
  • the group that belonged to disappears and each is newly tracked as a different group. However, if the group moves out of the angle of view of the camera 20 due to the movement and no longer appears, it is not determined that an abnormality has occurred. That is, the former is a case of running away in the same direction as a group, and the latter is a case of running away so as to be scattered all at once.
  • the human flow information analysis unit 103 detects that an abnormal change has occurred in the human flow by detecting a sudden change, for example, one of the persons belonging to the group starts moving in different directions without changing the speed. If this is the case, basically, this is not judged as an incident. In this case, since the destinations are simply different, there is a high possibility that they have started moving in different directions. On the other hand, when an abnormality occurs, the collective movement speed D2e is considered to become very large at that time and thereafter.
  • the person flow information analysis unit 103 determines that there is an abnormality, it generates the abnormality occurrence information D3 and outputs it to the suspicious person candidate detection unit 104. On the other hand, when it is determined that there is no abnormality, the human flow information analysis unit 103 does not output the abnormal occurrence occurrence information D3 to the suspicious person candidate detection unit 104 and waits until the human flow information D2 is updated.
  • the suspicious candidate detection unit 104 detects the suspicious candidate candidate by analyzing the person tracking information D1 generated by the person tracking unit 101 and the anomaly occurrence information D3 generated by the human flow information analyzing unit 103, and the suspicious candidate information D4 is detected. Generate. Further, the suspicious person candidate detecting unit 104 outputs the suspicious person candidate information D4 to the suspicious person candidate presenting part 105.
  • the suspicious candidate detection unit 104 operates only when the incident occurrence information D3 is input from the human flow information analysis unit 103 to the suspicious person candidate detection unit 104 or when the incident change information D3 is changed. It is good.
  • the suspicious person candidate information D4 includes a person number D1a assigned in the person tracking information D1 for a person detected by the suspicious person candidate detecting unit 104 as a suspicious person candidate, a detection frame number D1b that is information related to the person, a detection coordinate D1c, The information extracted from the person feature information D1d is included.
  • the suspicious person candidate detecting unit 104 compares the anomaly occurrence coordinates D3b with the person detection coordinates D1c, and identifies a person existing at a distance within a preset threshold from the anomaly occurrence coordinates D3b as a suspicious person candidate.
  • the suspicious person candidate detection unit 104 compares the anomaly occurrence coordinates D3b and the person detection coordinates D1c, and identifies a person existing at a distance closest to the anomaly occurrence coordinates D3b as a suspicious person candidate.
  • the suspicious person candidate detection part 104 specifies only the preset number of persons as a suspicious person candidate in an order from the person close
  • the preset threshold for distance and threshold for the number of suspicious candidates can be determined arbitrarily. For example, if you want to detect as many suspicious candidates as possible and ultimately identify them by visual inspection of guards, set a threshold to increase the distance, or set a threshold to increase the number of suspicious candidates be able to. When it is desired to narrow down candidates as much as possible, a threshold value can be set to shorten the distance, or a threshold value can be set to reduce the number of people to one.
  • the suspicious candidate presentation unit 105 inputs the suspicious candidate information D4 generated by the suspicious candidate detection unit 104 and the frame image output from the camera 20, and generates suspicious candidate presentation information D5 on the display 30. Output.
  • the display 30 inputs and displays the suspicious candidate presentation information D5 generated by the suspicious candidate presentation unit 105.
  • FIG. 4 is a diagram showing an example of the suspicious candidate presentation information D5 presented on the display 30. As shown in FIG. In FIG. 4, among the persons shown in the frame image output from the camera 20, a person who is detected as a suspicious candidate is displayed in a rectangle. By presenting the information on the display 30 in this manner, it is easy for the security guard to recognize at a glance whether or not there is a suspicious candidate and its position.
  • FIG. 5 is a flowchart showing processing of the video monitoring apparatus 10.
  • the person tracking unit 101 acquires this frame image (S002). If the camera 20 does not hold a new frame image (S001: No), the person tracking unit 101 waits until the camera 20 holds a new frame image.
  • the person tracking unit 101 acquires a frame image (S002), and then performs a person detection process on the frame image (S003).
  • the person tracking unit 101 obtains a similarity between the detected person feature of the person and the feature of the person detected in the previous frame. However, if no person is detected in the previous frame, the person tracking unit 101 does not perform this process (S005). If no person can be detected, the person tracking unit 101 waits until the camera 20 holds a new frame image (S004: No).
  • the person tracking unit 101 determines that the person is the same person and generates person tracking information D1 (S006: Yes). If a person with high similarity cannot be detected, the person tracking unit 101 waits until the camera 20 holds a new frame image (S006: No).
  • the human flow information generation unit 102 receives the person tracking information D1 from the person tracking unit 101, and calculates the moving direction and moving speed of each person being tracked (S007).
  • the human flow information generation unit 102 further recognizes, as a group, a person who is within a certain range (near) for each person being tracked and corresponds to the calculated moving direction and moving speed.
  • the human flow information D2 including the person number D2b, the central coordinate D2c of the group, the group movement direction D2d, and the group movement speed D2e is generated and transmitted to the human flow information analysis unit 103 (S008).
  • the human flow information analysis unit 103 receives the human flow information D2 from the human flow information generation unit 102, and determines whether or not an anomaly has occurred in the human flow (S009).
  • the human flow information analysis unit 103 determines that there is an abnormality
  • the human flow analysis unit 13 transmits the abnormality occurrence information D3 to the suspicious person candidate detection unit 104 (S009: Yes).
  • the human flow information analysis unit 103 waits until the camera 20 holds a new frame image (S009: No).
  • the suspicious person candidate detection unit 104 receives the anomaly occurrence information D3 from the human flow information analysis unit 103. Further, the suspicious person candidate detection unit 104 receives the person tracking information D ⁇ b> 1 from the person tracking unit 101. Thereafter, the suspicious person candidate detection unit 104 compares the anomaly occurrence coordinate D3b with the person detection coordinate D1c, and identifies a person existing at a distance within a certain threshold from the anomaly occurrence coordinate D3b as a suspicious person candidate (S010). ). Alternatively, the suspicious person candidate detection unit 104 compares the anomaly occurrence coordinates D3b and the person detection coordinates D1c, and identifies a person existing at a distance closest to the anomaly occurrence coordinates D3b as a suspicious person candidate. Alternatively, the suspicious person candidate detection unit 104 specifies a certain number of suspicious person candidates as a suspicious person candidate in order from a person close to the anomaly occurrence coordinate D3b.
  • the suspicious person candidate detecting unit 104 generates suspicious person candidate information D4 and transmits it to the suspicious person candidate presenting part 105.
  • the suspicious person candidate presentation unit 105 receives the suspicious person candidate information D4 from the suspicious person candidate detection unit 104. Further, the suspicious person candidate presenting unit 105 acquires a frame image from the camera 20. Further, the suspicious person candidate presenting unit 105 generates suspicious person candidate presentation information D5 as illustrated in FIG. 5 and transmits it to the display 30, and the display 30 displays it (S011). In FIG. 5, the frame image received from the camera 20 is surrounded by a rectangle of a person who is determined as a suspicious candidate so that the guard can easily recognize the presence or position of the suspicious candidate at a glance. ing.
  • the video monitoring apparatus 10 since the video monitoring apparatus 10 according to the present embodiment recognizes a plurality of detected persons as a group and detects suspicious candidate candidates from the change in the flow of the group, the number of persons reflected in the camera video is large. Even then, it can be grasped collectively by the group. From this, the video monitoring apparatus 10 has an effect that it is possible to suppress an increase in processing even if the number of persons shown in the camera video increases.
  • the video monitoring device 10 can detect the disturbance of the human flow caused by the behavior of the suspicious person escaping from the incident occurrence point or the surrounding person escaping when the incident occurs.
  • the video monitoring apparatus 10 can automatically detect suspicious candidate candidates that exist in the vicinity of the point where the abnormality has occurred, and can track the candidate person and display it on the display. For this reason, the video monitoring apparatus 10 according to the present invention can detect a suspicious person who moves based on a suspicious person behavior pattern learned in advance as a suspicious person candidate. Accordingly, the security guard can quickly grasp the position of the suspicious candidate candidate and the escape route from a large amount of video data in the video surveillance device, and thus can perform appropriate security promptly.
  • the above-mentioned suspicious candidate detection unit 104 detects suspicious candidates from human-flow anomaly information.
  • disguise equipment such as masks and sunglasses and weapons are used by image recognition technology using the frame image of the camera.
  • equipment information can be used for detection of a suspicious candidate.
  • a person possessing disguise equipment or a weapon can be determined as having a high degree of suspiciousness.
  • the color of the rectangle in the suspicious person candidate presentation information D5 generated by the suspicious person candidate presentation unit 105 is made more conspicuous for a person with a high suspicious degree. As a result, the guard can grasp the position of the suspicious person more quickly.
  • the suspicious person candidate detection unit 104 performs age estimation by an image recognition technique using a frame image of the camera.
  • the suspicious person candidate detection unit 104 can use the information on the estimated age for detecting the suspicious person candidate. For example, it is considered unlikely that a 10-year-old child is a suspicious person. For this reason, assuming that such a person has a lower suspicious degree, the suspicious person candidate detection unit 104 changes the rectangular color in the suspicious person candidate presentation information D5 to an inconspicuous color. Alternatively, the suspicious person candidate detection unit 104 does not display such a person as a suspicious person candidate. By doing in this way, since a suspicious candidate can be narrowed down, a guard can grasp a true suspicious person's position more quickly.
  • the suspicious person candidate detection unit 104 can narrow down the suspicious person candidates from the positional relationship and the movement state of the person. For example, it is considered that a person who is not a suspicious person in the vicinity takes an action of escaping from a suspicious person after the occurrence of an incident. For this reason, among the persons who move in the same direction after the occurrence of an anomaly, there is a high possibility that the person who is further away from the anomaly occurrence point is not a suspicious person. This is because even if a suspicious person pursues a person who is not a suspicious person, the opposite is not possible. By doing in this way, the suspicious person candidate detection part 104 can narrow down a suspicious person candidate. For this reason, the guard can grasp the position of the true suspicious person more quickly.
  • the suspicious candidate detection unit 104 may determine that the suspicious candidate candidate is not a suspicious candidate because it is highly likely that the escaping action is not taken. By doing in this way, the suspicious person candidate detection part 104 can narrow down a suspicious person candidate. For this reason, the guard can grasp the position of the true suspicious person more quickly.
  • FIG. FIG. 6 is a block diagram of the configuration of the video monitoring system according to the second embodiment.
  • the video monitoring apparatus 11 is an apparatus that can perform the video monitoring method according to the second embodiment.
  • the same or corresponding components as those shown in FIG. 1 are denoted by the same reference numerals as those in FIG.
  • the suspicious person candidate detection unit 117 is connected to the abnormal sound detection unit 116.
  • the abnormal sound detection unit 116 is connected to the microphone 40.
  • the video monitoring apparatus 11 according to the second embodiment is different from the video monitoring apparatus 10 according to the first embodiment in that abnormal sound detection is used to detect a suspicious candidate. Except for this point, the second embodiment is the same as the first embodiment.
  • the microphone 40 transmits audio data to the abnormal sound detection unit 116.
  • the microphone 40 may be a single microphone, or a plurality of microphones 40 may be used so that the direction of the sound source can be extracted with high accuracy.
  • the abnormal sound detection unit 116 acquires audio data from the microphone 40 and analyzes the audio data. For example, the abnormal sound detection unit 116 classifies the voice data from the voice data into “conversation”, “BGM”, “footstep”, “scream”, “anger”, and the like. The abnormal sound detection unit 116 detects a sound highly relevant to the incident, which is different from the life sound such as “scream” or “anger” from the classified sound data. In the analysis of voice data, the frequency (voice pitch) and amplitude (voice volume) can be compared with model data in each classification, and the classification can be discriminated by taking a correlation.
  • the abnormal sound detection unit 116 detects “scream” and “anger” by the classification of the voice data, and also estimates the direction of the sound source.
  • the sound source direction estimation technology has been actively researched and developed in recent years, and the same method can be adopted. In order to estimate the sound source direction of a plurality of microphones, there is a method of detecting a time difference between audio data collected by each microphone.
  • the abnormal sound detection unit 116 detects “scream” and “anger”, and transmits the position to the suspicious candidate detection unit 117 as a result of estimating the direction of the sound source. For this position, the correspondence between the estimated position in the microphone 40 and the coordinate system imaged by the camera 20 is investigated in advance, and the position estimated by the microphone is replaced with the position in the camera coordinate system and transmitted. Since the operation and action of the suspicious candidate detection unit 117 are the same as those of the first embodiment, description thereof is omitted.
  • the position where the abnormal sound has occurred is specified by detecting the abnormal sound. Therefore, even in a crowded state in which a person moves intricately and the variation in the human flow described in the first embodiment cannot be captured by the camera, the video monitoring device 11 is located at the position where the abnormality has occurred. Can be identified, and suspicious candidates can be detected from the surrounding area. This makes it possible for security guards to identify suspicious individuals from a large amount of video data from video surveillance devices and to track their actions, quickly grasp the situation, and execute appropriate security. I can do it.
  • FIG. 7 is a block diagram showing a modification (video monitoring device 12) of the video monitoring device according to the first and second embodiments of the present invention. 7, components that are the same as or correspond to the components shown in FIG. 1 are given the same reference numerals as those in FIG.
  • a sensing device 50 in FIG. 7 is a modification of the microphone 40 shown in FIG.
  • An anomaly detection unit 126 in FIG. 7 is a modification of the anomaly detection unit 116 shown in FIG.
  • the sensing device 50 is an odor sensor
  • the anomaly detection unit 126 is an odor detector.
  • a method for detecting anomalies using a frame image from the camera 20 as the sensing device 50 in addition to the example shown in the first embodiment, an area having a remarkably high brightness is detected.
  • a method for detecting anomalies there is a method for identifying a position by recognizing a fire or explosion by image processing.
  • the video monitoring apparatus 11 may report the detected suspicious candidate information to the security system.
  • the video monitoring apparatus 11 may transmit information to a facility that collects security information such as a central information center.
  • Video monitoring apparatus 101 Human tracking unit 102 Human flow information generation unit 103 Human flow information analysis unit 104, 117, 127 Suspicious person information detection unit 105 Suspicious person information presentation unit 116 Abnormal sound detection unit 126 Abnormality detection unit 20 Camera 30 Display 40 Microphone 50 Sensing device

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Business, Economics & Management (AREA)
  • Emergency Management (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Theoretical Computer Science (AREA)
  • Alarm Systems (AREA)
  • Closed-Circuit Television Systems (AREA)

Abstract

A video monitoring apparatus characterized by being provided with: a person tracking unit that detects persons from a frame image and tracks each of the detected persons over a plurality of frames; a people-flow information generation unit that, from a tracking result from the person tracking unit, recognizes persons, from among the detected persons, whose moving directions and moving speeds are estimated to be close to one another, to be groups, and that generates people-flow information including a group moving direction and a group moving speed for each of the groups; a people-flow information analysis unit that determines from the people-flow information whether an unusual event has occurred in a people flow, and specifies a position when it is determined that an unusual event has occurred; a suspicious person candidate detection unit that detects, as a suspicious person candidate, a person, from among the persons, who is present near the position specified by the people-flow information analysis unit; and a suspicious person candidate presentation unit that presents the detected suspicious person candidate.

Description

映像監視装置及び映像監視方法Image monitoring apparatus and image monitoring method
 この発明は、カメラを用いて不審者などを監視するための映像監視装置に関する。 This invention relates to a video monitoring device for monitoring a suspicious person or the like using a camera.
 近年、店内や施設内及び市街地など、屋内屋外問わず様々な箇所にカメラを設けて撮影された映像を利用して、不審者などを監視するための監視システムが開発されている。特に、イベント会場など人が多く集まる場所での安全対策が求められており、事件など問題が発生したときの不審者の特定や適切な避難誘導などが重要になる。 In recent years, surveillance systems have been developed to monitor suspicious individuals, etc., using images taken with cameras installed in various places, such as in stores, facilities, and urban areas. In particular, safety measures are required in places where many people gather, such as event venues, and it is important to identify suspicious persons and to appropriately guide evacuation when problems such as incidents occur.
 従来は、警備員など監視をする人がカメラ映像を表示したディスプレイを見て、ディスプレイに表示された映像に映る人物から不審者の推定を行ない、推定した不審者の疑いのある人物の行動をカメラ映像または直接現場で確認をすることで不審者の特定を行なうのが一般的であった。ただし、この方法では、人の数やカメラ映像の数が多ければ不審者の特定を行なうのに時間を要するという問題があった。 Conventionally, a security guard or other monitoring person looks at the display on which the camera image is displayed, estimates the suspicious person from the person appearing on the image displayed on the display, and performs the action of the person who is suspected of the estimated suspicious person In general, the suspicious person is identified by checking the camera image or directly in the field. However, this method has a problem that it takes time to identify a suspicious person if the number of people or the number of camera images is large.
 そこで、コンピュータを利用してカメラ映像から監視対象である人の移動軌跡に基づいて予め学習した不審行動モデルとの比較を実施することで異常な行動をしている不審者を特定して識別する方法があった(例えば、特許文献1参照)。 Therefore, by using a computer and comparing with a suspicious behavior model learned in advance based on the movement trajectory of the person being monitored from the camera image, a suspicious person who is performing abnormal behavior is identified and identified. There was a method (for example, refer patent document 1).
特開2008-217602号公報(第4-7頁、第3図)JP 2008-217602 A (page 4-7, FIG. 3)
 しかしながら、特許文献1の方法では、人が多く集まる場所であればカメラ映像に映る人物全ての移動軌跡を追う必要があり、人の数が多ければコンピュータの処理が莫大に必要になるという問題があった。 However, in the method of Patent Document 1, if there are many people gathering, it is necessary to follow the movement trajectory of all the people shown in the camera image. If the number of people is large, the computer processing is enormous. there were.
 この発明は、上述のような課題を解決するためになされたもので、カメラ映像に映る人物の数が多くなったときに処理の増大を抑えた映像監視装置を提供することを目的とする。 The present invention has been made to solve the above-described problems, and an object of the present invention is to provide a video monitoring apparatus that suppresses an increase in processing when the number of persons shown in a camera video increases.
 この発明にかかる映像監視装置においては、フレーム画像から、人物を検出し、検出したそれぞれの前記人物を複数フレームに渡って追跡する人物追跡部と、前記人物追跡部の追跡結果から、前記人物のうち、移動方向及び移動速度が近いと推定する前記人物を集団として認識し、前記集団ごとに集団移動方向及び集団移動速度を含む人流情報を生成する人流情報生成部と、前記人流情報から、人流の異変が発生したか否かを判別し、異変が発生したと判別したときの位置を特定する人流情報解析部と、前記人物のうち、前記人流情報解析部が特定した前記位置の付近に存在する前記人物を不審者候補として検出する不審者候補検出部と、前記検出した不審者候補を提示する不審者候補提示部とを備えることを特徴とするものである。 In the video monitoring apparatus according to the present invention, a person is detected from a frame image, a person tracking unit that tracks each detected person over a plurality of frames, and a tracking result of the person tracking unit, Among them, the human flow information generating unit that recognizes the person estimated to be close in moving direction and moving speed as a group, and generates human flow information including the group moving direction and the group moving speed for each group, A human flow information analysis unit for determining whether or not an abnormal event has occurred, and identifying a position when it is determined that an abnormal event has occurred; among the persons, the human flow information analysis unit is present in the vicinity of the position specified by the human flow information analysis unit A suspicious person candidate detecting unit that detects the person to be detected as a suspicious person candidate, and a suspicious person candidate presenting unit that presents the detected suspicious person candidate.
 この発明は、検出された複数の人物を集団として認識して集団の流れの異変から不審者候補を検出するので、カメラ映像に映る人物の数が多くなっても集団で纏めて把握することができる。このため、カメラ映像に映る人物の数が多くなっても処理の増大を抑えることができるという効果を奏する。 Since this invention recognizes a plurality of detected persons as a group and detects suspicious candidates from the change in the flow of the group, even if the number of persons appearing in the camera video increases, it can be grasped collectively by the group. it can. For this reason, there is an effect that it is possible to suppress an increase in processing even if the number of persons shown in the camera video increases.
実施の形態1にかかる映像監視システムの構成を示すブロック図である。1 is a block diagram showing a configuration of a video monitoring system according to a first exemplary embodiment. 実施の形態1にかかる人物追跡情報D1の一例を示す図である。It is a figure which shows an example of the person tracking information D1 concerning Embodiment 1. FIG. 実施の形態1にかかる人流情報D2の一例を示す図である。It is a figure which shows an example of the human flow information D2 concerning Embodiment 1. FIG. ディスプレイ30に提示する不審者候補提示情報D5の一例を示す図である。It is a figure which shows an example of the suspicious person candidate presentation information D5 shown on the display. 実施の形態1にかかる映像監視装置10の処理を示すフロー図である。FIG. 3 is a flowchart showing processing of the video monitoring apparatus 10 according to the first exemplary embodiment. 実施の形態2にかかる映像監視システムの構成を示すブロック図である。It is a block diagram which shows the structure of the video monitoring system concerning Embodiment 2. FIG. 実施の形態2にかかる映像監視システムの構成変形例を示すブロック図である。It is a block diagram which shows the structural modification of the video surveillance system concerning Embodiment 2. FIG.
実施の形態1.
 図1は、本実施の形態にかかる映像監視システムの構成を示すブロック図である。映像監視装置10は、カメラ20からの映像を入力としてディスプレイ30に不審者候補提示情報D5を提示する。映像監視装置10は、人物追跡部101、人流情報生成部102、人流情報解析部103、不審者情報検出部104、及び不審者情報提示部105を備える。
Embodiment 1 FIG.
FIG. 1 is a block diagram showing the configuration of the video monitoring system according to the present embodiment. The video monitoring apparatus 10 presents the suspicious person candidate presentation information D5 on the display 30 with the video from the camera 20 as an input. The video monitoring apparatus 10 includes a person tracking unit 101, a human flow information generation unit 102, a human flow information analysis unit 103, a suspicious person information detection unit 104, and a suspicious person information presentation unit 105.
 カメラ20は、撮像した画像をフレーム単位で保持する。また、カメラ20は、フレーム単位で映像監視装置10に出力する。以下、撮像した画像をフレーム画像ともいう。また、カメラ20は、人物の動線があるエリアを撮像できる位置に設けられる。なお、「動線」とは、人が自然に動く時に通ると思われる経路を表わした線のことである。 The camera 20 holds captured images in units of frames. Further, the camera 20 outputs to the video monitoring apparatus 10 in units of frames. Hereinafter, the captured image is also referred to as a frame image. The camera 20 is provided at a position where an area with a person's flow line can be imaged. The “flow line” is a line that represents a path that a person is supposed to take when moving naturally.
 人物追跡部101は、カメラ20から入力されたフレーム画像から人物を検出する。また、人物追跡部101は、検出した人物が複数フレームにわたって同一人物として認識した場合、その移動情報を示す人物追跡情報D1を生成する。人物追跡部101は、生成した人物追跡情報D1を人物情報生成部102へ出力する。また、人物追跡部101は、人物検出結果を不審者候補検出部104へ出力する。 The person tracking unit 101 detects a person from the frame image input from the camera 20. Further, when the detected person is recognized as the same person over a plurality of frames, the person tracking unit 101 generates person tracking information D1 indicating the movement information. The person tracking unit 101 outputs the generated person tracking information D1 to the person information generating unit 102. In addition, the person tracking unit 101 outputs the person detection result to the suspicious person candidate detecting unit 104.
 人物追跡部101による人物検出方法としては、事前に多くの人物画像を収集し、その輝度の勾配情報や輪郭情報などの人物特徴を学習する方法がある。この方法では、人物判別器を作成しておき、カメラ20からフレーム画像を取得すると、人物判別器によって、そのフレーム画像中に学習した人物特徴に類似する特徴を持つ領域が含まれているかどうかを判別する。領域が含まれていた場合、その領域の座標(始点座標および縦横の長さ)を特定する。また、HOG(Histogram of oriented Gradients)と呼ばれる特徴量を用いた人物検出方法が一般的であるが、本発明では、このように既知の人物検出手法のいずれかを用いて人物を検出する。 As a person detection method by the person tracking unit 101, there is a method in which a large number of person images are collected in advance and a person feature such as brightness gradient information and contour information is learned. In this method, when a person discriminator is created and a frame image is acquired from the camera 20, it is determined whether or not the person discriminator includes an area having a feature similar to the learned human feature in the frame image. Determine. If an area is included, the coordinates of the area (starting point coordinates and vertical and horizontal lengths) are specified. In addition, a person detection method using a feature amount called HOG (Histogram of orientated gradients) is common, but in the present invention, a person is detected using any of the known person detection methods.
 人物追跡部101による人物追跡方法としては、上述した人物検出方法で人物を検出する際に抽出する人物特徴を利用し、複数フレームにおいて検出した人物の人物特徴同士の類似度が一定の閾値以上となる場合に、それらの人物が同一であるとして判定する方法がある。但し、閾値を超える類似度となる組み合わせが複数ある場合には、そのうちの最も類似度が高いものを同一人物として判定する。 As a person tracking method by the person tracking unit 101, the person feature extracted when the person is detected by the person detection method described above is used, and the similarity between the person features detected in a plurality of frames is equal to or higher than a certain threshold. In such a case, there is a method for determining that these persons are the same. However, when there are a plurality of combinations having a similarity exceeding the threshold, the one with the highest similarity is determined as the same person.
 また、類似度が高い場合であっても、特定された座標が大きく異なる場合には同一人物ではないと判定することができる。例えば、人物追跡部101がカメラ20から連続して取得するフレーム間の時間が33ms、フレーム画像上の1ピクセルに対する実際の距離が0.5mとする。また、上記の人物検出方法によって特定し、その類似度が閾値以上となった人物の始点座標が連続するフレーム間で1000ピクセル離れていたとする。この場合、この人物が同一人物だとすれば33msの間に500m進んだことになり、現実的に不可能であるため、同一人物ではないとして判定することができる。このように、監視システムにおいて設置したカメラ20の位置関係とフレーム画像間の時間間隔との情報から、類似する人物の移動速度が推定できるので、人間の移動速度とは考えにくい移動速度であれば同一人物ではないとして判定する。 Also, even if the degree of similarity is high, it can be determined that they are not the same person if the specified coordinates differ greatly. For example, the time between frames that the person tracking unit 101 continuously acquires from the camera 20 is 33 ms, and the actual distance to one pixel on the frame image is 0.5 m. Also, it is assumed that the start point coordinates of a person specified by the above-described person detection method and whose similarity is equal to or greater than the threshold are 1000 pixels apart between consecutive frames. In this case, if this person is the same person, it will proceed 500 m in 33 ms, which is practically impossible, so it can be determined that they are not the same person. As described above, since the moving speed of a similar person can be estimated from the information on the positional relationship of the camera 20 installed in the monitoring system and the time interval between frame images, the moving speed is unlikely to be considered a human moving speed. It is determined that they are not the same person.
 図2は、人物追跡部101が生成する人物追跡情報D1の一例を示す図である。図2のとおり、人物追跡情報D1には、検出された各々の人物を示す人物番号D1aと、検出フレーム番号D1bと、検出座標D1cと、人物特徴情報D1dとが含まれる。 FIG. 2 is a diagram illustrating an example of the person tracking information D1 generated by the person tracking unit 101. As shown in FIG. 2, the person tracking information D1 includes a person number D1a indicating each detected person, a detection frame number D1b, a detection coordinate D1c, and person feature information D1d.
 人物番号D1aは、人物追跡部101が、異なる人物に対して識別を可能にするために別個に付与する番号である。 The person number D1a is a number that the person tracking unit 101 assigns separately to enable identification of different persons.
 検出フレーム番号D1bは、検出した人物ごとにその人物が映っていたフレーム画像のフレーム番号を示す。フレーム番号は、カメラ20が出力するフレーム画像ごとに割り当てた番号であっても構わない。また、フレーム番号は、人物追跡部101がフレーム画像ごとに割り当てた番号であっても構わない。 The detected frame number D1b indicates the frame number of the frame image in which the detected person is shown for each detected person. The frame number may be a number assigned to each frame image output from the camera 20. The frame number may be a number assigned by the person tracking unit 101 for each frame image.
 検出座標D1cは、人物追跡部101がフレーム画像においてその人物を検出した座標である。座標は例えばフレーム画像について左上のピクセル位置を(0,0)とする2次元軸で表現したものであり、検出した人物の中心位置を推定した位置を検出座標D1cとする。例えば頭部の中心位置など、検出した人物の一部分の位置を検出座標D1cとしても構わない。 The detected coordinates D1c are coordinates at which the person tracking unit 101 detects the person in the frame image. The coordinates are expressed by a two-dimensional axis with the pixel position at the upper left of the frame image as (0, 0), for example, and the position where the center position of the detected person is estimated is detected coordinates D1c. For example, the position of a part of the detected person such as the center position of the head may be used as the detection coordinate D1c.
 人物特徴情報D1dは、人物追跡部101が抽出した人物特徴である。人物特徴は、例えばHOG特徴量であって、フレーム画像を分割した画像領域内における勾配方向(エッジの方向)と勾配強度との関係をヒストグラム化した特徴量を示す。例えば、フレーム画像または人物が映っている画像領域の画像サイズが30ピクセル×60ピクセルの場合、分割した画像領域の単位1セルを5ピクセル×5ピクセルとしてセル単位でのそれぞれの勾配方向に対する勾配強度を求めてヒストグラム化する。ヒストグラム化するにあたって例えば、勾配方向を0[deg]から180[deg]までを20[deg]ずつの9方向に分割するものとする。 The person feature information D1d is a person feature extracted by the person tracking unit 101. The person feature is, for example, a HOG feature amount, and indicates a feature amount obtained by histogramating the relationship between the gradient direction (edge direction) and the gradient strength in the image area obtained by dividing the frame image. For example, when the image size of a frame image or an image area in which a person is shown is 30 pixels × 60 pixels, the gradient intensity for each gradient direction in units of cells, with 1 unit cell of the divided image region being 5 pixels × 5 pixels To obtain a histogram. For example, when the histogram is formed, the gradient direction is divided into 9 directions of 20 [deg] from 0 [deg] to 180 [deg].
 図2においては、フレーム番号1625における人物1の人物特徴は、勾配方向0[deg]での勾配強度が30、勾配方向20[deg]での勾配強度が19、勾配方向40[deg]での勾配強度が17であることを示している。同様に、フレーム番号1173における人物2の人物特徴は、勾配方向0[deg]での勾配強度が10、勾配方向20[deg]での勾配強度が2、勾配方向40[deg]での勾配強度が57であることを示している。 In FIG. 2, the person feature of the person 1 in the frame number 1625 has a gradient strength of 30 in the gradient direction 0 [deg], a gradient strength of 19 in the gradient direction 20 [deg], and a gradient direction of 40 [deg]. It shows that the gradient strength is 17. Similarly, the person feature of the person 2 in the frame number 1173 has a gradient strength of 10 in the gradient direction 0 [deg], a gradient strength of 2 in the gradient direction 20 [deg], and a gradient strength in the gradient direction of 40 [deg]. Is 57.
 人流情報生成部102は、人物追跡情報D1から、検出した人物のうち、移動方向及び移動速度が近い人物を集団として認識し、人流情報D2を生成する。ここで、人流とは、人物追跡部101によって追跡した人物から、同じ方向に同じ速度で移動し、一定の範囲内(近傍)に存在する人物を集団として認識したとき、それぞれの集団の1つ以上のフレームでの移動情報(集団の中心座標、移動方向、移動速度)を表す。集団の中心座標は、集団に含まれる人物の検出座標D1cの平均座標である。 From the person tracking information D1, the human flow information generation unit 102 recognizes, as a group, detected people whose movement direction and movement speed are close, and generates human flow information D2. Here, the human flow means that when a person who moves at the same speed in the same direction from the person tracked by the person tracking unit 101 and recognizes a person existing within a certain range (neighborhood) as a group, it is one of each group. The movement information in the above frame (group center coordinates, movement direction, movement speed) is represented. The center coordinates of the group are average coordinates of the detection coordinates D1c of the persons included in the group.
 図3は、人流情報生成部102から人流情報解析部103へ出力される人流情報D2の一例を示す図である。D2aは集団番号であって異なる集団に対して識別を可能にするために別個に付与する番号である。図3では、人物番号D1aとして1と4と7にあたる人物が一つの集団(集団番号1)として認識されたことを示す。また、集団番号1に該当する集団の中心座標D2cは(275,196)であることを示す。このように、人流情報D2には、それぞれの集団ごとに、含まれる人物の人物番号D2bと、集団の中心座標D2cと、集団移動方向D2dと、集団移動速度D2eとが含まれる。集団移動方向D2dは、特定の座標方向を0[deg]としたときに、連続するフレーム間において移動した方向の角度を表し、例えば集団移動方向D2dの単位を[deg]とする。 FIG. 3 is a diagram illustrating an example of the human flow information D2 output from the human flow information generation unit 102 to the human flow information analysis unit 103. D2a is a group number, which is a number assigned separately to enable identification for different groups. FIG. 3 shows that persons corresponding to person numbers D1a, 1, 4, and 7 are recognized as one group (group number 1). Further, the center coordinate D2c of the group corresponding to the group number 1 is (275, 196). As described above, the human flow information D2 includes the person number D2b of the included person, the center coordinates D2c of the group, the group movement direction D2d, and the group movement speed D2e for each group. The collective movement direction D2d represents an angle of a direction moved between consecutive frames when a specific coordinate direction is set to 0 [deg]. For example, the unit of the collective movement direction D2d is [deg].
 集団移動速度D2eは、連続するフレーム間における集団の中心座標D2cの距離を表し、例えば集団移動速度D2eの単位を1フレーム単位の移動画素距離[ピクセル/フレーム]とする。集団移動速度D2eは、1フレーム間における水平方向の移動画素数の二乗と垂直方向の移動画素数の二乗とを加算したとしてもよい。また、集団移動速度D2eは、1フレーム間における水平方向の移動画素数の二乗と垂直方向の移動画素数の二乗とを加算したものから平方根を求めた値としてもよい。1ピクセルあたりの距離はカメラ20を設けた位置関係から予め設定しておくことで換算することができる。例えば、カメラ20で撮像されたフレーム画像内の床のタイルのピクセル数と実際の床のタイルの距離とから1ピクセルあたりの距離換算の設定を決めることができる。 The collective movement speed D2e represents the distance of the central coordinates D2c of the collective between successive frames. For example, the unit of the collective movement speed D2e is a moving pixel distance [pixel / frame] in one frame unit. The collective moving speed D2e may be obtained by adding the square of the number of moving pixels in the horizontal direction and the square of the number of moving pixels in the vertical direction during one frame. The collective moving speed D2e may be a value obtained by calculating the square root from the sum of the square of the number of moving pixels in the horizontal direction and the square of the number of moving pixels in the vertical direction during one frame. The distance per pixel can be converted by setting in advance from the positional relationship in which the camera 20 is provided. For example, the distance conversion setting per pixel can be determined from the number of floor tile pixels in the frame image captured by the camera 20 and the actual floor tile distance.
 人流情報解析部103は、人流情報生成部102が生成した人流情報D2を解析して異変発生情報D3を生成する。また、異変発生情報D3を不審者候補検出部104へ出力する。 The human flow information analysis unit 103 analyzes the human flow information D2 generated by the human flow information generation unit 102 and generates anomaly occurrence information D3. Also, the incident occurrence information D <b> 3 is output to the suspicious person candidate detection unit 104.
 異変発生情報D3は、人流情報解析部103によって人流情報D2を解析した結果、人流に異変が発生していると判断された場合に、異変が発生したときのフレーム番号を異変発生フレーム番号D3aとし、異変が発生したときの座標を異変発生座標D3bとして含む。 The anomaly occurrence information D3 is the anomaly occurrence frame number D3a when the anomaly occurs when it is determined that an anomaly has occurred in the person flow as a result of analyzing the person flow information D2 by the person flow information analysis unit 103. The coordinates at the time of occurrence of an anomaly are included as the anomaly occurrence coordinates D3b.
 人流情報解析部103が、人流に異変が発生していると判断する方法について説明する。人流の異変は、追跡中の集団における集団移動速度D2eの突発的な変化や、追跡中の集団に属する人物が急激に減少する変化を指す。前者は、集団移動速度D2eの時間変動量が設定した閾値を超えた場合であって、例えば集団移動速度D2eが急激に早くなった場合、その近辺で異変が発生し、逃げる行動をとっていると考えられる。後者の例として、集団に3人が属しており、突如それぞれが異なる方向に移動を始めた場合、集団として認識するための「近傍」もしくは「同一方向」の条件を満たさなくなるため、それまで各々が属していた集団は消滅し、新たに各々が異なる集団として追跡されるようになることが挙げられる。但し、集団が移動によってカメラ20の画角外に移動して映らなくなった場合は、異変が発生していると判断しない。つまり、前者は集団のまま、同一の方向に逃げる場合であり、後者は一斉に散らばるように逃げる場合である。 A method will be described in which the human flow information analysis unit 103 determines that an anomaly has occurred in the human flow. An anomaly in the human flow indicates a sudden change in the group movement speed D2e in the group being tracked or a change in which the number of persons belonging to the group being tracked decreases rapidly. The former is a case where the time fluctuation amount of the collective movement speed D2e exceeds a set threshold value. For example, when the collective movement speed D2e is rapidly increased, an abnormality occurs in the vicinity of the collective movement speed D2e. it is conceivable that. As an example of the latter, if three people belong to a group and suddenly start moving in different directions, the conditions of “neighboring” or “same direction” for recognizing the group will not be met. The group that belonged to disappears and each is newly tracked as a different group. However, if the group moves out of the angle of view of the camera 20 due to the movement and no longer appears, it is not determined that an abnormality has occurred. That is, the former is a case of running away in the same direction as a group, and the latter is a case of running away so as to be scattered all at once.
 人流情報解析部103は、急激な変化を検知することで人流に異変が発生していると判断するため、例えば、集団に属する人物の1人が、速度は変えずに異なる方向に移動を始めた場合は、基本的に、これを異変として判断しない。この場合、単に目的地が異なるため、異なる方向に移動始めた可能性が高い。一方、異変が発生した場合は、その時点およびそれ以降において集団移動速度D2eが非常に大きくなると考えられる。 Since the human flow information analysis unit 103 detects that an abnormal change has occurred in the human flow by detecting a sudden change, for example, one of the persons belonging to the group starts moving in different directions without changing the speed. If this is the case, basically, this is not judged as an incident. In this case, since the destinations are simply different, there is a high possibility that they have started moving in different directions. On the other hand, when an abnormality occurs, the collective movement speed D2e is considered to become very large at that time and thereafter.
 人流情報解析部103は、異変ありと判断した場合に異変発生情報D3を生成して不審者候補検出部104に出力する。一方、人流情報解析部103は、異変なしと判断した場合には、異変発生情報D3を不審者候補検出部104に出力せずに、人流情報D2が更新されるまで待機する。 When the person flow information analysis unit 103 determines that there is an abnormality, it generates the abnormality occurrence information D3 and outputs it to the suspicious person candidate detection unit 104. On the other hand, when it is determined that there is no abnormality, the human flow information analysis unit 103 does not output the abnormal occurrence occurrence information D3 to the suspicious person candidate detection unit 104 and waits until the human flow information D2 is updated.
 不審者候補検出部104は、人物追跡部101が生成した人物追跡情報D1と人流情報解析部103が生成した異変発生情報D3とを解析して不審者候補を検出し、不審者候補情報D4を生成する。また、不審者候補検出部104は、不審者候補情報D4を不審者候補提示部105へ出力する。ここで、人流情報解析部103から異変発生情報D3が不審者候補検出部104に入力されたときは異変ありとの通知があったと推定することができる。このことから人流情報解析部103から異変発生情報D3が不審者候補検出部104に入力されたとき、または異変発生情報D3に変更があった場合にのみ不審者候補検出部104は動作を行なうものとしてもよい。 The suspicious candidate detection unit 104 detects the suspicious candidate candidate by analyzing the person tracking information D1 generated by the person tracking unit 101 and the anomaly occurrence information D3 generated by the human flow information analyzing unit 103, and the suspicious candidate information D4 is detected. Generate. Further, the suspicious person candidate detecting unit 104 outputs the suspicious person candidate information D4 to the suspicious person candidate presenting part 105. Here, when the anomaly occurrence information D3 is input from the human flow information analysis unit 103 to the suspicious person candidate detection unit 104, it can be estimated that there has been a notification of an anomaly. Therefore, the suspicious candidate detection unit 104 operates only when the incident occurrence information D3 is input from the human flow information analysis unit 103 to the suspicious person candidate detection unit 104 or when the incident change information D3 is changed. It is good.
 不審者候補情報D4は、不審者候補検出部104が不審者候補と検出した人物について、人物追跡情報D1において割り当てた人物番号D1a、その人物に関する情報である検出フレーム番号D1b、検出座標D1c、及び人物特徴情報D1dから抽出したものを含む。 The suspicious person candidate information D4 includes a person number D1a assigned in the person tracking information D1 for a person detected by the suspicious person candidate detecting unit 104 as a suspicious person candidate, a detection frame number D1b that is information related to the person, a detection coordinate D1c, The information extracted from the person feature information D1d is included.
 不審者候補検出部104は、異変発生座標D3bと人物の検出座標D1cとを比較し、異変発生座標D3bから予め設定した閾値以内の距離に存在する人物を、不審者候補として特定する。又は、不審者候補検出部104は、異変発生座標D3bと人物の検出座標D1cとを比較し、異変発生座標D3bに最も近い距離に存在する人物を不審者候補として特定する。あるいは、不審者候補検出部104は、異変発生座標D3bに近い人物から順に、予め設定した人数だけ不審者候補として特定する。 The suspicious person candidate detecting unit 104 compares the anomaly occurrence coordinates D3b with the person detection coordinates D1c, and identifies a person existing at a distance within a preset threshold from the anomaly occurrence coordinates D3b as a suspicious person candidate. Alternatively, the suspicious person candidate detection unit 104 compares the anomaly occurrence coordinates D3b and the person detection coordinates D1c, and identifies a person existing at a distance closest to the anomaly occurrence coordinates D3b as a suspicious person candidate. Or the suspicious person candidate detection part 104 specifies only the preset number of persons as a suspicious person candidate in an order from the person close | similar to the anomaly occurrence coordinate D3b.
 予め設定した距離の閾値及び不審者候補数の閾値は、任意で決定することができる。例えば不審者候補を出来るだけ多く検出し、警備員の目視によって最終的に特定したい場合には距離を長くするよう閾値を設定したり、不審者候補の人数を多くするよう閾値を設定したりすることができる。出来るだけ候補を絞り込みたい場合には、距離を短くするよう閾値を設定したり、人数を1人に絞るような閾値に設定することができる。 The preset threshold for distance and threshold for the number of suspicious candidates can be determined arbitrarily. For example, if you want to detect as many suspicious candidates as possible and ultimately identify them by visual inspection of guards, set a threshold to increase the distance, or set a threshold to increase the number of suspicious candidates be able to. When it is desired to narrow down candidates as much as possible, a threshold value can be set to shorten the distance, or a threshold value can be set to reduce the number of people to one.
 不審者候補提示部105は、不審者候補検出部104が生成した不審者候補情報D4と、カメラ20から出力されたフレーム画像とを入力し、不審者候補提示情報D5を生成してディスプレイ30に出力する。 The suspicious candidate presentation unit 105 inputs the suspicious candidate information D4 generated by the suspicious candidate detection unit 104 and the frame image output from the camera 20, and generates suspicious candidate presentation information D5 on the display 30. Output.
 ディスプレイ30は、不審者候補提示部105が生成した不審者候補提示情報D5を入力し、表示する。 The display 30 inputs and displays the suspicious candidate presentation information D5 generated by the suspicious candidate presentation unit 105.
 図4は、ディスプレイ30に提示する不審者候補提示情報D5の一例を示す図である。図4では、カメラ20から出力されたフレーム画像に映る人物のうち、不審者候補と検出した人物を矩形で囲む表示を行なっている。このようにディスプレイ30に提示することで、警備員が一目で不審者候補の有無やその位置を認識しやすくなる。 FIG. 4 is a diagram showing an example of the suspicious candidate presentation information D5 presented on the display 30. As shown in FIG. In FIG. 4, among the persons shown in the frame image output from the camera 20, a person who is detected as a suspicious candidate is displayed in a rectangle. By presenting the information on the display 30 in this manner, it is easy for the security guard to recognize at a glance whether or not there is a suspicious candidate and its position.
 次に映像監視装置10の処理の流れを説明する。図5は、映像監視装置10の処理を示すフロー図である。 Next, the processing flow of the video monitoring apparatus 10 will be described. FIG. 5 is a flowchart showing processing of the video monitoring apparatus 10.
 まず、人物追跡部101は、集団追跡処理を実行していない新たなフレーム画像をカメラ20が保持している場合(S001:Yes)、このフレーム画像を取得する(S002)。また、人物追跡部101は、新たなフレーム画像をカメラ20が保持していない場合(S001:No)、カメラ20が新たなフレーム画像を保持するまで待機する。 First, when the camera 20 holds a new frame image that has not been subjected to the group tracking process (S001: Yes), the person tracking unit 101 acquires this frame image (S002). If the camera 20 does not hold a new frame image (S001: No), the person tracking unit 101 waits until the camera 20 holds a new frame image.
 人物追跡部101は、フレーム画像を取得した後(S002)、このフレーム画像に対して人物検出処理を実行する(S003)。 The person tracking unit 101 acquires a frame image (S002), and then performs a person detection process on the frame image (S003).
 人物が検出できた場合(S004:Yes)、人物追跡部101は、検出した人物の人物特徴について、前のフレームで検出した人物の特徴との類似度を求める。但し、前のフレームで人物を検出していない場合、人物追跡部101は、この処理(S005)を行わない。また、人物が検出出来なかった場合、人物追跡部101は、カメラ20が新たなフレーム画像を保持するまで待機する(S004:No)。 When a person can be detected (S004: Yes), the person tracking unit 101 obtains a similarity between the detected person feature of the person and the feature of the person detected in the previous frame. However, if no person is detected in the previous frame, the person tracking unit 101 does not perform this process (S005). If no person can be detected, the person tracking unit 101 waits until the camera 20 holds a new frame image (S004: No).
 前フレームとの比較によって類似度が高い人物を検出した場合、人物追跡部101は、同一の人物として判定し、人物追跡情報D1を生成する(S006:Yes)。類似度が高い人物を検出できなかった場合、人物追跡部101は、カメラ20が新たなフレーム画像を保持するまで待機する(S006:No)。 When a person with a high degree of similarity is detected by comparison with the previous frame, the person tracking unit 101 determines that the person is the same person and generates person tracking information D1 (S006: Yes). If a person with high similarity cannot be detected, the person tracking unit 101 waits until the camera 20 holds a new frame image (S006: No).
 人流情報生成部102は、人物追跡部101から人物追跡情報D1を受信し、追跡している各々の人物の移動方向と移動速度とを算出する(S007)。 The human flow information generation unit 102 receives the person tracking information D1 from the person tracking unit 101, and calculates the moving direction and moving speed of each person being tracked (S007).
 人流情報生成部102は更に、追跡している各々の人物について、一定の範囲内(近傍)に存在し、かつ前記算出した移動方向と移動速度とが該当する人物を集団として認識し、集団に含まれる人物番号D2bと、集団の中心座標D2cと、集団移動方向D2dと、集団移動速度D2eとを含む人流情報D2を生成し、人流情報解析部103に送信する(S008)。 The human flow information generation unit 102 further recognizes, as a group, a person who is within a certain range (near) for each person being tracked and corresponds to the calculated moving direction and moving speed. The human flow information D2 including the person number D2b, the central coordinate D2c of the group, the group movement direction D2d, and the group movement speed D2e is generated and transmitted to the human flow information analysis unit 103 (S008).
 人流情報解析部103は、人流情報生成部102から人流情報D2を受信し、人流における異変発生の有無を判断する(S009)。 The human flow information analysis unit 103 receives the human flow information D2 from the human flow information generation unit 102, and determines whether or not an anomaly has occurred in the human flow (S009).
 人流情報解析部103は、異変ありと判断した場合、人流解析部13は異変発生情報D3を不審者候補検出部104に送信する(S009:Yes)。人流情報解析部103は、異変なしと判断した場合、カメラ20が新たなフレーム画像を保持するまで待機する(S009:No)。 When the human flow information analysis unit 103 determines that there is an abnormality, the human flow analysis unit 13 transmits the abnormality occurrence information D3 to the suspicious person candidate detection unit 104 (S009: Yes). When it is determined that there is no change, the human flow information analysis unit 103 waits until the camera 20 holds a new frame image (S009: No).
 不審者候補検出部104は、人流情報解析部103から異変発生情報D3を受信する。また、不審者候補検出部104は、人物追跡部101から人物追跡情報D1を受信する。その後、不審者候補検出部104は、異変発生座標D3bと人物の検出座標D1cとを比較し、異変発生座標D3bから一定の閾値以内の距離に存在する人物を、不審者候補として特定する(S010)。又は、不審者候補検出部104は、異変発生座標D3bと人物の検出座標D1cとを比較し、異変発生座標D3bに最も近い距離に存在する人物を不審者候補として特定する。あるいは、不審者候補検出部104は、異変発生座標D3bに近い人物から順に、一定の人数だけ不審者候補として特定する。 The suspicious person candidate detection unit 104 receives the anomaly occurrence information D3 from the human flow information analysis unit 103. Further, the suspicious person candidate detection unit 104 receives the person tracking information D <b> 1 from the person tracking unit 101. Thereafter, the suspicious person candidate detection unit 104 compares the anomaly occurrence coordinate D3b with the person detection coordinate D1c, and identifies a person existing at a distance within a certain threshold from the anomaly occurrence coordinate D3b as a suspicious person candidate (S010). ). Alternatively, the suspicious person candidate detection unit 104 compares the anomaly occurrence coordinates D3b and the person detection coordinates D1c, and identifies a person existing at a distance closest to the anomaly occurrence coordinates D3b as a suspicious person candidate. Alternatively, the suspicious person candidate detection unit 104 specifies a certain number of suspicious person candidates as a suspicious person candidate in order from a person close to the anomaly occurrence coordinate D3b.
 以上の処理により、不審者候補検出部104は、不審者候補情報D4を生成し、不審者候補提示部105に送信する。 Through the above processing, the suspicious person candidate detecting unit 104 generates suspicious person candidate information D4 and transmits it to the suspicious person candidate presenting part 105.
 不審者候補提示部105は、不審者候補検出部104から不審者候補情報D4を受信する。また、不審者候補提示部105は、カメラ20からフレーム画像を取得する。また、不審者候補提示部105は、図5に例示するような不審者候補提示情報D5を生成し、ディスプレイ30に送信し、ディスプレイ30がそれを表示する(S011)。なお、図5では、カメラ20から受信したフレーム画像に対し、不審者候補として判断した人物を矩形で囲むなどして、警備員が一目で不審者候補の有無やその位置を認識し易いようにしている。 The suspicious person candidate presentation unit 105 receives the suspicious person candidate information D4 from the suspicious person candidate detection unit 104. Further, the suspicious person candidate presenting unit 105 acquires a frame image from the camera 20. Further, the suspicious person candidate presenting unit 105 generates suspicious person candidate presentation information D5 as illustrated in FIG. 5 and transmits it to the display 30, and the display 30 displays it (S011). In FIG. 5, the frame image received from the camera 20 is surrounded by a rectangle of a person who is determined as a suspicious candidate so that the guard can easily recognize the presence or position of the suspicious candidate at a glance. ing.
 これにより、本実施の形態にかかる映像監視装置10は、検出された複数の人物を集団として認識して集団の流れの異変から不審者候補を検出するので、カメラ映像に映る人物の数が多くなっても集団で纏めて把握することができる。このことから、映像監視装置10は、カメラ映像に映る人物の数が多くなっても処理の増大を抑えることができるという効果を奏する。 Thereby, since the video monitoring apparatus 10 according to the present embodiment recognizes a plurality of detected persons as a group and detects suspicious candidate candidates from the change in the flow of the group, the number of persons reflected in the camera video is large. Even then, it can be grasped collectively by the group. From this, the video monitoring apparatus 10 has an effect that it is possible to suppress an increase in processing even if the number of persons shown in the camera video increases.
 また、映像監視装置10は、異変発生時、異変発生地点から不審者が逃亡したり、その周辺の人物が逃げたりする行動によって起こる人流の乱れを検知することができる。また、映像監視装置10は、その異変が生じた地点の周辺に存在する不審者候補を自動的に検出し、候補とされた人物を追跡してディスプレイ上に表示することができる。このため、本発明における映像監視装置10は、事前に学習した不審者行動パターンによらない動きをする不審者についても、不審者候補として検出することができる。これにより、警備員は映像監視装置における大量の映像データの中から、不審者候補の位置や逃走経路を迅速に把握することが可能になるため、適切な警備を迅速に実行することが出来る。 In addition, the video monitoring device 10 can detect the disturbance of the human flow caused by the behavior of the suspicious person escaping from the incident occurrence point or the surrounding person escaping when the incident occurs. In addition, the video monitoring apparatus 10 can automatically detect suspicious candidate candidates that exist in the vicinity of the point where the abnormality has occurred, and can track the candidate person and display it on the display. For this reason, the video monitoring apparatus 10 according to the present invention can detect a suspicious person who moves based on a suspicious person behavior pattern learned in advance as a suspicious person candidate. Accordingly, the security guard can quickly grasp the position of the suspicious candidate candidate and the escape route from a large amount of video data in the video surveillance device, and thus can perform appropriate security promptly.
 上述の不審者候補検出部104は、人流の異変情報から不審者候補を検出することとしているが、カメラのフレーム画像を利用した画像認識技術によってマスクやサングラスなどの変装用の装備や、凶器を検出することで、装備情報を不審者候補の検出に利用することができる。例えば、変装用の装備または凶器を所持する人物は不審度が高いとして判定することができる。この場合、不審度が高い人物には不審者候補提示部105が生成する不審者候補提示情報D5における矩形の色をより目立つ色にする。これにより、警備員はより迅速に不審者の位置を把握することが可能になる。 The above-mentioned suspicious candidate detection unit 104 detects suspicious candidates from human-flow anomaly information. However, disguise equipment such as masks and sunglasses and weapons are used by image recognition technology using the frame image of the camera. By detecting, equipment information can be used for detection of a suspicious candidate. For example, a person possessing disguise equipment or a weapon can be determined as having a high degree of suspiciousness. In this case, the color of the rectangle in the suspicious person candidate presentation information D5 generated by the suspicious person candidate presentation unit 105 is made more conspicuous for a person with a high suspicious degree. As a result, the guard can grasp the position of the suspicious person more quickly.
 また、不審者候補検出部104は、カメラのフレーム画像を利用した画像認識技術によって年齢推定を行う。このことで、不審者候補検出部104は、推定年齢の情報を不審者候補の検出に利用することができる。例えば、10歳の子供が不審者である可能性は低いと考えられる。このため、このような人物はより不審度が低いとして、不審者候補検出部104は、不審者候補提示情報D5における矩形の色を目立たない色にする。または、不審者候補検出部104は、このような人物を不審者候補として表示しないようにする。このようにすることで、不審者候補を絞り込むことができるため、警備員はより迅速に真の不審者の位置を把握することが可能になる。 In addition, the suspicious person candidate detection unit 104 performs age estimation by an image recognition technique using a frame image of the camera. Thus, the suspicious person candidate detection unit 104 can use the information on the estimated age for detecting the suspicious person candidate. For example, it is considered unlikely that a 10-year-old child is a suspicious person. For this reason, assuming that such a person has a lower suspicious degree, the suspicious person candidate detection unit 104 changes the rectangular color in the suspicious person candidate presentation information D5 to an inconspicuous color. Alternatively, the suspicious person candidate detection unit 104 does not display such a person as a suspicious person candidate. By doing in this way, since a suspicious candidate can be narrowed down, a guard can grasp a true suspicious person's position more quickly.
 また、不審者候補検出部104は、人物の位置関係と移動状態とから不審者候補を絞り込むことが出来る。例えば、異変発生後、周辺の不審者ではない人物は、不審者から逃げる行動をとると考えられる。このため、異変発生後に同じ方向に移動する人物のうち、より異変発生地点から遠い位置に存在する人物は、不審者ではない可能性が高い。不審者が不審者でない人物を追うことはあっても、その逆はあり得ないと考えられるからである。このようにすることで、不審者候補検出部104は、不審者候補を絞り込むことができる。このため、警備員はより迅速に真の不審者の位置を把握することが可能になる。 In addition, the suspicious person candidate detection unit 104 can narrow down the suspicious person candidates from the positional relationship and the movement state of the person. For example, it is considered that a person who is not a suspicious person in the vicinity takes an action of escaping from a suspicious person after the occurrence of an incident. For this reason, among the persons who move in the same direction after the occurrence of an anomaly, there is a high possibility that the person who is further away from the anomaly occurrence point is not a suspicious person. This is because even if a suspicious person pursues a person who is not a suspicious person, the opposite is not possible. By doing in this way, the suspicious person candidate detection part 104 can narrow down a suspicious person candidate. For this reason, the guard can grasp the position of the true suspicious person more quickly.
 また、不審者候補検出部104は、不審者候補の移動速度が特定の閾値以下の場合には、逃げる行動を取っていない可能性が高いとして、不審者候補ではないと判定しても良い。このようにすることで、不審者候補検出部104は、不審者候補を絞り込むことができる。このため、警備員はより迅速に真の不審者の位置を把握することが可能になる。 Further, when the moving speed of the suspicious candidate candidate is equal to or lower than a specific threshold, the suspicious candidate detection unit 104 may determine that the suspicious candidate candidate is not a suspicious candidate because it is highly likely that the escaping action is not taken. By doing in this way, the suspicious person candidate detection part 104 can narrow down a suspicious person candidate. For this reason, the guard can grasp the position of the true suspicious person more quickly.
実施の形態2.
 図6は、実施の形態2にかかる映像監視システムの構成を示すブロック図である。映像監視装置11は、実施の形態2にかかる映像監視方法を実施することができる装置である。図6において、図1に示される構成要素と同一又は対応する構成要素には、図1における符号と同じ符号を付している。実施の形態2における映像監視装置11は、不審者候補検出部117が異音検知部116に接続されている。また、実施の形態2における映像監視装置11は、異音検知部116がマイク40に接続されている。実施の形態2にかかる映像監視装置11は、不審者候補を検出するために異音検知を利用する点において、実施の形態1にかかる映像監視装置10と異なる。この点を除き、実施の形態2は、実施の形態1と同じである。
Embodiment 2. FIG.
FIG. 6 is a block diagram of the configuration of the video monitoring system according to the second embodiment. The video monitoring apparatus 11 is an apparatus that can perform the video monitoring method according to the second embodiment. In FIG. 6, the same or corresponding components as those shown in FIG. 1 are denoted by the same reference numerals as those in FIG. In the video monitoring apparatus 11 according to the second embodiment, the suspicious person candidate detection unit 117 is connected to the abnormal sound detection unit 116. In the video monitoring apparatus 11 according to the second embodiment, the abnormal sound detection unit 116 is connected to the microphone 40. The video monitoring apparatus 11 according to the second embodiment is different from the video monitoring apparatus 10 according to the first embodiment in that abnormal sound detection is used to detect a suspicious candidate. Except for this point, the second embodiment is the same as the first embodiment.
 マイク40は、音声データを異音検知部116へ送信する。このマイク40は1台のマイクでもよいし、複数台用いて音源方向を高精度に抽出できるように構成してもよい。 The microphone 40 transmits audio data to the abnormal sound detection unit 116. The microphone 40 may be a single microphone, or a plurality of microphones 40 may be used so that the direction of the sound source can be extracted with high accuracy.
 異音検知部116は、マイク40から音声データを取得し、音声データを分析する。例えば、異音検知部116は、音声データから「会話」「BGM」「足音」「悲鳴」「怒号」などに音声データを分類する。異音検知部116は、分類された音声データから「悲鳴」や「怒号」といった生活音などと異なる、事件に関連性の高い音声を検出する。音声データの分析は周波数(声の高さ)や振幅(声の大きさ)をそれぞれの分類におけるモデルデータと比較し、相関を取ることで、その分類を判別するなどできる。 The abnormal sound detection unit 116 acquires audio data from the microphone 40 and analyzes the audio data. For example, the abnormal sound detection unit 116 classifies the voice data from the voice data into “conversation”, “BGM”, “footstep”, “scream”, “anger”, and the like. The abnormal sound detection unit 116 detects a sound highly relevant to the incident, which is different from the life sound such as “scream” or “anger” from the classified sound data. In the analysis of voice data, the frequency (voice pitch) and amplitude (voice volume) can be compared with model data in each classification, and the classification can be discriminated by taking a correlation.
 異音検知部116は、音声データの分類により「悲鳴」や「怒号」を検出するとともに、音源の方向推定を行う。音源方向推定技術も近年盛んに研究開発がされており、同様の方式をとることができる。複数台のマイクでの音源方向を推定するには、それぞれのマイクで集音した音声データの時間差を検出する方式等がある。 The abnormal sound detection unit 116 detects “scream” and “anger” by the classification of the voice data, and also estimates the direction of the sound source. The sound source direction estimation technology has been actively researched and developed in recent years, and the same method can be adopted. In order to estimate the sound source direction of a plurality of microphones, there is a method of detecting a time difference between audio data collected by each microphone.
 異音検知部116は、「悲鳴」や「怒号」を検出し、それらの音源の方向推定を行った結果として、その位置を不審者候補検出部117に送信する。この位置は、予めマイク40での推定位置と、カメラ20で撮像する座標系との対応を調査しておき、マイクで推定した位置をカメラ座標系での位置に置き換えて送信する。不審者候補検出部117の動作及び作用は、実施の形態1の内容と同じため、説明を省略する。 The abnormal sound detection unit 116 detects “scream” and “anger”, and transmits the position to the suspicious candidate detection unit 117 as a result of estimating the direction of the sound source. For this position, the correspondence between the estimated position in the microphone 40 and the coordinate system imaged by the camera 20 is investigated in advance, and the position estimated by the microphone is replaced with the position in the camera coordinate system and transmitted. Since the operation and action of the suspicious candidate detection unit 117 are the same as those of the first embodiment, description thereof is omitted.
 以上に説明したように、本実施の形態にかかる映像監視装置11においては、異音が発生した位置を異音検知によって特定する。このことから、人が入り組んで移動するような激しい混雑状態にあって、実施の形態1で説明した人流の変異をカメラによって捉えられない場合にも、映像監視装置11は、異変が発生した位置を特定でき、その周辺から不審者候補を検出することができる。これにより、警備員は大量の映像監視装置による映像データの中から、不審者を特定し、その行動を追跡することを可能にでき、迅速に状況を把握し、適切な警備を実行することが出来る。 As described above, in the video monitoring apparatus 11 according to the present embodiment, the position where the abnormal sound has occurred is specified by detecting the abnormal sound. Therefore, even in a crowded state in which a person moves intricately and the variation in the human flow described in the first embodiment cannot be captured by the camera, the video monitoring device 11 is located at the position where the abnormality has occurred. Can be identified, and suspicious candidates can be detected from the surrounding area. This makes it possible for security guards to identify suspicious individuals from a large amount of video data from video surveillance devices and to track their actions, quickly grasp the situation, and execute appropriate security. I can do it.
 図7は、本発明の実施の形態1及び2にかかる映像監視装置の変形例(映像監視装置12)を示すブロック図である。図7において、図1に示される構成要素と同一又は対応する構成要素には、図1における符号と同じ符号を付している。図7におけるセンシング装置50は、図6で示されたマイク40の変形例である。図7における異変検知部126は、図6で示された異音検知部116の変形例である。例えばセンシング装置50は、臭いセンサーであり、異変検知部126は、異臭検知器である。また、センシング装置50としてカメラ20からのフレーム画像を流用した異変検知の方法として、実施の形態1に示した例の他にも、爆発物から生じる閃光を、輝度が著しく高いエリアを検出することで認識し、位置を特定する方法がある。また、異変検知の方法として、火災や爆発を画像処理により認識して、位置を特定する方法もある。更に、前記のように単体のセンサーを利用するだけでなく、複数のセンサーを利用し、それらセンサーデータの解析の結果を複合的に勘案して異常を検知し、位置を特定する方法もある。このようにすることで、映像監視装置12は、より精度良く異常が発生した位置を特定でき、不審者の推定が実現できる。 FIG. 7 is a block diagram showing a modification (video monitoring device 12) of the video monitoring device according to the first and second embodiments of the present invention. 7, components that are the same as or correspond to the components shown in FIG. 1 are given the same reference numerals as those in FIG. A sensing device 50 in FIG. 7 is a modification of the microphone 40 shown in FIG. An anomaly detection unit 126 in FIG. 7 is a modification of the anomaly detection unit 116 shown in FIG. For example, the sensing device 50 is an odor sensor, and the anomaly detection unit 126 is an odor detector. In addition to the example shown in the first embodiment, as a method for detecting anomalies using a frame image from the camera 20 as the sensing device 50, in addition to the example shown in the first embodiment, an area having a remarkably high brightness is detected. There is a method of recognizing and specifying the position. In addition, as a method for detecting anomalies, there is a method for identifying a position by recognizing a fire or explosion by image processing. Furthermore, there is a method of not only using a single sensor as described above but also using a plurality of sensors, detecting an abnormality by combining the results of analysis of the sensor data, and specifying the position. By doing in this way, the video monitoring apparatus 12 can specify the position where the abnormality has occurred with higher accuracy, and can estimate the suspicious person.
 本実施の形態にかかる映像監視装置11は、検出した不審者候補の情報を警備システムへ通報してもよい。また、映像監視装置11は、中央情報センター等のセキュリティ情報を集約する施設へ、情報を送信するようにしてもよい。 The video monitoring apparatus 11 according to the present embodiment may report the detected suspicious candidate information to the security system. In addition, the video monitoring apparatus 11 may transmit information to a facility that collects security information such as a central information center.
  10、11、12    映像監視装置
  101         人物追跡部
  102         人流情報生成部
  103         人流情報解析部
  104、117、127 不審者情報検出部
  105         不審者情報提示部
  116         異音検知部
  126         異変検知部
  20          カメラ
  30          ディスプレイ
  40          マイク
  50          センシング装置
10, 11, 12 Video monitoring apparatus 101 Human tracking unit 102 Human flow information generation unit 103 Human flow information analysis unit 104, 117, 127 Suspicious person information detection unit 105 Suspicious person information presentation unit 116 Abnormal sound detection unit 126 Abnormality detection unit 20 Camera 30 Display 40 Microphone 50 Sensing device

Claims (14)

  1.  フレーム画像から、人物を検出し、検出したそれぞれの前記人物を複数フレームに渡って追跡する人物追跡部と、
     前記人物追跡部の追跡結果から、前記人物のうち、移動方向及び移動速度が近いと推定する前記人物を集団として認識し、前記集団ごとに集団移動方向及び集団移動速度を含む人流情報を生成する人流情報生成部と、
     前記人流情報から、人流の異変が発生したか否かを判別し、異変が発生したと判別したときの位置を特定する人流情報解析部と、
     前記人物のうち、前記人流情報解析部が特定した前記位置の付近に存在する前記人物を不審者候補として検出する不審者候補検出部と、
     前記検出した不審者候補を提示する不審者候補提示部と、
    を備えることを特徴とする映像監視装置。
    A person tracking unit for detecting a person from a frame image and tracking each detected person over a plurality of frames;
    Based on the tracking result of the person tracking unit, the person who is estimated to be close in moving direction and moving speed is recognized as a group, and human flow information including the group moving direction and the group moving speed is generated for each group. A human flow information generator,
    From the human flow information, it is determined whether or not a human flow anomaly has occurred, and a human flow information analysis unit that identifies a position when it is determined that an anomaly has occurred,
    Among the persons, a suspicious person candidate detection unit that detects the person existing near the position specified by the human flow information analysis unit as a suspicious person candidate;
    A suspicious candidate presentation section for presenting the detected suspicious candidate,
    A video surveillance apparatus comprising:
  2.  前記人流情報生成部は、認識した前記集団のうち、前記集団に属するそれぞれの前記人物の前記移動方向がそれぞれ異なる方向を取ったと推定した場合に、前記人流情報から前記集団が消滅させて新たに異なる集団として認識して新たな前記人流情報を生成し、
     前記人流情報解析部は、前記人流情報から前記集団が消滅した場合に、異変が発生した
    と判別する
    ことを特徴とする請求項1に記載の映像監視装置。
    When the person flow information generation unit estimates that the movement direction of each person belonging to the group has taken a different direction among the recognized group, the group disappears from the person flow information and newly Recognize as a different group and generate new human flow information,
    The video monitoring apparatus according to claim 1, wherein the human flow information analysis unit determines that an abnormality has occurred when the group disappears from the human flow information.
  3.  前記人流情報解析部は、前記集団移動速度の時間変動量が予め設定した閾値を超えた場合に、異変が発生したと判別する
    ことを特徴とする請求項1または請求項2に記載の映像監視装置。
    The video surveillance according to claim 1, wherein the human flow information analysis unit determines that an anomaly has occurred when a temporal variation amount of the collective movement speed exceeds a preset threshold value. apparatus.
  4.  前記不審者候補検出部は、前記人物のうち、前記人流情報解析部が特定した前記位置に最も近く存在する前記人物を前記不審者候補として検出する
    ことを特徴とする請求項1から3のいずれか1項に記載の映像監視装置。
    The suspicious individual candidate detection unit detects the person who is closest to the position specified by the human flow information analysis unit among the persons as the suspicious individual candidate. The video monitoring apparatus according to claim 1.
  5.  前記不審者候補検出部は、前記人物のうち、前記人流情報解析部が特定した前記位置から近くに存在している順に、予め設定した人数分を前記不審者候補として検出する
    ことを特徴とする請求項1から4のいずれか1項に記載の映像監視装置。
    The suspicious person candidate detecting unit detects a predetermined number of persons as the suspicious person candidates in the order in which the person is present near the position specified by the human flow information analyzing unit. The video monitoring apparatus according to any one of claims 1 to 4.
  6.  前記不審者候補検出部は、前記人物のうち、前記人流情報解析部が特定した前記位置から特定の閾値以内の距離に存在する前記人物を前記不審者候補として検出する
    ことを特徴とする請求項1から5のいずれか1項に記載の映像監視装置。
    The suspicious person candidate detecting unit detects the person existing at a distance within a specific threshold from the position specified by the human flow information analyzing unit among the persons as the suspicious person candidate. The video monitoring apparatus according to any one of 1 to 5.
  7.  前記不審者候補検出部は、前記人流情報解析部が特定した前記位置の周辺で複数の前記不審者候補の前記移動方向が同じ場合は、前記人物のうち、前記人流情報解析部が特定した前記位置に近い側の前記人物を前記不審者候補として検出する
    ことを特徴とする請求項1から6のいずれか1項に記載された映像監視装置。
    The suspicious person candidate detection unit, when the movement direction of the plurality of suspicious person candidates is the same around the position specified by the human flow information analysis unit, among the persons, the human flow information analysis unit specified The video monitoring apparatus according to claim 1, wherein the person close to the position is detected as the suspicious person candidate.
  8.  前記不審者候補検出部は、前記人物のうち、前記移動速度が予め設定した閾値以下の前記人物を前記不審者候補から除外する
    ことを特徴とする請求項1から7のいずれか1項に記載された映像監視装置。
    The suspicious person candidate detection unit excludes the person whose moving speed is equal to or less than a preset threshold among the persons from the suspicious person candidate. Video surveillance device.
  9.  前記不審者候補検出部は、前記映像データから前記人物の年齢情報を推定し、前記検出した不審者候補の絞り込みに用いる
    ことを特徴とする請求項1から8のいずれか1項に記載された映像監視装置。
    The said suspicious person candidate detection part estimates the age information of the said person from the said video data, and uses it for narrowing down the detected suspicious person candidate, The one described in any one of Claim 1 to 8 characterized by the above-mentioned. Video surveillance device.
  10.  音声データを収集して悲鳴または怒号を含む特定音を検出し、検出した前記特定音の音源方向の推定を行う異音検知部をさらに備え、
     前記不審者候補検出部は、前記異音検知部で検出した前記特定音及び推定した前記音源方向を用いて前記不審者候補を検出する
    ことを特徴とする請求項1から9のいずれか1項に記載された映像監視装置。
    It further includes an abnormal sound detection unit that collects voice data to detect a specific sound including a scream or an anger, and estimates a sound source direction of the detected specific sound,
    10. The suspicious person candidate detection unit detects the suspicious person candidate using the specific sound detected by the abnormal sound detection unit and the estimated sound source direction. 11. The video surveillance device described in 1.
  11.  予め設定した異変状態を検知し、前記異変状態の発生位置の推定を行う異変検知部をさらに備え、
     前記不審者候補検出部は、前記異変状態の発生位置を用いて前記不審者候補を検出する
    ことを特徴とする請求項1から9のいずれか1項に記載された映像監視装置。
    An anomaly detection unit that detects an anomaly state set in advance and estimates an occurrence position of the anomaly state, further includes,
    10. The video monitoring apparatus according to claim 1, wherein the suspicious person candidate detecting unit detects the suspicious person candidate using the occurrence position of the abnormal state. 11.
  12.  前記不審者候補検出部の検出結果を警備システムへ通報する
    ことを特徴とする請求項1から11のいずれか1項に記載された映像監視装置。
    The video monitoring apparatus according to any one of claims 1 to 11, wherein a detection result of the suspicious person candidate detection unit is reported to a security system.
  13.  フレーム画像から、人物を検出し、検出したそれぞれの前記人物を複数フレームに渡って追跡する人物追跡ステップと、
     前記人物追跡ステップの追跡結果から、前記人物のうち、移動方向及び移動速度が近いと推定する前記人物を集団として認識し、前記集団ごとに集団移動方向及び集団移動速度を含む人流情報を生成する人流情報生成ステップと、
     前記人流情報から、人流の異変が発生したか否かを判別し、異変が発生したと判別したときの位置を特定する人流情報解析ステップと、
     前記人物のうち、前記人流情報解析ステップが特定した前記位置の付近に存在する前記人物を不審者候補として検出する不審者候補検出ステップと、
     前記検出した不審者候補を提示する不審者候補提示ステップと、
    を備えることを特徴とする映像監視方法。
    A person tracking step of detecting a person from the frame image and tracking each detected person over a plurality of frames;
    Based on the tracking result of the person tracking step, the person who is estimated to be close in moving direction and moving speed is recognized as a group, and human flow information including the group moving direction and the group moving speed is generated for each group. Human flow information generation step,
    From the human flow information, it is determined whether or not a human flow anomaly has occurred, and a human flow information analysis step for specifying a position when it is determined that an anomaly has occurred,
    Among the persons, a suspicious person candidate detecting step for detecting the person existing near the position specified by the human flow information analyzing step as a suspicious person candidate;
    A suspicious candidate presentation step for presenting the detected suspicious candidate,
    A video surveillance method comprising:
  14.  音声データを収集して悲鳴または怒号を含む特定音を検出し、検出した前記特定音の音源方向の推定を行う異音検知ステップをさらに備え、
     前記不審者候補検出ステップは、前記異音検知ステップで検出した前記特定音及び推定した前記音源方向を用いて前記不審者候補を検出する
    ことを特徴とする請求項13に記載の映像監視方法。
    It further includes an abnormal sound detection step of collecting sound data to detect a specific sound including a scream or an anger, and estimating a sound source direction of the detected specific sound,
    14. The video monitoring method according to claim 13, wherein the suspicious candidate detection step detects the suspicious candidate using the specific sound detected in the abnormal sound detection step and the estimated sound source direction.
PCT/JP2016/082598 2016-01-20 2016-11-02 Video monitoring apparatus and video monitoring method WO2017126187A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2016008841A JP2019050438A (en) 2016-01-20 2016-01-20 Video monitoring device and video monitoring method
JP2016-008841 2016-01-20

Publications (1)

Publication Number Publication Date
WO2017126187A1 true WO2017126187A1 (en) 2017-07-27

Family

ID=59362705

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2016/082598 WO2017126187A1 (en) 2016-01-20 2016-11-02 Video monitoring apparatus and video monitoring method

Country Status (2)

Country Link
JP (1) JP2019050438A (en)
WO (1) WO2017126187A1 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107995689A (en) * 2017-11-13 2018-05-04 苏州微站通信科技有限公司 A kind of TD-LTE monitoring systems
CN109740444A (en) * 2018-12-13 2019-05-10 深圳云天励飞技术有限公司 Flow of the people information displaying method and Related product
WO2019136777A1 (en) * 2018-01-13 2019-07-18 四川远大创景科技有限公司 Method for monitoring and guiding regional tourist flow and system thereof
WO2019187288A1 (en) * 2018-03-27 2019-10-03 日本電気株式会社 Information processing device, data generation method, and non-transient computer-readable medium whereon program has been stored
JP2021097285A (en) * 2019-12-16 2021-06-24 株式会社アジラ Abnormal behavior detection device
CN114764895A (en) * 2020-12-31 2022-07-19 清华大学 Abnormal behavior detection device and method
CN116931487A (en) * 2023-07-25 2023-10-24 西安速度时空大数据科技有限公司 Data acquisition monitoring system applied to military enterprises

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP7440332B2 (en) 2020-04-21 2024-02-28 株式会社日立製作所 Event analysis system and method
CN117391877A (en) * 2023-12-07 2024-01-12 武汉大千信息技术有限公司 Method for quickly generating suspicious personnel relationship network

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012243128A (en) * 2011-05-20 2012-12-10 Jvc Kenwood Corp Monitoring device, monitoring method and monitoring program
WO2014174738A1 (en) * 2013-04-26 2014-10-30 日本電気株式会社 Monitoring device, monitoring method and monitoring program
JP2015046732A (en) * 2013-08-28 2015-03-12 キヤノン株式会社 Image processing apparatus and image processing method
WO2015068854A1 (en) * 2013-11-11 2015-05-14 日本電気株式会社 Avoidance-behavior detection device, avoidance-reason processing device, and methods for same

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012243128A (en) * 2011-05-20 2012-12-10 Jvc Kenwood Corp Monitoring device, monitoring method and monitoring program
WO2014174738A1 (en) * 2013-04-26 2014-10-30 日本電気株式会社 Monitoring device, monitoring method and monitoring program
JP2015046732A (en) * 2013-08-28 2015-03-12 キヤノン株式会社 Image processing apparatus and image processing method
WO2015068854A1 (en) * 2013-11-11 2015-05-14 日本電気株式会社 Avoidance-behavior detection device, avoidance-reason processing device, and methods for same

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107995689A (en) * 2017-11-13 2018-05-04 苏州微站通信科技有限公司 A kind of TD-LTE monitoring systems
WO2019136777A1 (en) * 2018-01-13 2019-07-18 四川远大创景科技有限公司 Method for monitoring and guiding regional tourist flow and system thereof
WO2019187288A1 (en) * 2018-03-27 2019-10-03 日本電気株式会社 Information processing device, data generation method, and non-transient computer-readable medium whereon program has been stored
JPWO2019187288A1 (en) * 2018-03-27 2021-03-11 日本電気株式会社 Information processing equipment, data generation methods, and programs
US11341774B2 (en) 2018-03-27 2022-05-24 Nec Corporation Information processing apparatus, data generation method, and non-transitory computer readable medium storing program
JP7081659B2 (en) 2018-03-27 2022-06-07 日本電気株式会社 Information processing equipment, data generation method, and program
CN109740444A (en) * 2018-12-13 2019-05-10 深圳云天励飞技术有限公司 Flow of the people information displaying method and Related product
JP2021097285A (en) * 2019-12-16 2021-06-24 株式会社アジラ Abnormal behavior detection device
CN114764895A (en) * 2020-12-31 2022-07-19 清华大学 Abnormal behavior detection device and method
CN116931487A (en) * 2023-07-25 2023-10-24 西安速度时空大数据科技有限公司 Data acquisition monitoring system applied to military enterprises

Also Published As

Publication number Publication date
JP2019050438A (en) 2019-03-28

Similar Documents

Publication Publication Date Title
WO2017126187A1 (en) Video monitoring apparatus and video monitoring method
CN109300471B (en) Intelligent video monitoring method, device and system for field area integrating sound collection and identification
CN108053427B (en) Improved multi-target tracking method, system and device based on KCF and Kalman
Vishwakarma et al. Automatic detection of human fall in video
JP7040463B2 (en) Analysis server, monitoring system, monitoring method and program
JP5560397B2 (en) Autonomous crime prevention alert system and autonomous crime prevention alert method
JP5992276B2 (en) Person recognition apparatus and method
KR102195706B1 (en) Method and Apparatus for Detecting Intruder
JP6532106B2 (en) Monitoring device, monitoring method and program for monitoring
US20150262068A1 (en) Event detection apparatus and event detection method
US9761248B2 (en) Action analysis device, action analysis method, and action analysis program
KR101541272B1 (en) Apparatus and method for detecting terrorism using irregular motion of peoples
JPWO2017047687A1 (en) Monitoring system
JP2013196684A (en) Object counting method and object counting device
JP6233624B2 (en) Information processing system, information processing method, and program
JP6729793B2 (en) Information processing apparatus, control method, and program
US20160217330A1 (en) Image processing system and image processing method
JP2015070401A (en) Image processing apparatus, image processing method, and image processing program
US20210174653A1 (en) System for detecting persons in an area of interest
WO2019220589A1 (en) Video analysis device, video analysis method, and program
KR20200017594A (en) Method for Recognizing and Tracking Large-scale Object using Deep learning and Multi-Agent
KR101407952B1 (en) Elevator crime prvent system and method of controlling the same
KR101238788B1 (en) Elevator crime prvent system and method of controlling the same
CN114764895A (en) Abnormal behavior detection device and method
KR20210043960A (en) Behavior Recognition Based Safety Monitoring System and Method using Artificial Intelligence Technology and IoT

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16886426

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 16886426

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP