WO2022195837A1 - Image analysis device, image analysis system, and image analysis method - Google Patents

Image analysis device, image analysis system, and image analysis method Download PDF

Info

Publication number
WO2022195837A1
WO2022195837A1 PCT/JP2021/011313 JP2021011313W WO2022195837A1 WO 2022195837 A1 WO2022195837 A1 WO 2022195837A1 JP 2021011313 W JP2021011313 W JP 2021011313W WO 2022195837 A1 WO2022195837 A1 WO 2022195837A1
Authority
WO
WIPO (PCT)
Prior art keywords
condition
image
state
detection
event
Prior art date
Application number
PCT/JP2021/011313
Other languages
French (fr)
Japanese (ja)
Inventor
友輔 生内
圭吾 長谷川
Original Assignee
株式会社日立国際電気
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社日立国際電気 filed Critical 株式会社日立国際電気
Priority to PCT/JP2021/011313 priority Critical patent/WO2022195837A1/en
Priority to JP2023506655A priority patent/JPWO2022195837A1/ja
Publication of WO2022195837A1 publication Critical patent/WO2022195837A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion

Definitions

  • the present invention relates to an image analysis device, an image analysis system, and an image analysis method, and more particularly to an image analysis device, an image analysis system, and an image analysis method that analyze captured images and detect objects.
  • AI Artificial Intelligence
  • DL Deep Learning
  • AI such as DL often uses supervised learning.
  • the pairs of input images and expected outputs are called training data.
  • generalization performance is required to output (infer) appropriate results even for unknown input images.
  • a neural network exists as a representative method.
  • Patent Literature 1 a display terminal displays specific object information notified from a surveillance camera, and according to an operation instruction from an administrator, a nearby surveillance camera is displayed based on position information/time information from the surveillance camera.
  • a surveillance camera system is disclosed that sends a control signal to an object to track based on specific object information.
  • confidence a statistical measure of how certain AI's inference results are
  • the operational environment there are various events in the operational environment that cause a decrease in inference accuracy due to a decrease in the confidence of the object to be detected.
  • the events here include, for example, differences in shooting environment, background, angle of view, and time zone at the time of learning data collection, and existence of objects not included in learning data.
  • inference accuracy may be degraded if a detection target object is partially hidden by other objects, and the detection target object may be overlooked.
  • Patent Document 1 discloses a technique for tracking based on specific object information, but does not describe a detection method when the detection target object is shielded as described above.
  • an object of the present invention is to provide an image analysis device, an image analysis system, and an image analysis method that enable more accurate detection of a hidden object.
  • one representative image analysis apparatus of the present invention includes an image acquisition unit that acquires a photographed image, and an image that outputs analysis results for the input image acquired by the image acquisition unit.
  • a processing unit and a correction data acquisition unit that acquires data used for correction conditions, and the image processing unit determines that the detection target object exists in the image based on the conditions for detecting the detection target object.
  • a state of occurrence of an event to be detected a state of no event in which it is determined that the detection target object does not exist in the image based on a condition of non-detection of the detection target object, and a state in which the detection target object is shielded using the correction condition.
  • the condition for transition from the neutral state to the event occurrence state is set looser than the transition condition from the no-event state to the event occurrence state. It is characterized by being
  • one of the image analysis methods of the present invention is an image analysis method for performing image analysis using a processing device, and includes steps of acquiring a photographed image, acquiring data used for correction conditions, An event occurrence state for determining that the detection target object exists in the image based on the object detection condition, and determining that the detection target object does not exist in the image based on the detection target object non-detection condition. and a neutral state in which it is determined that the object to be detected is shielded using the correction condition, and a transition from the neutral state to the event occurrence state is performed.
  • a condition for transition is set looser than a condition for transition from the no-event state to the event-occurred state.
  • an image analysis apparatus can more accurately detect a shielded object. Problems, configurations, and effects other than those described above will be clarified by the following embodiments.
  • FIG. 1 is a system configuration diagram showing an example of the image analysis system of the present invention.
  • FIG. 2 is a hardware configuration diagram showing an example of the image analysis apparatus of the present invention.
  • FIG. 3 is a functional block diagram showing an example of the image analysis device of the present invention.
  • 4 is a functional block diagram showing an example of an image processing unit in FIG. 3.
  • FIG. 5 is a diagram of a first example for explaining an example of the flow of processing when correction conditions are not used.
  • FIG. 5 is a diagram of a second example for explaining an example of the flow of processing when correction conditions are not used.
  • FIG. 7 is a diagram of a first example for explaining the flow of processing according to the first correction condition of the image analysis apparatus of the present invention.
  • FIG. 1 is a system configuration diagram showing an example of the image analysis system of the present invention.
  • FIG. 2 is a hardware configuration diagram showing an example of the image analysis apparatus of the present invention.
  • FIG. 3 is a functional block diagram showing an example of the image analysis device
  • FIG. 8 is a diagram of a second example for explaining the flow of processing according to the first correction condition of the image analysis apparatus of the present invention.
  • FIG. 9 is a diagram of a third example for explaining the flow of processing according to the first correction condition of the image analysis apparatus of the present invention.
  • FIG. 10 is a diagram of a first example for explaining the flow of processing according to the second correction condition of the image analysis apparatus of the present invention.
  • FIG. 11 is a diagram of a second example for explaining the flow of processing according to the second correction condition of the image analysis apparatus of the present invention.
  • FIG. 12 is a diagram explaining an example of state transition of the image analysis apparatus of the present invention.
  • FIG. 13 is an example of comparing the results of the image analysis apparatus of the present invention with the case where there is no correction condition.
  • FIG. 14 shows an example of a processing flowchart of the image analysis apparatus according to the present invention.
  • FIG. 1 is a system configuration diagram showing an example of the image analysis system of the present invention.
  • the analysis server 101, camera 102, and database server 103 are connected via a network 104.
  • a network 104 is a line capable of data communication that connects each server. Any type of line, such as a dedicated line, an intranet, an IP network such as the Internet, etc., does not matter.
  • Video data acquired by the camera 102 is analyzed by the analysis server 101 and the output result is stored in the database server 103 .
  • FIG. 1 the configuration in FIG. 1 is an example, and various modifications are possible, such as performing AI inference and image analysis system processing on the camera 102 .
  • the analysis server 101 can be applied as an image analysis device that analyzes images, and its configuration and processing details will be described later.
  • a configuration of a camera in which information is obtained by forming an image of incident light on an imaging device via a lens and an aperture can be applied.
  • the imaging device here include a CCD (Charge-Coupled Device) image sensor and a CMOS (Complementary Metal Oxide Semiconductor) image sensor.
  • the camera 102 shoots an image at, for example, 3 frames per second (3 fps) or more, and the information is sent to the analysis server 101 and the database server 103 .
  • a plurality of cameras 102 can be installed according to the situation, and can be arranged in various places. For example, it may be installed at a monitoring location as a monitoring camera.
  • the database server 103 is a device that records images captured by the camera 102, information necessary for processing by the analysis server 101, processing results of the analysis server 101, and the like.
  • a device for recording for example, HDD (Hard Disk Drive), SSD (Solid State Drive), DDS (Digital Data Storage), etc., can be applied according to need.
  • FIG. 2 is a hardware configuration diagram showing an example of the image analysis apparatus of the present invention.
  • a hardware configuration example of the analysis server 101 will be described with reference to FIG.
  • the hardware consists of a computer system equipped with a processing unit such as a CPU (Central Processing Unit), and each function is executed.
  • a processing unit such as a CPU (Central Processing Unit), and each function is executed.
  • a processing device in addition to the CPU, a Digital Signal Processor (DSP), a Field-Programmable Gate Array (FPGA), a Graphics Processing Unit (GPU), etc. may be applied.
  • the analysis server 101 includes a processor section 201 , a main storage section 202 , an auxiliary storage section 203 , an input/output interface section 204 , a display interface section 205 and a network interface section 206 , which are connected by a bus 207 .
  • An input/output interface unit 204 is connected to an input device 208 such as a keyboard and mouse to provide a user interface.
  • the display interface unit 205 is connected to the display output device 209 .
  • a network interface unit 206 is an interface for connecting the analysis server 101 and the network 104 .
  • the auxiliary storage unit 203 is usually composed of a non-volatile memory such as an HDD or flash memory, and stores programs executed by the analysis server 101 and data to be processed by the programs.
  • the main storage unit 202 is composed of a RAM, and temporarily stores programs, data necessary for executing the programs, and the like according to instructions from the processor unit 201 .
  • the processor unit 201 executes programs loaded from the auxiliary storage unit 203 to the main storage unit 202 . Note that the configuration in FIG. 2 is an example, and various modifications are possible.
  • FIG. 3 is a functional block diagram showing an example of the image analysis device of the present invention. Functional blocks of the analysis server 101 will be described with reference to FIG.
  • the analysis server 101 is composed of an auxiliary storage unit 203 , an image acquisition unit 301 , a correction data acquisition unit 302 , an image processing unit 303 , a storage control unit 304 and a display control unit 305 .
  • the image acquisition unit 301 acquires the signal obtained from the auxiliary storage unit 203 as an image.
  • a correction data acquisition unit 302 acquires the signal obtained from the auxiliary storage unit 203 as an image or time-series data.
  • An image processing unit 303 receives the image obtained by the image obtaining unit 301 and the data obtained by the correction data obtaining unit 302, performs AI inference processing and shielding detection processing, and determines the state of the system from the results.
  • a storage control unit 304 performs storage control of the output result using the result of the image processing unit 303 and stores it in the auxiliary storage unit 203 .
  • a display control unit 305 controls the display of the result of the image processing unit 303 and the information stored in the auxiliary storage unit 203 and outputs the information to the display output device 209 .
  • the image acquisition unit 301 acquires image data from a video signal input from a video storage device or the like in which image data is stored.
  • the image data may be subjected to preprocessing such as a smoothing filter, an edge enhancement filter, and density conversion.
  • preprocessing such as a smoothing filter, an edge enhancement filter, and density conversion.
  • a data format such as RGB color, YUV, or monochrome may be selected according to the application.
  • the image data may be reduced to a predetermined size.
  • FIG. 4 is a functional block diagram showing an example of the image processing unit 303 in FIG. Next, functional blocks of the image processing unit 303 will be described with reference to FIG.
  • the image processing unit 303 is composed of an AI inference processing unit 401 , a shielding detection processing unit 402 , and an inference result correction unit 403 .
  • the AI inference processing unit 401 performs inference processing (eg, object detection) using AI on the input image acquired by the image acquisition unit 301 .
  • the shielding detection processing unit 402 uses the input image acquired by the image acquisition unit 301 and the correction data acquired by the correction data acquisition unit 302 (background image, previous frame image, inference result, etc.) to determine the correction conditions. Determine the presence or absence of the object to be detected.
  • the inference result correction unit 403 uses the results obtained by the AI inference processing unit 401 and the shielding detection processing unit 402 to determine the final determination of the image analysis system, and outputs the result to the storage control unit 304 .
  • Inference processing by AI is, for example, the process of extracting feature values using neural networks, deep learning, etc., and the degree of certainty can be estimated from the feature values related to the detection target in the image.
  • CNN Convolution Neural Networks
  • FIG. 5 and 6 are diagrams for explaining an example of the flow of processing when correction conditions are not used.
  • FIG. 5 shows an example of transition from "no event” to "event occurrence”.
  • FIG. 6 shows an example of transition from "event occurrence” to "no event”.
  • Processing when no correction condition is used can determine two states: "no event” (first state) and “event occurrence” (second state).
  • the “no event” state here is a state in which the existence of a person, who is a detection target object, is not confirmed in the image based on the non-detection condition of the detection target object.
  • the “event occurrence” state is a state in which it is determined that a person, who is a detection target object, exists in the image based on the conditions for detecting the detection target object.
  • the detection result as a system is determined by changing the state of the system.
  • the "detection event occurrence condition” (first condition) is a condition for transitioning from the "no event” state to the "event occurrence” state.
  • the 'detection event continuation condition' (second condition) is a condition for 'event occurrence' to continue from the 'event occurrence' state.
  • the “detection event occurrence condition” (first condition) is the case where there are 3 out of 3 consecutive frames with a certainty of 60% or higher when an object is detected.
  • the “detection event continuation condition” (second condition) is a case where one out of three consecutive frames has a confidence of 60% or more at the time of object detection.
  • the state is "event occurrence" at first.
  • frames 22, 23, and 24 the confidence factor of the person to be detected is not obtained.
  • none of the three consecutive frames has a certainty of 60% or higher when an object is detected, and the second condition, the "detection event continuation condition,” is not satisfied. Therefore, in frame 24, the state of "event occurred” is changed to the state of "no event". After that, the state is "no event” until the first condition "detection event generation condition" is satisfied.
  • FIG. 7 to 9 are diagrams for explaining the flow of processing according to the first correction condition of the image analysis apparatus of the present invention.
  • FIG. 7 shows an example of transition from “no event” to "event occurrence”.
  • FIG. 8 shows an example of transitions in the order of "event occurrence", “neutral”, and “event occurrence”.
  • FIG. 9 shows an example of transition from "event occurrence” to "no event”.
  • the processing is performed by the analysis server 101, which is an image analysis device.
  • the "correction condition" (third condition) is satisfied. do. In this case, it is determined that "the object to be detected is shielded" and the inference result is corrected.
  • the setting of the correction range it is also possible to set the range in advance. For example, if it is a fixed camera, the range can be assumed in advance. Further, the correction range may be calculated from the moving state of the detection target object. It is possible to determine the correction range by estimating that the object to be detected is not out of the screen with the amount of movement up to that frame. For example, if the moving speed of the object to be detected in the previous frames is slow, there is a high possibility that the object will still exist in the screen in the previous frame.
  • the "detection event occurrence condition” (first condition) is the case where there are 3 out of 3 consecutive frames with a degree of confidence of 60% or more at the time of object detection. Further, the “detection event continuation condition” (second condition) is a case where one out of three consecutive frames has a confidence of 60% or more at the time of object detection. Also, the "correction condition” (third condition) is satisfied if the position of the object is within a preset correction range in the last frame in which the object to be detected is detected during tracking. Further, when the "correction condition” (third condition) is satisfied, the transition is made to "neutral" (third state).
  • the first frame 41 is in the "event occurrence" state.
  • Frames 42, 43, and 44 do not have the confidence of the person to be detected, so the confidence is less than 60%. Since the frames 42 and 43 satisfy the "detection event continuation condition" (second condition), they are “event occurrence”. In frame 44, the second condition, ie, the "detection event continuation condition” is not satisfied.
  • the last detected object during tracking is the object 41a in the frame 41 with a certainty of 60% or higher.
  • the object 41a here exists within the set correction range 41b. Therefore, the frame 44 satisfies the "correction condition” (third condition) and transitions to the "neutral" state (third state).
  • FIG. 10 and 11 are diagrams for explaining the flow of processing according to the second correction condition of the image analysis apparatus of the present invention.
  • FIG. 10 shows an example of transition in order of "event occurrence”, “neutral”, and “event occurrence”.
  • FIG. 11 shows an example of transition in order of "event occurrence", “neutral”, and "no event”.
  • the processing here is performed by the analysis server 101, which is an image analysis device.
  • auxiliary information (results of analysis by background subtraction technology or the like) is used to correct AI inference results.
  • the processing under the second correction condition in FIGS. 10 and 11 is an example of correction of AI inference results using background subtraction technology as a correction method.
  • the difference from the background image captured in advance in the frame that no longer satisfies the second condition is binarized with a certain threshold value. If the difference area is larger than the area of the detection frame of the frame in which the detection target object was last detected, it is determined that "the target object is shielded". In this case, the frame satisfies the "correction condition” (third condition) and transitions to the "neutral" state.
  • the "detection event occurrence condition” (first condition) is the case where there are 3 out of 3 consecutive frames with a degree of confidence of 60% or more at the time of object detection. Further, the “detection event continuation condition” (second condition) is a case where one out of three consecutive frames has a confidence of 60% or more at the time of object detection.
  • the "correction condition” (third condition) is satisfied when the difference area between the frame that no longer satisfies the "detection event continuation condition" and the background image is larger than the detection frame of the last detected object. Further, when the "correction condition” (third condition) is satisfied, the state shifts to "neutral" (third state).
  • the first frame 61 is in the "event occurrence" state.
  • the confidence of the person to be detected is not obtained, so the confidence is less than 60%. Since the frames 62 and 63 satisfy the "detection event continuation condition" (second condition), they are “event occurrence”.
  • the second condition ie, the "detection event continuation condition” is not satisfied.
  • the difference between the pre-detection image 68 which is a background image captured in advance, and the frame 64 to be determined now is obtained.
  • Frame 64 now shows additional cars and people compared to pre-detection image 68 where only trees were shown. Therefore, the car and person portions correspond to the difference and are displayed in white in the difference image 69 .
  • This white area is larger than the area of the person of the detection target object surrounded by the frame of the frame 61 where the event occurred, because the car and the person are shown. Therefore, it is presumed that there was a large shielding object in front of the person, and it is determined that the "correction condition" is satisfied.
  • this "neutral” state if the "detection event continuation condition” (second condition) is satisfied within the correction frame (three frames in FIGS. 10 and 11), the state shifts to the "event occurrence" state.
  • the detection target object is detected with a certainty of 80%, so the "detection event continuation condition" is satisfied, and the state transitions to the "event occurrence” state.
  • the detection frame will gradually become smaller when there is a shielding object. At that time, when the detection frame becomes smaller than a predetermined amount, it is determined that the object to be detected is occluded, and the "correction condition" is satisfied, and the same processing as in FIGS. 7 to 11 can be performed. .
  • FIG. 12 is a diagram explaining an example of state transition of the image analysis apparatus of the present invention.
  • Arrow 2 indicates transition from "event occurrence” to "neutral".
  • the condition here corresponds to the case where the “detection event continuation condition” (second condition) is not satisfied but the “correction condition” (third condition) is satisfied.
  • Arrow 3 indicates transition from “neutral” to "event occurrence”.
  • the condition here corresponds to the case where the “detection event continuation condition” (second condition) is satisfied within the correction frame.
  • Arrow 4 indicates a transition from "neutral” to "no event".
  • the condition here corresponds to the case where the "detection event continuation condition” (second condition) is not satisfied within the correction frame.
  • Arrow 5 indicates transition from “event occurrence” to "no event".
  • the condition here corresponds to the case where the “detection event continuation condition” (second condition) and the “correction condition” (third condition) are not satisfied. In this case, the transition is made from “event occurrence” to "no event” without going through "neutral".
  • Arrow 6 indicates transition from “no event” to "event occurrence”.
  • the condition here corresponds to the case where the “detection event occurrence condition” (first condition) is satisfied.
  • the "detection event continuation condition” (first condition) is a loose condition that is easier to achieve than the “detection event occurrence condition” (second condition).
  • the "detection event generation condition” includes M frames out of N consecutive frames with a certainty of X% or more of the detection target object at the time of object detection (N and M are integers and N ⁇ M)
  • the "detection The “event continuation condition” is, for example, the case where there are P out of N consecutive frames (P is an integer and M>P) where the certainty of the detection target object at the time of object detection is Y% or higher.
  • the certainty factors of both may be the same, or may be changed. In this case X ⁇ Y is preferred. In this way, the condition can be loosened by lowering the appearance probability of the frame at a certainty or more.
  • the condition for transitioning from "event occurrence” to “neutral” indicated by arrow 3 is not limited to the above. If it is Therefore, it is possible to set a looser condition than the "detection event generation condition" regardless of the “detection event continuation condition”. For example, if the “detection event occurrence condition” is that there are M frames out of N consecutive frames with a certainty of X% or more at the time of object detection (N and M are integers, N ⁇ M), “neutral” to “event The condition for transitioning to "occurrence” is, for example, that there are Q out of N consecutive frames with a certainty Z% or more at the time of object detection (Q is an integer and M>Q).
  • condition of arrow 4 is a case where this condition is not satisfied.
  • certainty factors of both may be the same, or may be changed.
  • X ⁇ Z is preferred. In this way, the condition can be loosened by lowering the appearance probability of the frame at a certainty or more.
  • FIG. 13 is an example of comparing the results of the image analysis apparatus of the present invention with the case where there is no correction condition.
  • a comparison between the results of the image analysis apparatus of the present invention (method with correction conditions) and the method without correction conditions will be described using the table of FIG.
  • "O" indicates the occurrence of an event
  • "X" indicates no event
  • " ⁇ " indicates a neutral state.
  • frame no. 1, no. 2 is the state of "no event".
  • no. Up to 7 the "detection event continuation condition” is satisfied and the state is "event occurrence”.
  • frame no. 8 to frame no. Up to 13 the second condition "detection event continuation condition” is not satisfied. Therefore, frame no. 8 to frame no. Up to 13, the state of the image transits to the state of "no event", and a loss of alarm occurs.
  • FIG. 14 shows an example of a processing flowchart of the image analysis apparatus of the present invention.
  • the processor unit 201 of the analysis server 101 executes the program loaded from the auxiliary storage unit 203 to the main storage unit 202 to activate the image analysis system.
  • the image analysis system may allow the user to check the result on a GUI (Graphical User Interface), or may allow the user to check only the states of "event occurrence", “no event”, and "neutral".
  • the first condition "detection event occurrence condition”, the second condition “detection event continuation condition”, the third condition "correction condition”, the threshold for object detection, the correction range, etc. are prepared in advance. You can read the configuration file and set it. Alternatively, the user may be allowed to select using a GUI. It should be noted that it is desirable that the “detection event continuation condition” (second condition) is easier to achieve than the “detection event occurrence condition” (first condition). Also, one “correction condition” (third condition) may be used, or a combination of a plurality of conditions may be used for determination.
  • step 1001 the user selects an object to be detected and the camera 102 to be used for detection.
  • the type of object to be detected and the number of cameras may be one or more.
  • an example in which the number of objects and cameras to be detected is one will be described.
  • the image acquired from the camera 102 is read.
  • the frame rate, image size, and the like of the camera 102 may be set in advance on the camera 102 side, or may be arbitrarily selected by the user through the GUI.
  • step 1003 it is determined whether the end command has been executed.
  • the end command may be a keyboard operation or a GUI operation.
  • step 1004 AI inference processing is performed using the read image, and the result is output. Already explained processing can be used for the inference processing of AI here.
  • step 1005 the state of the system is determined, and subsequent processing is determined. If the state of the system is "no event”, go to step 1006; On the other hand, in the case of the “event occurrence” state or the “neutral” state, the process proceeds to step 1007 .
  • step 1006 it is determined whether the AI inference result satisfies the "detection event occurrence condition" (first condition). to read the image of the next frame.
  • step 1007 it is determined whether the AI inference result satisfies the "detection event continuation condition" (second condition). move on.
  • step 1008 the state of the image analysis system is changed to "event occurrence”.
  • step 1009 the user is notified of the occurrence of the event.
  • the occurrence notification may be made on the GUI, or the event occurrence notification may be delivered to the small terminal via the network 104 or the like.
  • the process advances to step 1002 to read the image of the next frame.
  • step 1010 the state of the image analysis system is determined, and subsequent processing is determined. If the state of the image analysis system is "event occurrence", the process proceeds to step 1011; On the other hand, in the case of the "neutral” state, the process proceeds to step 1012 .
  • step 1011 it is determined whether or not the "correction condition” (third condition) is satisfied using the background subtraction technique, the tracking technique, etc. If the "correction condition" is satisfied, the process proceeds to step 1013; Proceed to 1015.
  • step 1012 it is determined whether or not the "correction condition" is satisfied. If the "correction condition” is satisfied, the process proceeds to step 1013, and if the "correction condition" is not satisfied, the process proceeds to step 1015.
  • step 1013 the state of the image analysis system is changed to "neutral".
  • a warning notification is issued to notify the user of the "neutral" state.
  • the warning notification may be made on the GUI, or the event occurrence notification may be delivered to the small terminal via the network 104 or the like.
  • the process proceeds to step 1002 to read the image of the next frame.
  • step 1015 the state of the image analysis system is changed to "no event", and the process proceeds to step 1002 to read the image of the next frame.
  • the correction conditions at this time can be set to more suitable correction conditions by setting a correction range using a tracking technique or using a background subtraction technique.
  • Situations in which the present invention is expected to be applied include obstacles that partially or completely hide the object to be detected during operation with moving images or live images.
  • it can be applied to environments such as outdoors where brightness and color tend to change.
  • the present invention is not limited to the above-described embodiments, and includes various modifications.
  • the above-described embodiments have been described in detail in order to explain the present invention in an easy-to-understand manner, and are not necessarily limited to those having all the configurations described.
  • analysis server 101 and the camera 102 have been described as separate members, they may be integrated. That is, it can be configured as an image analysis apparatus having the function of the analysis server 101 on the side of the camera 102 without going through the network 104 .

Abstract

Provided are an image analysis device, an image analysis system, and an image analysis method capable of more accurate detection of a shielded object. The present invention comprises an image acquisition unit (301) which acquires a captured image; an image processing unit (303) which outputs an analysis result for an input image acquired by the image acquisition unit (301); and a correction data acquisition unit (302) which acquires data used for a correction condition, wherein the image processing unit (303) discriminates an event occurrence state in which a detection target object is determined to be present in the image on the basis of a detection condition of the detection target object, an event-free state in which the detection target object is determined not to be present in the image on the basis of a non-detection condition of the detection target object, and a neural state in which the detection target object is determined to be shielded by using a correction condition, and a condition of transition from the neutral state to the event occurrence state is set looser than a condition of transition from the event-free state to the event occurrence state.

Description

画像解析装置、画像解析システム及び画像解析方法Image analysis device, image analysis system and image analysis method
 本発明は、画像解析装置、画像解析システム及び画像解析方法に関し、特に、撮影された画像を解析して、物体を検知する画像解析装置、画像解析システム及び画像解析方法に関する。 The present invention relates to an image analysis device, an image analysis system, and an image analysis method, and more particularly to an image analysis device, an image analysis system, and an image analysis method that analyze captured images and detect objects.
 撮影した画像を解析するための手法は多く存在する。例えば、画像を入力とした物体認識・属性識別・解析・予測などのタスクに対して、Deep Learning(DL)などのAI(Artificial Intelligence)の適用が進んでいる。このような機能を製品やシステムに組み込むことで人の仕事を代替して人件費の削減や商品の付加価値向上を図ることができる。これらは、製品やシステムの販売促進に寄与するため、多くの企業がAIに対し高い関心を抱いている。 There are many methods for analyzing captured images. For example, the application of AI (Artificial Intelligence) such as Deep Learning (DL) is progressing for tasks such as object recognition, attribute identification, analysis, and prediction using images as input. By incorporating such functions into products and systems, it is possible to reduce labor costs and improve the added value of products by substituting human work. Many companies have high interest in AI because it contributes to sales promotion of products and systems.
 DLなどのAIでは教師有り学習が多く用いられる。ある問題に対する入力画像と画像を入力した結果として期待する出力信号のペアの集合を用いて、入力画像から正しい出力信号を出力するという問題を解くモデルを学習する。入力画像と期待する出力のペアを学習データと呼ぶ。実用上では未知の入力画像に対しても適切な結果を出力(推論)する汎化性能が求められる。代表的な手法としてはニューラルネットワークが存在する。 AI such as DL often uses supervised learning. Using a set of pairs of input images and expected output signal pairs for a given problem, we train a model that solves the problem of outputting the correct output signal from the input images. The pairs of input images and expected outputs are called training data. In practice, generalization performance is required to output (infer) appropriate results even for unknown input images. A neural network exists as a representative method.
 一方、特許文献1には、表示端末が、監視カメラから通知された特定の物体情報を表示すると共に、管理者からの操作指示により監視カメラからの位置情報/時刻情報を元に近接する監視カメラに制御信号を送信して特定の物体情報に基づく追跡を行う監視カメラシステムが開示されている。 On the other hand, in Patent Literature 1, a display terminal displays specific object information notified from a surveillance camera, and according to an operation instruction from an administrator, a nearby surveillance camera is displayed based on position information/time information from the surveillance camera. A surveillance camera system is disclosed that sends a control signal to an object to track based on specific object information.
特許第6403784号公報Japanese Patent No. 6403784
 物体認識の処理では、検知/非検知の判定に確信度(AIの推論結果がどのくらい確実であるかの統計的な尺度)を用いることができる。例えば、確信度が60%以上であれば検知、60%未満を非検知とする、というように確信度に閾値を設け、検知/非検知の判定をする。  In the object recognition process, it is possible to use confidence (a statistical measure of how certain AI's inference results are) to determine detection/non-detection. For example, if the certainty is 60% or higher, it is detected, and if it is less than 60%, it is judged as non-detected.
 しかし、上述した従来の画像を解析する方法では、運用環境においては、検知対象物体の確信度の低下にともなう推論精度低下の原因となる様々な事象が存在する。ここでの事象は、例えば、学習データ収集時の撮影環境、背景、画角、時間帯などの違いや、学習データに含まれない物体の存在などである。具体的には、物体認識のタスクにおいて、検知対象物体が他の物体に一部隠れることで推論精度が低下し、検知対象物体の見逃しが発生する恐れがある。 However, in the conventional image analysis method described above, there are various events in the operational environment that cause a decrease in inference accuracy due to a decrease in the confidence of the object to be detected. The events here include, for example, differences in shooting environment, background, angle of view, and time zone at the time of learning data collection, and existence of objects not included in learning data. Specifically, in the task of object recognition, inference accuracy may be degraded if a detection target object is partially hidden by other objects, and the detection target object may be overlooked.
 また、動画やライブ映像での運用時に、単一フレームではなく複数フレームのAIの推論結果を利用して物体をトラッキングして検知/非検知を判定し、検知対象の見逃しを抑えることもできる。しかしこれらの手法でも、判定対象となる複数フレーム間にて検知条件やトラッキング条件を満たす必要がある。このため、一度検知したりトラッキングしたりした物体であっても、検知対象物体が他の物体で遮蔽されるなどして、学習したモデルでは検知できないフレームが続く場合は、見逃してしまうことが想定される。 In addition, when operating with video or live video, it is possible to use the AI inference results of multiple frames instead of a single frame to track objects and determine whether they are detected or not, thereby reducing the possibility of overlooking detection targets. However, even with these methods, it is necessary to satisfy detection conditions and tracking conditions between multiple frames to be determined. For this reason, even if an object has been detected or tracked once, it is assumed that if the object to be detected is blocked by another object and the frames cannot be detected by the learned model, the frames will be missed. be done.
 遮蔽された検知対象物体を見逃さないためには、学習したモデルでは検知できないフレームに対して検知対象物体が遮蔽されていることを検知し、推論結果を補正して検知できる仕組みが必要となる。しかし推論結果を補正して検知と判定するだけでは、検知対象物体が既に検知範囲外に出ているにもかかわらず検知状態と判定し、誤報が発生することも想定される。そのため、検知対象物体の遮蔽を検知し推論結果の補正をするとともにその補正による誤報を防ぐ仕組みが必要となる。  In order not to miss the occluded detection target object, it is necessary to detect that the detection target object is occluded for frames that cannot be detected by the learned model, and to correct the inference results for detection. However, it is conceivable that if the inference result is only corrected and the detection is determined, the detection state is determined even though the object to be detected is already out of the detection range, and an erroneous alarm occurs. Therefore, it is necessary to detect the shielding of the object to be detected, correct the inference result, and prevent false alarms due to the correction.
 また、特許文献1には、特定の物体情報に基づく追跡を行う技術が開示されているが、上記のように検知対象物体が遮蔽されている場合についての検知の方法については記載されていない。 In addition, Patent Document 1 discloses a technique for tracking based on specific object information, but does not describe a detection method when the detection target object is shielded as described above.
 本発明は、上記課題に鑑みて、遮蔽されている物体についてのより正確な検知が可能な画像解析装置、画像解析システム及び画像解析方法を提供することを目的とする。 In view of the above problems, an object of the present invention is to provide an image analysis device, an image analysis system, and an image analysis method that enable more accurate detection of a hidden object.
 上記目的を達成するため、代表的な本発明の画像解析装置の一つは、撮影した画像を取得する画像取得部と、前記画像取得部により取得した入力画像に対して解析結果を出力する画像処理部と、補正条件に用いるデータを取得する補正用データ取得部とを備え、前記画像処理部は、検知対象物体の検知の条件に基づき前記検知対象物体が画像内に存在していると判定するイベント発生の状態と、検知対象物体の非検知の条件に基づき前記検知対象物体が画像内に存在しないと判定するイベントなしの状態と、前記補正条件を用いて前記検知対象物体が遮蔽されていると判定するニュートラルの状態とを、判別し、前記ニュートラルの状態から前記イベント発生の状態への遷移の条件は、前記イベントなしの状態から前記イベント発生の状態への遷移の条件よりも緩く設定されていることを特徴とする。 In order to achieve the above object, one representative image analysis apparatus of the present invention includes an image acquisition unit that acquires a photographed image, and an image that outputs analysis results for the input image acquired by the image acquisition unit. A processing unit and a correction data acquisition unit that acquires data used for correction conditions, and the image processing unit determines that the detection target object exists in the image based on the conditions for detecting the detection target object. a state of occurrence of an event to be detected, a state of no event in which it is determined that the detection target object does not exist in the image based on a condition of non-detection of the detection target object, and a state in which the detection target object is shielded using the correction condition. The condition for transition from the neutral state to the event occurrence state is set looser than the transition condition from the no-event state to the event occurrence state. It is characterized by being
 さらに本発明の画像解析方法の一つは、処理装置を用いて画像解析を行う画像解析方法であって、撮影した画像を取得するステップと、補正条件に用いるデータを取得するステップと、検知対象物体の検知の条件に基づき前記検知対象物体が画像内に存在していると判定するイベント発生の状態と、検知対象物体の非検知の条件に基づき前記検知対象物体が画像内に存在しないと判定するイベントなしの状態と、前記補正条件を用いて前記検知対象物体が遮蔽されていると判定するニュートラルの状態とを、判別するステップとを備え、前記ニュートラルの状態から前記イベント発生の状態への遷移の条件は、前記イベントなしの状態から前記イベント発生の状態への遷移の条件よりも緩く設定されていることを特徴とする。 Further, one of the image analysis methods of the present invention is an image analysis method for performing image analysis using a processing device, and includes steps of acquiring a photographed image, acquiring data used for correction conditions, An event occurrence state for determining that the detection target object exists in the image based on the object detection condition, and determining that the detection target object does not exist in the image based on the detection target object non-detection condition. and a neutral state in which it is determined that the object to be detected is shielded using the correction condition, and a transition from the neutral state to the event occurrence state is performed. A condition for transition is set looser than a condition for transition from the no-event state to the event-occurred state.
 本発明によれば、画像解析装置、画像解析システム及び画像解析方法において、遮蔽されている物体についてのより正確な検知をすることができる。
 上記以外の課題、構成及び効果は、以下の実施形態により明らかにされる。
According to the present invention, an image analysis apparatus, an image analysis system, and an image analysis method can more accurately detect a shielded object.
Problems, configurations, and effects other than those described above will be clarified by the following embodiments.
図1は、本発明の画像解析システムの一例を示すシステム構成図である。FIG. 1 is a system configuration diagram showing an example of the image analysis system of the present invention. 図2は、本発明の画像解析装置の一例を示すハードウェア構成図である。FIG. 2 is a hardware configuration diagram showing an example of the image analysis apparatus of the present invention. 図3は、本発明の画像解析装置の一例を示す機能ブロック図である。FIG. 3 is a functional block diagram showing an example of the image analysis device of the present invention. 図4は、図3における画像処理部の一例を示す機能ブロック図である。4 is a functional block diagram showing an example of an image processing unit in FIG. 3. FIG. 図5は、補正条件を用いない場合の処理の流れの一例を説明する第1の例の図である。FIG. 5 is a diagram of a first example for explaining an example of the flow of processing when correction conditions are not used. 図5は、補正条件を用いない場合の処理の流れの一例を説明する第2の例の図である。FIG. 5 is a diagram of a second example for explaining an example of the flow of processing when correction conditions are not used. 図7は、本発明の画像解析装置の第1の補正条件による処理の流れを説明する第1の例の図である。FIG. 7 is a diagram of a first example for explaining the flow of processing according to the first correction condition of the image analysis apparatus of the present invention. 図8は、本発明の画像解析装置の第1の補正条件による処理の流れを説明する第2の例の図である。FIG. 8 is a diagram of a second example for explaining the flow of processing according to the first correction condition of the image analysis apparatus of the present invention. 図9は、本発明の画像解析装置の第1の補正条件による処理の流れを説明する第3の例の図である。FIG. 9 is a diagram of a third example for explaining the flow of processing according to the first correction condition of the image analysis apparatus of the present invention. 図10は、本発明の画像解析装置の第2の補正条件による処理の流れを説明する第1の例の図である。FIG. 10 is a diagram of a first example for explaining the flow of processing according to the second correction condition of the image analysis apparatus of the present invention. 図11は、本発明の画像解析装置の第2の補正条件による処理の流れを説明する第2の例の図である。FIG. 11 is a diagram of a second example for explaining the flow of processing according to the second correction condition of the image analysis apparatus of the present invention. 図12は、本発明の画像解析装置の状態遷移の例を説明する図である。FIG. 12 is a diagram explaining an example of state transition of the image analysis apparatus of the present invention. 図13は、補正条件がない場合と本発明の画像解析装置の結果を比較した一例である。FIG. 13 is an example of comparing the results of the image analysis apparatus of the present invention with the case where there is no correction condition. 図14は、本発明に係る画像解析装置の処理フローチャートの一例を示す。FIG. 14 shows an example of a processing flowchart of the image analysis apparatus according to the present invention.
 本発明を実施するための形態を説明する。 A form for carrying out the present invention will be described.
(システム構成)
 図1は、本発明の画像解析システムの一例を示すシステム構成図である。
(System configuration)
FIG. 1 is a system configuration diagram showing an example of the image analysis system of the present invention.
 解析サーバ101とカメラ102、データベースサーバ103が、ネットワーク104にて接続されている。ネットワーク104は各サーバを結ぶデータ通信可能な回線である。専用線、イントラネット、インターネット等のIPネットワーク等、回線の種類は問わない。カメラ102で取得した映像データは解析サーバ101にて解析され、出力結果はデータベースサーバ103に記憶される。なお図1の構成は一例であり、カメラ102上でAIの推論と画像解析システムの処理を行う等様々な変形が可能である。 The analysis server 101, camera 102, and database server 103 are connected via a network 104. A network 104 is a line capable of data communication that connects each server. Any type of line, such as a dedicated line, an intranet, an IP network such as the Internet, etc., does not matter. Video data acquired by the camera 102 is analyzed by the analysis server 101 and the output result is stored in the database server 103 . Note that the configuration in FIG. 1 is an example, and various modifications are possible, such as performing AI inference and image analysis system processing on the camera 102 .
 解析サーバ101は、画像の解析を行う画像解析装置として適用でき、その構成や処理の内容は後述する。 The analysis server 101 can be applied as an image analysis device that analyzes images, and its configuration and processing details will be described later.
 カメラ102は、レンズや絞りを介して撮像素子に入射光を結像して情報を得るカメラの構成を適用できる。ここでの撮像素子の例としては、CCD(Charge-Coupled Device)イメージセンサやCMOS(Complementary Metal Oxide Semiconductor)イメージセンサ等があげられる。カメラ102は、映像として、例えば、1秒間に3フレーム(3fps)以上等で撮影して、その情報は、解析サーバ101やデータベースサーバ103へ送られる。カメラ102は、状況に応じて複数設置可能であり、様々な場所に配置可能である。例えば、監視カメラとして監視箇所に配置する等である。 For the camera 102, a configuration of a camera in which information is obtained by forming an image of incident light on an imaging device via a lens and an aperture can be applied. Examples of the imaging device here include a CCD (Charge-Coupled Device) image sensor and a CMOS (Complementary Metal Oxide Semiconductor) image sensor. The camera 102 shoots an image at, for example, 3 frames per second (3 fps) or more, and the information is sent to the analysis server 101 and the database server 103 . A plurality of cameras 102 can be installed according to the situation, and can be arranged in various places. For example, it may be installed at a monitoring location as a monitoring camera.
 データベースサーバ103は、カメラ102で撮影した画像や解析サーバ101の処理のために必要な情報、解析サーバ101の処理結果等を記録する装置である。記録するための装置としては、例えば、HDD(Hard Disk Drive)、SSD(Solid State Drive)、DDS(Digital Data Storage)等、必要に応じて適した方式を適用できる。 The database server 103 is a device that records images captured by the camera 102, information necessary for processing by the analysis server 101, processing results of the analysis server 101, and the like. As a device for recording, for example, HDD (Hard Disk Drive), SSD (Solid State Drive), DDS (Digital Data Storage), etc., can be applied according to need.
(画像解析装置のハードウェア構成)
 図2は、本発明の画像解析装置の一例を示すハードウェア構成図である。図2を用いて解析サーバ101のハードウェア構成例を説明する。
(Hardware configuration of image analysis device)
FIG. 2 is a hardware configuration diagram showing an example of the image analysis apparatus of the present invention. A hardware configuration example of the analysis server 101 will be described with reference to FIG.
 ハードウェアとしては、CPU(Central Processing Unit)等の処理装置を備えた電子計算機システムにより構成され、それぞれの機能が実行されるようになっている。処理装置としては、CPUの他に、Digital Signal Processor(DSP)やField-Programmable Gate Array(FPGA)、Graphics Processing Unit(GPU)などを適用してもよい。解析サーバ101は、プロセッサ部201、主記憶部202、補助記憶部203、入出力インターフェース部204、表示インターフェース部205、ネットワークインターフェース部206を含み、これらはバス207により結合されている。入出力インターフェース部204は、キーボードやマウス等の入力装置208に接続されてユーザインタフェースを提供する。表示インターフェース部205は、表示出力装置209に接続される。ネットワークインターフェース部206は解析サーバ101とネットワーク104を接続するためのインターフェースである。 The hardware consists of a computer system equipped with a processing unit such as a CPU (Central Processing Unit), and each function is executed. As the processing device, in addition to the CPU, a Digital Signal Processor (DSP), a Field-Programmable Gate Array (FPGA), a Graphics Processing Unit (GPU), etc. may be applied. The analysis server 101 includes a processor section 201 , a main storage section 202 , an auxiliary storage section 203 , an input/output interface section 204 , a display interface section 205 and a network interface section 206 , which are connected by a bus 207 . An input/output interface unit 204 is connected to an input device 208 such as a keyboard and mouse to provide a user interface. The display interface unit 205 is connected to the display output device 209 . A network interface unit 206 is an interface for connecting the analysis server 101 and the network 104 .
 補助記憶部203は通常、HDDやフラッシュメモリなどの不揮発性メモリで構成され、解析サーバ101が実行するプログラムやプログラムが処理対象とするデータ等を記憶する。主記憶部202はRAMで構成され、プロセッサ部201の命令により、プログラムやプログラムの実行に必要なデータ等を一時的に記憶する。プロセッサ部201は、補助記憶部203から主記憶部202にロードしたプログラムを実行する。なお、図2の構成は一例であり、様々な変形が可能である。 The auxiliary storage unit 203 is usually composed of a non-volatile memory such as an HDD or flash memory, and stores programs executed by the analysis server 101 and data to be processed by the programs. The main storage unit 202 is composed of a RAM, and temporarily stores programs, data necessary for executing the programs, and the like according to instructions from the processor unit 201 . The processor unit 201 executes programs loaded from the auxiliary storage unit 203 to the main storage unit 202 . Note that the configuration in FIG. 2 is an example, and various modifications are possible.
(画像解析装置の機能ブロック)
 図3は、本発明の画像解析装置の一例を示す機能ブロック図である。図3を用いて解析サーバ101の機能ブロックを説明する。
(Functional block of image analysis device)
FIG. 3 is a functional block diagram showing an example of the image analysis device of the present invention. Functional blocks of the analysis server 101 will be described with reference to FIG.
 解析サーバ101は補助記憶部203、画像取得部301、補正用データ取得部302、画像処理部303、記憶制御部304、表示制御部305で構成される。画像取得部301は補助記憶部203から得られる信号を画像として取得する。補正用データ取得部302は補助記憶部203から得られる信号を画像、または時系列データとして取得する。画像処理部303は画像取得部301により得られた画像と補正用データ取得部302により得られたデータを入力として、AI推論処理、遮蔽検知処理を行いその結果からシステムの状態を判定する。記憶制御部304は画像処理部303の結果を用いて出力結果の記憶制御を行い補助記憶部203に保存する。表示制御部305は画像処理部303の結果や補助記憶部203に保存された情報の表示を制御し、表示出力装置209に出力する。 The analysis server 101 is composed of an auxiliary storage unit 203 , an image acquisition unit 301 , a correction data acquisition unit 302 , an image processing unit 303 , a storage control unit 304 and a display control unit 305 . The image acquisition unit 301 acquires the signal obtained from the auxiliary storage unit 203 as an image. A correction data acquisition unit 302 acquires the signal obtained from the auxiliary storage unit 203 as an image or time-series data. An image processing unit 303 receives the image obtained by the image obtaining unit 301 and the data obtained by the correction data obtaining unit 302, performs AI inference processing and shielding detection processing, and determines the state of the system from the results. A storage control unit 304 performs storage control of the output result using the result of the image processing unit 303 and stores it in the auxiliary storage unit 203 . A display control unit 305 controls the display of the result of the image processing unit 303 and the information stored in the auxiliary storage unit 203 and outputs the information to the display output device 209 .
 画像取得部301においては画像データが記憶されている映像記憶装置などから入力された映像信号から画像データとして取得する。この画像データにおいてはノイズやフリッカなどの影響を低減するために前処理として平滑化フィルタや輪郭強調フィルタや濃度変換などの処理を施してもよい。また用途に応じてRGBカラーやYUV、モノクロなどのデータ形式を選択してもよい。さらには処理コスト低減のために所定の大きさで画像データに縮小処理を施してもよい。 The image acquisition unit 301 acquires image data from a video signal input from a video storage device or the like in which image data is stored. In order to reduce the influence of noise, flicker, etc., the image data may be subjected to preprocessing such as a smoothing filter, an edge enhancement filter, and density conversion. A data format such as RGB color, YUV, or monochrome may be selected according to the application. Furthermore, in order to reduce the processing cost, the image data may be reduced to a predetermined size.
 図4は、図3における画像処理部303の一例を示す機能ブロック図である。次に図4を用いて画像処理部303の機能ブロックを説明する。 FIG. 4 is a functional block diagram showing an example of the image processing unit 303 in FIG. Next, functional blocks of the image processing unit 303 will be described with reference to FIG.
 画像処理部303はAI推論処理部401と遮蔽検知処理部402、推論結果補正部403で構成される。AI推論処理部401は画像取得部301で取得した入力画像に対してAIを用いた推論処理(例:物体検知)を行う。また遮蔽検知処理部402は画像取得部301で取得した入力画像および補正用データ取得部302で取得した補正用データ(背景画像や直前のフレームの画像、推論結果等)を用いて、補正条件を元に検知対象物体の有無を決定する。推論結果補正部403では、AI推論処理部401および遮蔽検知処理部402で得られた結果を用いて最終的な画像解析システムの判定を決定し、その結果を記憶制御部304に出力する。 The image processing unit 303 is composed of an AI inference processing unit 401 , a shielding detection processing unit 402 , and an inference result correction unit 403 . The AI inference processing unit 401 performs inference processing (eg, object detection) using AI on the input image acquired by the image acquisition unit 301 . The shielding detection processing unit 402 uses the input image acquired by the image acquisition unit 301 and the correction data acquired by the correction data acquisition unit 302 (background image, previous frame image, inference result, etc.) to determine the correction conditions. Determine the presence or absence of the object to be detected. The inference result correction unit 403 uses the results obtained by the AI inference processing unit 401 and the shielding detection processing unit 402 to determine the final determination of the image analysis system, and outputs the result to the storage control unit 304 .
 AIによる推論処理は、例えば、ニューラルネットワークやディープラーニング等を用いて特徴量を抽出する処理等であり、画像内の検知対象に関する特徴量などから、確信度を推定できる。具体例としては、CNN(Convolution Neural Networks)等を用いることができる。 Inference processing by AI is, for example, the process of extracting feature values using neural networks, deep learning, etc., and the degree of certainty can be estimated from the feature values related to the detection target in the image. As a specific example, CNN (Convolution Neural Networks) or the like can be used.
(補正条件を用いない場合の処理)
 図5、6は、補正条件を用いない場合の処理の流れの一例を説明する図である。図5では、「イベントなし」から「イベント発生」に遷移する例を示している。図6では、「イベント発生」から「イベントなし」に遷移する例を示している。
(Process when correction conditions are not used)
5 and 6 are diagrams for explaining an example of the flow of processing when correction conditions are not used. FIG. 5 shows an example of transition from "no event" to "event occurrence". FIG. 6 shows an example of transition from "event occurrence" to "no event".
 補正条件を用いない場合の処理は「イベントなし」(第1の状態)と、「イベント発生」(第2の状態)の2つの状態を判定可能である。ここでの「イベントなし」の状態は、検知対象物体の非検知の条件に基づき検知対象物体である人が画像内に存在が確認されない状態である。「イベント発生」の状態は、検知対象物体の検知の条件に基づき検知対象物体である人が画像内に存在していると判定される状態である。動画やライブ映像での運用時において、AIの推論結果を用いて第1の条件である「検知イベント発生条件」と第2の条件である「検知イベント継続条件」を組み合わせた遷移条件に基づいてシステムを状態遷移させることでシステムとしての検知結果を決定する。  Processing when no correction condition is used can determine two states: "no event" (first state) and "event occurrence" (second state). The "no event" state here is a state in which the existence of a person, who is a detection target object, is not confirmed in the image based on the non-detection condition of the detection target object. The “event occurrence” state is a state in which it is determined that a person, who is a detection target object, exists in the image based on the conditions for detecting the detection target object. When operating with video or live video, based on the transition condition that combines the first condition "detection event occurrence condition" and the second condition "detection event continuation condition" using the inference result of AI The detection result as a system is determined by changing the state of the system.
 ここで、「検知イベント発生条件」(第1の条件)は、「イベントなし」の状態から「イベント発生」の状態へ移行する条件である。「検知イベント継続条件」(第2の条件)は、「イベント発生」の状態から「イベント発生」が継続するための条件である。図5、6の具体的な条件例において、「検知イベント発生条件」(第1の条件)は、物体検知時の確信度60%以上のフレームが連続3枚中3枚存在する場合である。また、「検知イベント継続条件」(第2の条件)は、物体検知時の確信度60%以上のフレームが連続3枚中1枚存在する場合である。 Here, the "detection event occurrence condition" (first condition) is a condition for transitioning from the "no event" state to the "event occurrence" state. The 'detection event continuation condition' (second condition) is a condition for 'event occurrence' to continue from the 'event occurrence' state. In the specific condition examples shown in FIGS. 5 and 6, the “detection event occurrence condition” (first condition) is the case where there are 3 out of 3 consecutive frames with a certainty of 60% or higher when an object is detected. Further, the "detection event continuation condition" (second condition) is a case where one out of three consecutive frames has a confidence of 60% or more at the time of object detection.
 図5、6では、時系列的に連続したフレームに対して物体検知処理をかける。図5の例では、最初は「イベントなし」の状態である。そして、画像のフレーム12、13、14において、検知対象である検知枠で囲まれた人の確信度が80%、85%、85%で推移している。この場合、物体検知時の確信度60%以上のフレームが連続3枚中3枚存在することになり、第1の条件である「検知イベント発生条件」を満たす。このため、フレーム14において、「イベントなし」の状態から「イベント発生」の状態に移行する。これ以降第2の条件である「検知イベント継続条件」を満たす場合、「イベント発生」状態となる。  In Figures 5 and 6, object detection processing is applied to consecutive frames in time series. In the example of FIG. 5, the initial state is "no event". In frames 12, 13, and 14 of the image, the degree of certainty of the person surrounded by the detection frame, which is the detection target, changes between 80%, 85%, and 85%. In this case, there are 3 out of 3 consecutive frames with a degree of confidence of 60% or higher when an object is detected, which satisfies the first condition, the "detection event generation condition." Therefore, in frame 14, the state of "no event" is changed to the state of "event occurrence". After that, when the second condition, ie, the condition for continuation of the detection event, is satisfied, the state becomes an event occurrence state.
 図6の例では、最初は「イベント発生」の状態である。フレーム22、23、24において、検知対象である人の確信度が得られていない。この場合、物体検知時の確信度60%以上のフレームが連続3枚中1枚も存在せず、第2の条件である「検知イベント継続条件」を満たさない。このため、フレーム24において、「イベント発生」の状態から「イベントなし」状態に移行する。以降、第1の条件の「検知イベント発生条件」を満たすまで「イベントなし」の状態となる。 In the example of FIG. 6, the state is "event occurrence" at first. In frames 22, 23, and 24, the confidence factor of the person to be detected is not obtained. In this case, none of the three consecutive frames has a certainty of 60% or higher when an object is detected, and the second condition, the "detection event continuation condition," is not satisfied. Therefore, in frame 24, the state of "event occurred" is changed to the state of "no event". After that, the state is "no event" until the first condition "detection event generation condition" is satisfied.
(第1の補正条件による処理)
 図7~9は、本発明の画像解析装置の第1の補正条件による処理の流れを説明する図である。図7では、「イベントなし」から「イベント発生」に遷移する例を示している。図8では、「イベント発生」、「ニュートラル」、「イベント発生」の順に遷移する例を示している。図9では、「イベント発生」から「イベントなし」に遷移する例を示している。ここでの処理は、画像解析装置である解析サーバ101で行われる。
(Processing under the first correction condition)
7 to 9 are diagrams for explaining the flow of processing according to the first correction condition of the image analysis apparatus of the present invention. FIG. 7 shows an example of transition from "no event" to "event occurrence". FIG. 8 shows an example of transitions in the order of "event occurrence", "neutral", and "event occurrence". FIG. 9 shows an example of transition from "event occurrence" to "no event". The processing here is performed by the analysis server 101, which is an image analysis device.
 図7~9の第1の補正条件による処理では、補正条件を用いない図5、6と比べて「イベントなし」(第1の状態)と「イベント発生」(第2の状態)に加えて、「ニュートラル」(第3の状態)を備える点が異なる。「ニュートラル」は、検知対象物体である人が遮蔽されていると判定する状態である。ここでの処理は、補助情報(トラッキング技術等による解析結果)を用いてAIの推論結果を補正する。さらに、第1の条件と第2の条件に加えて、第3の条件である補正条件を組み合わせた処理を行う。このような遷移条件に基づいてシステムを状態遷移させる。 In the processing according to the first correction condition in FIGS. 7 to 9, in addition to "no event" (first state) and "event occurrence" (second state), compared to FIGS. , "neutral" (third state). "Neutral" is a state in which it is determined that a person who is a detection target object is shielded. In this process, auxiliary information (results of analysis by tracking technology or the like) is used to correct AI inference results. Furthermore, in addition to the first condition and the second condition, processing is performed by combining the correction condition, which is the third condition. The state of the system is changed based on such transition conditions.
 図7~9では、トラッキング(追跡)時に検知対象物体を最後に検知したフレームにおいて、その物体の位置が事前に設定した補正範囲内であれば「補正条件」(第3の条件)を満たすとする。この場合、「検知対象物体の遮蔽あり」と判定して推論結果を補正するものとする。ここでの補正範囲の設定は、あらかじめ範囲を設定しておくことも可能である。例えば、固定カメラであれば、あらかじめ範囲を想定できる。また、補正範囲は、検知対象物体の移動状態から補正範囲を算出してもよい。そのフレームまでの移動量で検知対象物体が、画面の外へでていないことを推定して、補正範囲を決めることが可能である。例えば、それまでのフレームでの検知対象物体の移動速度がゆっくりなら、先のフレームでも画面内に存在している可能性が高いため、その移動速度に応じて補正範囲を広く設定する。 In FIGS. 7 to 9, in the frame in which the object to be detected was last detected during tracking, if the position of the object is within a preset correction range, the "correction condition" (third condition) is satisfied. do. In this case, it is determined that "the object to be detected is shielded" and the inference result is corrected. As for the setting of the correction range here, it is also possible to set the range in advance. For example, if it is a fixed camera, the range can be assumed in advance. Further, the correction range may be calculated from the moving state of the detection target object. It is possible to determine the correction range by estimating that the object to be detected is not out of the screen with the amount of movement up to that frame. For example, if the moving speed of the object to be detected in the previous frames is slow, there is a high possibility that the object will still exist in the screen in the previous frame.
 図7~9の具体的な条件例について説明する。「検知イベント発生条件」(第1の条件)は、物体検知時の確信度60%以上のフレームが連続3枚中3枚存在する場合である。また、「検知イベント継続条件」(第2の条件)は、物体検知時の確信度60%以上のフレームが連続3枚中1枚存在する場合である。また「補正条件」(第3の条件)は、トラッキング時に検知対象物体を最後に検知したフレームにおいて、その物体の位置が事前に設定した補正範囲内であれば条件を満たすとする。さらに、「補正条件」(第3の条件)を満たした場合「ニュートラル」(第3の状態)に遷移する。この場合、補正フレームである3フレーム以内に「検知イベント継続条件」(第2の条件)を満たす場合は「イベント発生」状態にし、3フレーム以内に「検知イベント継続条件」を満たさない場合は「イベントなし」状態に遷移する。 Specific example conditions in FIGS. 7 to 9 will be explained. The "detection event occurrence condition" (first condition) is the case where there are 3 out of 3 consecutive frames with a degree of confidence of 60% or more at the time of object detection. Further, the "detection event continuation condition" (second condition) is a case where one out of three consecutive frames has a confidence of 60% or more at the time of object detection. Also, the "correction condition" (third condition) is satisfied if the position of the object is within a preset correction range in the last frame in which the object to be detected is detected during tracking. Further, when the "correction condition" (third condition) is satisfied, the transition is made to "neutral" (third state). In this case, if the "detection event continuation condition" (second condition) is satisfied within 3 frames, which are the correction frames, the state is changed to "event occurrence", and if the "detection event continuation condition" is not satisfied within 3 frames, " transition to the "no event" state.
 図7~9では、時系列的に連続したフレームに対して物体検知処理をかける。図7の例では、最初は「イベントなし」の状態である。そして、画像フレーム32、33、34において、検知対象である検知枠で囲まれた人の確信度が80%、85%、85%で推移している。この場合、物体検知時の確信度60%以上のフレームが連続3枚中3枚存在することになり、第1の条件である「検知イベント発生条件」を満たす。このため、フレーム34において、「イベントなし」の状態から「イベント発生」状態に移行する。  In Figures 7 to 9, object detection processing is applied to consecutive frames in time series. In the example of FIG. 7, the initial state is "no event". In image frames 32, 33, and 34, the degree of certainty of the person surrounded by the detection frame, which is the detection target, changes between 80%, 85%, and 85%. In this case, there are 3 out of 3 consecutive frames with a degree of confidence of 60% or higher when an object is detected, which satisfies the first condition, the "detection event generation condition." Therefore, in frame 34, the state of "no event" is changed to the state of "event occurred".
 図8の例では、最初のフレーム41は「イベント発生」の状態である。フレーム42、43、44で検知対象である人の確信度が得られていないため確信度が60%未満である。フレーム42、43は、「検知イベント継続条件」(第2の条件)を満たすため、「イベント発生」となる。そして、フレーム44においては、第2の条件である「検知イベント継続条件」を満たさない。さらに、トラッキング時に最後に検知した物体は、確信度60%以上である、フレーム41の物体41aとなる。ここでの物体41aは、設定された補正範囲41b内に存在する。このため、フレーム44は、「補正条件」(第3の条件)を満たし、「ニュートラル」状態(第3の状態)に移行する。この「ニュートラル」状態にて補正フレーム(図7~9では3フレーム)以内に「検知イベント継続条件」(第2の条件)を満たした場合、「イベント発生」状態に移行する。図8では、フレーム45において、確信度80%で検知対象物体が検知されるため、「検知イベント継続条件」を満たすことになる。 In the example of FIG. 8, the first frame 41 is in the "event occurrence" state. Frames 42, 43, and 44 do not have the confidence of the person to be detected, so the confidence is less than 60%. Since the frames 42 and 43 satisfy the "detection event continuation condition" (second condition), they are "event occurrence". In frame 44, the second condition, ie, the "detection event continuation condition" is not satisfied. Furthermore, the last detected object during tracking is the object 41a in the frame 41 with a certainty of 60% or higher. The object 41a here exists within the set correction range 41b. Therefore, the frame 44 satisfies the "correction condition" (third condition) and transitions to the "neutral" state (third state). In this "neutral" state, if the "detection event continuation condition" (second condition) is satisfied within the correction frame (three frames in FIGS. 7 to 9), the state shifts to the "event occurrence" state. In FIG. 8, in frame 45, the detection target object is detected with a certainty of 80%, so the "detection event continuation condition" is satisfied.
 図9では、フレーム51において、物体51aは、設定された補正範囲51b内に存在しない。このため、フレーム54で「検知イベント継続条件」(第2の条件)を満たさなくなった場合に「補正条件」(第3の条件)を満たさない。すなわち、フレーム54では、第2の条件と第3の条件の両方を満たさないため、「ニュートラル」状態を介さず「イベントなし」状態に移行する。 In FIG. 9, in the frame 51, the object 51a does not exist within the set correction range 51b. Therefore, when the "detection event continuation condition" (second condition) is no longer satisfied at frame 54, the "correction condition" (third condition) is not satisfied. That is, at frame 54, since both the second condition and the third condition are not satisfied, the transition is made to the "no event" state without going through the "neutral" state.
(第2の補正条件による処理)
 図10、11は、本発明の画像解析装置の第2の補正条件による処理の流れを説明する図である。図10では、「イベント発生」「ニュートラル」「イベント発生」の順に遷移する例を示している。図11では、「イベント発生」「ニュートラル」「イベントなし」の順に遷移する例を示している。ここでの処理は、画像解析装置である解析サーバ101で行われる。
(Processing under the second correction condition)
10 and 11 are diagrams for explaining the flow of processing according to the second correction condition of the image analysis apparatus of the present invention. FIG. 10 shows an example of transition in order of "event occurrence", "neutral", and "event occurrence". FIG. 11 shows an example of transition in order of "event occurrence", "neutral", and "no event". The processing here is performed by the analysis server 101, which is an image analysis device.
 図10、11の第2の補正条件による処理では、図7~9の第1の補正条件と同様に、「イベントなし」(第1の状態)、「イベント発生」(第2の状態)、「ニュートラル」(第3の状態)を備える。ここでの処理は、補助情報(背景差分技術等による解析結果)を用いてAIの推論結果を補正する。 10 and 11, similarly to the first correction conditions shown in FIGS. A "neutral" (third state) is provided. In this process, auxiliary information (results of analysis by background subtraction technology or the like) is used to correct AI inference results.
 また図10、11の第2の補正条件による処理は、補正方法として背景差分技術を用いたAIの推論結果の補正例である。ここでは、第2の条件である「検知イベント継続条件」を満たさなくなったフレームにて事前に撮影した背景画像との差分を取り、ある閾値で2値化する。その差分領域が最後に検知対象物体を検知したフレームの検知枠の領域よりも大きい場合、「対象物体の遮蔽あり」と判定する。この場合、そのフレームは、「補正条件」(第3の条件)を満たし、「ニュートラル」状態に遷移する。 The processing under the second correction condition in FIGS. 10 and 11 is an example of correction of AI inference results using background subtraction technology as a correction method. Here, the difference from the background image captured in advance in the frame that no longer satisfies the second condition, the “detection event continuation condition,” is binarized with a certain threshold value. If the difference area is larger than the area of the detection frame of the frame in which the detection target object was last detected, it is determined that "the target object is shielded". In this case, the frame satisfies the "correction condition" (third condition) and transitions to the "neutral" state.
 図10、11の具体的な条件例について説明する。「検知イベント発生条件」(第1の条件)は、物体検知時の確信度60%以上のフレームが連続3枚中3枚存在する場合である。また、「検知イベント継続条件」(第2の条件)は、物体検知時の確信度60%以上のフレームが連続3枚中1枚存在する場合である。また「補正条件」(第3の条件)は、「検知イベント継続条件」を満たさなくなったフレームと背景画像の差分領域が、最後に検知した物体の検知枠より大きい場合は条件を満たすとする。さらに、「補正条件」(第3の条件)を満たした場合「ニュートラル」(第3の状態)に移行する。この場合、「検知対象物体の遮蔽あり」と判定して推論結果を補正するものとする。そして、補正フレームである3フレーム以内に「検知イベント継続条件」(第2の条件)を満たす場合は「イベント発生」状態に遷移し、3フレーム以内に「検知イベント継続条件」を満たさない場合は「イベントなし」状態に遷移する。 A specific example of conditions in FIGS. 10 and 11 will be explained. The "detection event occurrence condition" (first condition) is the case where there are 3 out of 3 consecutive frames with a degree of confidence of 60% or more at the time of object detection. Further, the "detection event continuation condition" (second condition) is a case where one out of three consecutive frames has a confidence of 60% or more at the time of object detection. The "correction condition" (third condition) is satisfied when the difference area between the frame that no longer satisfies the "detection event continuation condition" and the background image is larger than the detection frame of the last detected object. Further, when the "correction condition" (third condition) is satisfied, the state shifts to "neutral" (third state). In this case, it is determined that "the object to be detected is shielded" and the inference result is corrected. Then, if the "detection event continuation condition" (second condition) is satisfied within 3 frames, which are the correction frames, the state transitions to "event occurrence", and if the "detection event continuation condition" is not satisfied within 3 frames, Transition to the "no event" state.
 図10の例では、最初のフレーム61は「イベント発生」の状態である。フレーム62、63、64で検知対象である人の確信度が得られていないので確信度が60%未満である。フレーム62、63は、「検知イベント継続条件」(第2の条件)を満たすため、「イベント発生」となる。そして、フレーム64においては、第2の条件である「検知イベント継続条件」を満たさない。ここで、「補正条件」(第3の条件)を満たすかどうかが判定される。まず、事前に撮影された背景画像である検知前画像68と、今、判定しようとしているフレーム64の差分をとる。すると、フレーム64には、木しか映っていなかった検知前画像68と比べて、車と人が追加で写っている。このため、この車と人の部分が差分に相当して、差分画像69において白で表示されている。この白の面積は、イベントが発生していたフレーム61の枠で囲まれた検知対象物体の人の面積よりも、車と人が写っている分、大きくなる。このため、人の前に大きい遮蔽物があったことが推定され、「補正条件」を満たすと判断される。これにより、「ニュートラル」状態に移行する。この「ニュートラル」状態にて補正フレーム(図10、11では3フレーム)以内に「検知イベント継続条件」(第2の条件)を満たした場合、「イベント発生」状態に移行する。図10では、フレーム65において、確信度80%で検知対象物体が検知されるため、「検知イベント継続条件」を満たし、「イベント発生」状態に遷移する。 In the example of FIG. 10, the first frame 61 is in the "event occurrence" state. In frames 62, 63, and 64, the confidence of the person to be detected is not obtained, so the confidence is less than 60%. Since the frames 62 and 63 satisfy the "detection event continuation condition" (second condition), they are "event occurrence". In frame 64, the second condition, ie, the "detection event continuation condition" is not satisfied. Here, it is determined whether or not the "correction condition" (third condition) is satisfied. First, the difference between the pre-detection image 68, which is a background image captured in advance, and the frame 64 to be determined now is obtained. Frame 64 now shows additional cars and people compared to pre-detection image 68 where only trees were shown. Therefore, the car and person portions correspond to the difference and are displayed in white in the difference image 69 . This white area is larger than the area of the person of the detection target object surrounded by the frame of the frame 61 where the event occurred, because the car and the person are shown. Therefore, it is presumed that there was a large shielding object in front of the person, and it is determined that the "correction condition" is satisfied. This transitions to the "neutral" state. In this "neutral" state, if the "detection event continuation condition" (second condition) is satisfied within the correction frame (three frames in FIGS. 10 and 11), the state shifts to the "event occurrence" state. In FIG. 10, in frame 65, the detection target object is detected with a certainty of 80%, so the "detection event continuation condition" is satisfied, and the state transitions to the "event occurrence" state.
 図11では、フレーム71~74の処理は、図10と同様であり、フレーム74では、「補正条件」(第3の条件)を満たし、「ニュートラル」状態に移行する。しかし、フレーム74以降に2フレーム分は、検知対象物体の検知がなされていないため確信度が60%未満となる。このため、「ニュートラル」状態にて補正フレーム(図10、11では3フレーム)以内に「検知イベント継続条件」を満たさない。このため、フレーム75では、「イベントなし」状態に移行する。 In FIG. 11, the processing of frames 71 to 74 is the same as in FIG. 10, and in frame 74, the "correction condition" (third condition) is satisfied and the state shifts to the "neutral" state. However, since the object to be detected is not detected for two frames after the frame 74, the certainty is less than 60%. Therefore, the "detection event continuation condition" is not satisfied within the correction frame (three frames in FIGS. 10 and 11) in the "neutral" state. Therefore, at frame 75, the transition to the "no event" state is made.
 図7~11以外の検知対象物体の遮蔽を検知する例について説明する。「検知イベント継続条件」(第2の条件)を満たさなくなったフレーム画像と最後に検知対象物体を検知したフレーム画像を入力して、検知対象物体の遮蔽の有無をディープラーニングで判定してもよい。ここで、検知対象物体が遮蔽されていると判定されれば、「補正条件」を満たすとして、図7~11と同様の処理を行うことができる。例えば、検知対象物体である人の前に車が通過した場合等である。また、この他、対象物体検知の検知枠が徐々に小さくなった時に「対象物体の遮蔽あり」と判定してもよい。遮蔽物があった場合に検知枠がだんだん小さくなってくることが想定される。その際に、所定以上、検知枠が小さくなった場合に、検知対象物体が遮蔽されていると判定して、「補正条件」を満たすとして、図7~11と同様の処理を行うことができる。 An example of detecting the shielding of a detection target object other than FIGS. 7 to 11 will be described. By inputting the frame image that no longer satisfies the "detection event continuation condition" (second condition) and the frame image that finally detected the detection target object, deep learning may be used to determine whether or not the detection target object is shielded. . Here, if it is determined that the object to be detected is shielded, the same processing as in FIGS. 7 to 11 can be performed on the assumption that the "correction condition" is satisfied. For example, there is a case where a car passes in front of a person who is a detection target object. In addition, it may be determined that "target object is shielded" when the detection frame for target object detection gradually becomes smaller. It is assumed that the detection frame will gradually become smaller when there is a shielding object. At that time, when the detection frame becomes smaller than a predetermined amount, it is determined that the object to be detected is occluded, and the "correction condition" is satisfied, and the same processing as in FIGS. 7 to 11 can be performed. .
(状態遷移の説明)
 図12は、本発明の画像解析装置の状態遷移の例を説明する図である。
(Description of state transition)
FIG. 12 is a diagram explaining an example of state transition of the image analysis apparatus of the present invention.
 矢印1は、「イベント発生」を継続することを示す。ここでの条件は、「検知イベント継続条件」(第2の条件)が該当する。 Arrow 1 indicates that "event occurrence" continues. The condition here corresponds to the "detection event continuation condition" (second condition).
 矢印2は、「イベント発生」から「ニュートラル」へ遷移することを示す。ここでの条件は、「検知イベント継続条件」(第2の条件)を満たさず、「補正条件」(第3の条件)を満たす場合が該当する。 Arrow 2 indicates transition from "event occurrence" to "neutral". The condition here corresponds to the case where the "detection event continuation condition" (second condition) is not satisfied but the "correction condition" (third condition) is satisfied.
 矢印3は、「ニュートラル」から「イベント発生」へ遷移することを示す。ここでの条件は、補正フレーム以内に「検知イベント継続条件」(第2の条件)を満たす場合が該当する。 Arrow 3 indicates transition from "neutral" to "event occurrence". The condition here corresponds to the case where the "detection event continuation condition" (second condition) is satisfied within the correction frame.
 矢印4は、「ニュートラル」から「イベントなし」へ遷移することを示す。ここでの条件は、補正フレーム以内に「検知イベント継続条件」(第2の条件)を満たさない場合が該当する。 Arrow 4 indicates a transition from "neutral" to "no event". The condition here corresponds to the case where the "detection event continuation condition" (second condition) is not satisfied within the correction frame.
 矢印5は、「イベント発生」から「イベントなし」へ遷移することを示す。ここでの条件は、「検知イベント継続条件」(第2の条件)及び「補正条件」(第3の条件)を満たさない場合が該当する。この場合は、「ニュートラル」を介さずに、「イベント発生」から「イベントなし」へ遷移する。 Arrow 5 indicates transition from "event occurrence" to "no event". The condition here corresponds to the case where the "detection event continuation condition" (second condition) and the "correction condition" (third condition) are not satisfied. In this case, the transition is made from "event occurrence" to "no event" without going through "neutral".
 矢印6は、「イベントなし」から「イベント発生」へ遷移することを示す。ここでの条件は、「検知イベント発生条件」(第1の条件)を満たす場合が該当する。 Arrow 6 indicates transition from "no event" to "event occurrence". The condition here corresponds to the case where the "detection event occurrence condition" (first condition) is satisfied.
 矢印7は、「イベントなし」を継続することを示す。ここでの条件は、「検知イベント発生条件」(第1の条件)を満たさない場合が該当する。 Arrow 7 indicates continuation of "no event". The condition here corresponds to the case where the "detection event occurrence condition" (first condition) is not satisfied.
 ここで、「検知イベント継続条件」(第1の条件)は、「検知イベント発生条件」(第2の条件)よりも達成が容易な緩い条件である。例えば、「検知イベント発生条件」が、物体検知時の検知対象物体の確信度X%以上のフレームが連続N枚中M枚存在する場合(N、Mは整数でN≧M)として、「検知イベント継続条件」は、物体検知時の検知対象物体の確信度Y%以上のフレームが連続N枚中P枚存在する場合(Pは整数でM>P)とする等である。また、両者の確信度は同じでもよいし、変更することも可能である。この場合X≧Yが好ましい。このように所定以上の確信度におけるフレームの出現確率を下げることで条件を緩くできる。 Here, the "detection event continuation condition" (first condition) is a loose condition that is easier to achieve than the "detection event occurrence condition" (second condition). For example, assuming that the "detection event generation condition" includes M frames out of N consecutive frames with a certainty of X% or more of the detection target object at the time of object detection (N and M are integers and N≧M), the "detection The “event continuation condition” is, for example, the case where there are P out of N consecutive frames (P is an integer and M>P) where the certainty of the detection target object at the time of object detection is Y% or higher. Moreover, the certainty factors of both may be the same, or may be changed. In this case X≧Y is preferred. In this way, the condition can be loosened by lowering the appearance probability of the frame at a certainty or more.
 ここで、矢印3で示した、「イベント発生」から「ニュートラル」へ遷移するための条件は、上記に限らず、「検知イベント発生条件」(第1の条件)よりも達成が容易な緩い条件であればよい。このため、「検知イベント継続条件」とは関係なしに、「検知イベント発生条件」よりも緩い条件を設定することができる。例えば、「検知イベント発生条件」が、物体検知時の確信度X%以上のフレームが連続N枚中M枚存在する場合(N、Mは整数でN≧M)として、「ニュートラル」から「イベント発生」へ遷移するための条件は、物体検知時の確信度Z%以上のフレームが連続N枚中Q枚存在する場合(Qは整数でM>Q)とする等である。この場合、矢印4の条件は、この条件を満たさない場合となる。また、両者の確信度は同じでもよいし、変更することも可能である。この場合X≧Zが好ましい。このように所定以上の確信度におけるフレームの出現確率を下げることで条件を緩くできる。 Here, the condition for transitioning from "event occurrence" to "neutral" indicated by arrow 3 is not limited to the above. If it is Therefore, it is possible to set a looser condition than the "detection event generation condition" regardless of the "detection event continuation condition". For example, if the “detection event occurrence condition” is that there are M frames out of N consecutive frames with a certainty of X% or more at the time of object detection (N and M are integers, N≧M), “neutral” to “event The condition for transitioning to "occurrence" is, for example, that there are Q out of N consecutive frames with a certainty Z% or more at the time of object detection (Q is an integer and M>Q). In this case, the condition of arrow 4 is a case where this condition is not satisfied. Moreover, the certainty factors of both may be the same, or may be changed. In this case X≧Z is preferred. In this way, the condition can be loosened by lowering the appearance probability of the frame at a certainty or more.
(比較例)
 図13は、補正条件がない場合と本発明の画像解析装置の結果を比較した一例である。図13の表を用いて本発明の画像解析装置の結果(補正条件ありの手法)と補正条件なしの手法の比較について述べる。前提として、フレームno.1からフレームno.15まで検知対象物体が画面上の検知範囲内に存在しており、フレームno.5における検知対象物体が補正条件を満たすとする。図中、「〇」がイベント発生、「×」がイベントなし、「△」がニュートラルの状態を示す。
(Comparative example)
FIG. 13 is an example of comparing the results of the image analysis apparatus of the present invention with the case where there is no correction condition. A comparison between the results of the image analysis apparatus of the present invention (method with correction conditions) and the method without correction conditions will be described using the table of FIG. As a premise, frame no. 1 to frame no. Up to frame no. 15, the object to be detected exists within the detection range on the screen. Assume that the object to be detected in 5 satisfies the correction conditions. In the figure, "O" indicates the occurrence of an event, "X" indicates no event, and "△" indicates a neutral state.
 補正条件なしの場合、第1の条件である「検知イベント発生条件」を満たすまで、フレームno.1、no.2は、「イベントなし」の状態である。その後、no.7までは、「検知イベント継続条件」を満たし「イベント発生」の状態となる。フレームno.8からフレームno.13までは、第2の条件である「検知イベント継続条件」を満たさなくなる。このため、フレームno.8からフレームno.13まで画像の状態が「イベントなし」の状態に遷移し、失報が発生する。no.14、no.15では、「検知イベント発生条件」を満たし「イベント発生」の状態となる。  In the case of no correction condition, frame no. 1, no. 2 is the state of "no event". After that, no. Up to 7, the "detection event continuation condition" is satisfied and the state is "event occurrence". frame no. 8 to frame no. Up to 13, the second condition "detection event continuation condition" is not satisfied. Therefore, frame no. 8 to frame no. Up to 13, the state of the image transits to the state of "no event", and a loss of alarm occurs. no. 14, no. In 15, the "detection event generation condition" is satisfied and the state becomes "event generation".
 これに対して本発明の画像解析装置(補正条件ありの手法)では、no.7までは、補正条件なしの手法と同様である。一方、第3の条件である「補正条件」を満たすフレームno.8では「ニュートラル」状態に遷移する。この状態で、物体検知確信度60%以上の物体があるフレームno.9で「検知イベント継続条件」(第2の条件)を満たして「イベント発生」状態に遷移する。この状態はno.15まで続く。このことで、no.9~13の失報を防ぐことができる。 On the other hand, in the image analysis apparatus of the present invention (method with correction conditions), no. Up to 7, it is the same as the method without correction conditions. On the other hand, frame no. 8 transitions to the "neutral" state. In this state, frame no. At 9, the "detection event continuation condition" (second condition) is satisfied, and the state changes to the "event occurrence" state. This state is no. Continues up to 15. With this, no. 9 to 13 misreports can be prevented.
(フローチャート)
 図14に本発明の画像解析装置の処理フローチャートの一例を示す。
(flowchart)
FIG. 14 shows an example of a processing flowchart of the image analysis apparatus of the present invention.
 初めに解析サーバ101のプロセッサ部201が補助記憶部203から主記憶部202にロードしたプログラムを実行して画像解析のシステムを起動する。画像解析のシステムはGUI(Graphical User Interface)でユーザーが結果を確認できるようにしてもよいし、「イベント発生」、「イベントなし」、「ニュートラル」の状態のみを確認できるようにしてもよい。 First, the processor unit 201 of the analysis server 101 executes the program loaded from the auxiliary storage unit 203 to the main storage unit 202 to activate the image analysis system. The image analysis system may allow the user to check the result on a GUI (Graphical User Interface), or may allow the user to check only the states of "event occurrence", "no event", and "neutral".
 また第1の条件である「検知イベント発生条件」、第2の条件である「検知イベント継続条件」、第3の条件である「補正条件」、物体検知の閾値、補正範囲等は予め用意した設定ファイルを読み込んで設定してもよい。また、ユーザーがGUIで選択できるようにしてもよい。なお、「検知イベント継続条件」(第2の条件)は「検知イベント発生条件」(第1の条件)より条件達成が容易であることが望ましい。また「補正条件」(第3の条件)は1つでもよいし、複数の条件を組み合わせて判定してもよい。 In addition, the first condition "detection event occurrence condition", the second condition "detection event continuation condition", the third condition "correction condition", the threshold for object detection, the correction range, etc. are prepared in advance. You can read the configuration file and set it. Alternatively, the user may be allowed to select using a GUI. It should be noted that it is desirable that the "detection event continuation condition" (second condition) is easier to achieve than the "detection event occurrence condition" (first condition). Also, one "correction condition" (third condition) may be used, or a combination of a plurality of conditions may be used for determination.
 システムを起動後、ステップ1001ではユーザーが検知対象とする物体および検知に使用するカメラ102を選択する。なお、検知対象とする物体の種類やカメラの数は1つでも複数でもよい。なお以降は検知対象とする物体およびカメラの数は1つの場合の例を説明する。 After starting the system, in step 1001, the user selects an object to be detected and the camera 102 to be used for detection. Note that the type of object to be detected and the number of cameras may be one or more. Hereinafter, an example in which the number of objects and cameras to be detected is one will be described.
 ステップ1002では、カメラ102から取得した画像を読み込む。ここで、カメラ102のフレームレートや画像サイズ等は、事前にカメラ102側で設定しておいてもよいし、ユーザーがGUIで任意に選択できるようにしてもよい。 At step 1002, the image acquired from the camera 102 is read. Here, the frame rate, image size, and the like of the camera 102 may be set in advance on the camera 102 side, or may be arbitrarily selected by the user through the GUI.
 ステップ1003では、終了コマンドが実行されたかどうかを判定する。終了コマンドが実行された場合、画像解析のシステムを終了する。実行されていない場合はステップ1004に進む。ここで、終了コマンドはキーボードの操作でもよいし、GUI上の操作でもよい。 At step 1003, it is determined whether the end command has been executed. When the termination command is executed, the image analysis system is terminated. If not, go to step 1004 . Here, the end command may be a keyboard operation or a GUI operation.
 ステップ1004では、読み込んだ画像を用いてAIの推論処理を行い、その結果を出力する。ここでのAIの推論処理はすでに説明した処理を用いることができる。 In step 1004, AI inference processing is performed using the read image, and the result is output. Already explained processing can be used for the inference processing of AI here.
 ステップ1005では、システムの状態がどの状態かを判定し、その後の処理を決定する。システムの状態が「イベントなし」の場合、ステップ1006に進む。一方、「イベント発生」の状態または「ニュートラル」の状態の場合、ステップ1007に進む。 At step 1005, the state of the system is determined, and subsequent processing is determined. If the state of the system is "no event", go to step 1006; On the other hand, in the case of the “event occurrence” state or the “neutral” state, the process proceeds to step 1007 .
 ステップ1006では、AIの推論結果が「検知イベント発生条件」(第1の条件)を満たすかどうかを判定し、「検知イベント発生条件」を満たす場合はステップ1007に進み、満たさない場合はステップ1002に進んで次のフレームの画像の読み込みを実施する。 At step 1006, it is determined whether the AI inference result satisfies the "detection event occurrence condition" (first condition). to read the image of the next frame.
 ステップ1007では、AIの推論結果が「検知イベント継続条件」(第2の条件)を満たすかどうかを判定し、「検知イベント継続条件」を満たす場合はステップ1008に、満たさない場合はステップ1010に進む。 In step 1007, it is determined whether the AI inference result satisfies the "detection event continuation condition" (second condition). move on.
 ステップ1008では、画像解析システムの状態を「イベント発生」に遷移する。 At step 1008, the state of the image analysis system is changed to "event occurrence".
 ステップ1009では、ユーザーに対してイベントの発生通知を行う。ここで、発生通知はGUI上で行ってもよいし、ネットワーク104を介する等して小型端末にイベント発生通知が届くようにしてもよい。イベントの発生通知を完了したらステップ1002に進んで次のフレームの画像の読み込みを実施する。 In step 1009, the user is notified of the occurrence of the event. Here, the occurrence notification may be made on the GUI, or the event occurrence notification may be delivered to the small terminal via the network 104 or the like. When the notification of the occurrence of the event is completed, the process advances to step 1002 to read the image of the next frame.
 ステップ1010では、画像解析システムの状態がどの状態かを判定し、その後の処理を決定する。画像解析システムの状態が「イベント発生」の場合、ステップ1011に進む。一方、「ニュートラル」の状態の場合、ステップ1012に進む。 At step 1010, the state of the image analysis system is determined, and subsequent processing is determined. If the state of the image analysis system is "event occurrence", the process proceeds to step 1011; On the other hand, in the case of the "neutral" state, the process proceeds to step 1012 .
 ステップ1011では、背景差分技術やトラッキング技術等を用いて「補正条件」(第3の条件)を満たすかどうかを判定し、「補正条件」を満たす場合はステップ1013に進み、満たさない場合はステップ1015に進む。 In step 1011, it is determined whether or not the "correction condition" (third condition) is satisfied using the background subtraction technique, the tracking technique, etc. If the "correction condition" is satisfied, the process proceeds to step 1013; Proceed to 1015.
 ステップ1012では、「補正条件」を満たすかどうかを判定し、「補正条件」を満たす場合はステップ1013に進み、「補正条件」を満たさない場合はステップ1015に進む。 In step 1012, it is determined whether or not the "correction condition" is satisfied.If the "correction condition" is satisfied, the process proceeds to step 1013, and if the "correction condition" is not satisfied, the process proceeds to step 1015.
 ステップ1013では、画像解析システムの状態を「ニュートラル」に遷移する。 At step 1013, the state of the image analysis system is changed to "neutral".
 ステップ1014では、ユーザーに対して「ニュートラル」状態であることを知らせるため、警告通知を行う。ここで、警告通知はGUI上で行ってもよいし、ネットワーク104を介する等して小型端末にイベント発生通知が届くようにしてもよい。警告通知を完了したらステップ1002に進んで次のフレームの画像の読み込みを実施する。 At step 1014, a warning notification is issued to notify the user of the "neutral" state. Here, the warning notification may be made on the GUI, or the event occurrence notification may be delivered to the small terminal via the network 104 or the like. After completing the warning notification, the process proceeds to step 1002 to read the image of the next frame.
 ステップ1015では、画像解析システムの状態を「イベントなし」に遷移し、ステップ1002に進んで次のフレームの画像の読み込みを実施する。 At step 1015, the state of the image analysis system is changed to "no event", and the process proceeds to step 1002 to read the image of the next frame.
 なお上記の全体処理の説明は具体的操作の一例であり、これに限るものではない。 The above description of the overall processing is an example of specific operations, and is not limited to this.
(効果)
 以上のように、上記の実施形態では、「検知イベント発生条件」よりも緩い条件(例えば「検知イベント継続条件」)を満たすと即座に「イベント発生」状態へ遷移する「ニュートラル」状態を有するようにした。これにより、AIの推論結果を補正することで遮蔽された検知対象物体に対して、物体検知の処理をより的確に行うことができる。同時に、検知対象物体が検知範囲外に出ているにもかかわらず「イベント発生」状態に留まり続ける誤報を防ぐことができる。このことで、失報や誤報を防ぐことができ、より精度の高い画像解析の処理を行うことが可能となる。
(effect)
As described above, in the above embodiment, when a condition looser than the "detection event occurrence condition" (for example, the "detection event continuation condition") is met, the "neutral" state immediately transitions to the "event occurrence" state. made it As a result, by correcting the inference result of AI, it is possible to more accurately perform object detection processing for a detection target object that has been shielded. At the same time, it is possible to prevent erroneous reporting in which the "event occurrence" state continues even though the object to be detected is out of the detection range. This makes it possible to prevent false alarms and false alarms, and to perform image analysis processing with higher accuracy.
 さらに、「ニュートラル」に移行する場合の条件として「補正条件」を設けることで、遮蔽された検知対象物体があることを、より正確に推定することが可能となる。すなわち、検知対象物体が、画面外に出ている状態であると誤認識することを防止できる。この時の補正条件は、トラッキング技術を用いた補正範囲の設定や、背景差分技術を用いることで、より適した補正条件とすることができる。 Furthermore, by setting a "correction condition" as a condition for shifting to "neutral", it is possible to more accurately estimate that there is a hidden object to be detected. That is, it is possible to prevent erroneous recognition that the object to be detected is out of the screen. The correction conditions at this time can be set to more suitable correction conditions by setting a correction range using a tracking technique or using a background subtraction technique.
 本発明の適用が想定される状況としては、動画やライブ映像での運用時に、検知対象物体を一部もしくは全部隠すような遮蔽物が含まれることが挙げられる。また検知対象物体の輝度・色合いの変化も撮影環境による遮蔽であると捉えることで、輝度・色合いが変化しやすい屋外等の環境にも適用可能である。 Situations in which the present invention is expected to be applied include obstacles that partially or completely hide the object to be detected during operation with moving images or live images. In addition, by recognizing changes in the brightness and color of the object to be detected as shielding due to the shooting environment, it can be applied to environments such as outdoors where brightness and color tend to change.
 以上の様に、本発明の実施形態について説明してきたが、本発明は上記した実施形態に限定されるものではなく、様々な変形例が含まれる。例えば、上記した実施形態は本発明を分かりやすく説明するために詳細に説明したものであり、必ずしも説明した全ての構成を備えるものに限定されるものではない。また、各実施形態の構成の一部について、他の構成の追加・削除・置換をすることが可能である。 As described above, the embodiments of the present invention have been described, but the present invention is not limited to the above-described embodiments, and includes various modifications. For example, the above-described embodiments have been described in detail in order to explain the present invention in an easy-to-understand manner, and are not necessarily limited to those having all the configurations described. Moreover, it is possible to add, delete, or replace part of the configuration of each embodiment with another configuration.
 例えば、上記の実施形態では、AIによる推論処理を用いることを説明した。しかし、AIによる推論処理を用いる以外のトラッキングの画像解析方法でも、上記の実施形態に適用することが可能である。この場合、この画像解析により検出する検知対象物体を同様に扱えばよい。 For example, in the above embodiment, it was explained that inference processing by AI was used. However, a tracking image analysis method other than using inference processing by AI can also be applied to the above embodiment. In this case, the object to be detected detected by this image analysis may be handled in the same way.
 また、解析サーバ101とカメラ102は別部材として説明したが、これらは一体に構成してもよい。すなわち、ネットワーク104を介さずカメラ102側に解析サーバ101の機能を有し画像解析装置として構成することができる。 Also, although the analysis server 101 and the camera 102 have been described as separate members, they may be integrated. That is, it can be configured as an image analysis apparatus having the function of the analysis server 101 on the side of the camera 102 without going through the network 104 .
101…解析サーバ、102…カメラ、103…データベースサーバ、104…ネットワーク、201…プロセッサ部、202…主記憶部、203…補助記憶部、204…入出力インターフェース部、205…表示インターフェース部、206…ネットワークインターフェース部、207…バス、208…入力装置、209…表示出力装置、301…画像取得部、302…補正用データ部、303…画像処理部、304…記憶制御部、305…表示制御部、401…AI推論処理部、402…遮蔽検知処理部、403…推論結果補正部 101... Analysis server 102... Camera 103... Database server 104... Network 201... Processor unit 202... Main storage unit 203... Auxiliary storage unit 204... Input/output interface unit 205... Display interface unit 206... Network interface unit 207 Bus 208 Input device 209 Display output device 301 Image acquisition unit 302 Correction data unit 303 Image processing unit 304 Storage control unit 305 Display control unit 401 ... AI inference processing unit, 402 ... shielding detection processing unit, 403 ... inference result correction unit

Claims (7)

  1.  撮影した画像を取得する画像取得部と、前記画像取得部により取得した入力画像に対して解析結果を出力する画像処理部と、補正条件に用いるデータを取得する補正用データ取得部とを備え、
     前記画像処理部は、検知対象物体の検知の条件に基づき前記検知対象物体が画像内に存在していると判定するイベント発生の状態と、検知対象物体の非検知の条件に基づき前記検知対象物体が画像内に存在しないと判定するイベントなしの状態と、前記補正条件を用いて前記検知対象物体が遮蔽されていると判定するニュートラルの状態とを、判別し、前記ニュートラルの状態から前記イベント発生の状態への遷移の条件は、前記イベントなしの状態から前記イベント発生の状態への遷移の条件よりも緩く設定されていることを特徴とする画像解析装置。
    An image acquisition unit that acquires a captured image, an image processing unit that outputs analysis results for the input image acquired by the image acquisition unit, and a correction data acquisition unit that acquires data used for correction conditions,
    The image processing unit determines an event occurrence state that the detection target object exists in an image based on a detection target object detection condition, and determines the detection target object non-detection condition based on the detection target object detection condition. is not present in the image, and a neutral state is determined by using the correction condition to determine that the detection target object is shielded, and the event occurrence is determined from the neutral state. wherein a condition for transition to the state of is set looser than a condition for transition from the no-event state to the event-occurred state.
  2.  請求項1に記載の画像解析装置において、
     前記イベント発生の状態を継続する条件は、前記イベントなしの状態から前記イベント発生の状態へ遷移する条件よりも緩く設定され、
     前記ニュートラルの状態へは、前記イベント発生の状態を継続できなくなったときに、前記補正条件を満たす場合に前記イベント発生の状態から遷移することを特徴とする画像解析装置。
    In the image analysis device according to claim 1,
    the condition for continuing the event occurrence state is set looser than the condition for transitioning from the no-event state to the event occurrence state;
    The image analysis apparatus, wherein the transition from the event occurrence state to the neutral state is made when the correction condition is satisfied when the event occurrence state cannot be continued.
  3.  請求項2に記載の画像解析装置において、
     前記補正条件は、検知対象物体を最後に検知した画像のフレームにおいて、その物体の位置が事前に設定した補正範囲内である場合に満たす条件であることを特徴とする画像解析装置。
    In the image analysis device according to claim 2,
    The image analysis apparatus, wherein the correction condition is a condition that is satisfied when the position of the detection target object is within a preset correction range in a frame of an image in which the detection target object is last detected.
  4.  請求項2に記載の画像解析装置において、
     前記補正条件は、前記イベント発生の状態を満たさなくなった画像のフレームと背景画像の差分領域が、前記検知対象物体を最後に検知したフレームにおける検知対象物体の検知枠より大きい場合に満たす条件であることを特徴とする画像解析装置。
    In the image analysis device according to claim 2,
    The correction condition is a condition that is satisfied when the difference area between the frame of the image that no longer satisfies the event occurrence state and the background image is larger than the detection frame of the detection target object in the frame in which the detection target object was last detected. An image analysis device characterized by:
  5.  カメラと、当該カメラで撮影した画像を取得する請求項1に記載の画像解析装置とを備え、
     前記カメラと前記画像解析装置は、ネットワークを介して通信可能であることを特徴とする画像解析システム。
    A camera and the image analysis device according to claim 1 for acquiring an image taken by the camera,
    An image analysis system, wherein the camera and the image analysis device are capable of communicating via a network.
  6.  処理装置を用いて画像解析を行う画像解析方法であって、
     撮影した画像を取得するステップと、
     補正条件に用いるデータを取得するステップと、
     検知対象物体の検知の条件に基づき前記検知対象物体が画像内に存在していると判定するイベント発生の状態と、検知対象物体の非検知の条件に基づき前記検知対象物体が画像内に存在しないと判定するイベントなしの状態と、前記補正条件を用いて前記検知対象物体が遮蔽されていると判定するニュートラルの状態とを、判別するステップとを備え、
     前記ニュートラルの状態から前記イベント発生の状態への遷移の条件は、前記イベントなしの状態から前記イベント発生の状態への遷移の条件よりも緩く設定されていることを特徴とする画像解析方法。
    An image analysis method for performing image analysis using a processing device,
    obtaining a captured image;
    a step of acquiring data used for correction conditions;
    An event occurrence state in which it is determined that the detection target object exists in the image based on a detection target object detection condition, and the detection target object does not exist in the image based on the detection target object non-detection condition. and a neutral state in which it is determined that the detection target object is blocked using the correction condition,
    An image analysis method, wherein a condition for transition from the neutral state to the event occurrence state is set looser than a condition for transition from the no-event state to the event occurrence state.
  7.  請求項6に記載の画像解析方法において、
     前記イベント発生の状態を継続する条件は、前記イベントなしの状態から前記イベント発生の状態へ遷移する条件よりも緩く設定され、
     前記ニュートラルの状態へは、前記イベント発生の状態を継続できなくなったときに、前記補正条件を満たす場合に前記イベント発生の状態から遷移することを特徴とする画像解析方法。
    In the image analysis method according to claim 6,
    the condition for continuing the event occurrence state is set looser than the condition for transitioning from the no-event state to the event occurrence state;
    An image analysis method, wherein transition from the event occurrence state to the neutral state is performed when the correction condition is satisfied when the event occurrence state cannot be continued.
PCT/JP2021/011313 2021-03-19 2021-03-19 Image analysis device, image analysis system, and image analysis method WO2022195837A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/JP2021/011313 WO2022195837A1 (en) 2021-03-19 2021-03-19 Image analysis device, image analysis system, and image analysis method
JP2023506655A JPWO2022195837A1 (en) 2021-03-19 2021-03-19

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2021/011313 WO2022195837A1 (en) 2021-03-19 2021-03-19 Image analysis device, image analysis system, and image analysis method

Publications (1)

Publication Number Publication Date
WO2022195837A1 true WO2022195837A1 (en) 2022-09-22

Family

ID=83320237

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2021/011313 WO2022195837A1 (en) 2021-03-19 2021-03-19 Image analysis device, image analysis system, and image analysis method

Country Status (2)

Country Link
JP (1) JPWO2022195837A1 (en)
WO (1) WO2022195837A1 (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2011243155A (en) * 2010-05-21 2011-12-01 Panasonic Corp Flow line creation device and flow line creation method
JP2016081095A (en) * 2014-10-10 2016-05-16 キヤノン株式会社 Subject tracking device, control method thereof, image-capturing device, display device, and program

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2011243155A (en) * 2010-05-21 2011-12-01 Panasonic Corp Flow line creation device and flow line creation method
JP2016081095A (en) * 2014-10-10 2016-05-16 キヤノン株式会社 Subject tracking device, control method thereof, image-capturing device, display device, and program

Also Published As

Publication number Publication date
JPWO2022195837A1 (en) 2022-09-22

Similar Documents

Publication Publication Date Title
US8041075B2 (en) Identifying spurious regions in a video frame
US8134596B2 (en) Classifying an object in a video frame
AU2009243442B2 (en) Detection of abnormal behaviour in video objects
JP5241782B2 (en) Surveillance camera system having a camera abnormality detection device
US8837781B2 (en) Video object fragmentation detection and management
JP6525545B2 (en) INFORMATION PROCESSING APPARATUS, INFORMATION PROCESSING METHOD, AND COMPUTER PROGRAM
US7460689B1 (en) System and method of detecting, recognizing, and tracking moving targets
JP4668921B2 (en) Object detection in images
US10713798B2 (en) Low-complexity motion detection based on image edges
JP5644097B2 (en) Image processing apparatus, image processing method, and program
JP7014161B2 (en) Image processing equipment, image processing systems, and image processing methods, and programs
CN104966304A (en) Kalman filtering and nonparametric background model-based multi-target detection tracking method
CN110647818A (en) Identification method and device for shielding target object
KR20160037480A (en) Method for establishing region of interest in intelligent video analytics and video analysis apparatus using the same
US8311345B2 (en) Method and system for detecting flame
KR101581162B1 (en) Automatic detection method, apparatus and system of flame, smoke and object movement based on real time images
WO2022195837A1 (en) Image analysis device, image analysis system, and image analysis method
JP5710230B2 (en) Monitoring system and monitoring method
CN111784750A (en) Method, device and equipment for tracking moving object in video image and storage medium
US20190102888A1 (en) Image processing apparatus and method and monitoring system
JP2007510994A (en) Object tracking in video images
WO2020139071A1 (en) System and method for detecting aggressive behaviour activity
JP6308612B2 (en) Image processing apparatus, image processing method, and image processing program
JP2011123562A (en) Image processing apparatus, image processing method, and program
JP2022034699A (en) Information processing program, information processing method, and information processing device

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21931584

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2023506655

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21931584

Country of ref document: EP

Kind code of ref document: A1