US20180144206A1 - Setting apparatus, setting method, and storage medium - Google Patents
Setting apparatus, setting method, and storage medium Download PDFInfo
- Publication number
- US20180144206A1 US20180144206A1 US15/872,835 US201815872835A US2018144206A1 US 20180144206 A1 US20180144206 A1 US 20180144206A1 US 201815872835 A US201815872835 A US 201815872835A US 2018144206 A1 US2018144206 A1 US 2018144206A1
- Authority
- US
- United States
- Prior art keywords
- image
- detection
- detection region
- setting
- region
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G06K9/2081—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/52—Surveillance or monitoring of activities, e.g. for recognising suspicious objects
-
- G06K9/00771—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/22—Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition
- G06V10/235—Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition based on user input or interaction
Definitions
- the present invention relates to a setting apparatus and method which set a detection region corresponding to a detection process of detecting a change of an image within the detection region for an object of detection.
- a captured image may be analyzed, and when a specific event is detected, an image in which the event is occurring may be stored or may be viewed and checked by a person in charge.
- Such specific events to be analyzed may include removal detection which is detection of removal of an object of surveillance, for example.
- the present invention may reduce the number of steps of a region-setting procedure by an operator.
- the present invention may further prevent deterioration of accuracy of detection.
- An aspect of the present invention may provide a setting apparatus which sets a detection region for a detection process of detecting a change of an image within a detection region corresponding to an object of detection, the apparatus including an input unit usable for inputting a first image in which the object of detection is present and a second image in which the object of detection is not present, and a determination unit configured to determine the detection region from the first image and the second image such that the detection process may be performed on a detection region of a third image.
- Another aspect of the present invention may further provide a method for determining a detection region in a setting apparatus which sets a detection region for a detection process of detecting a change of an image within a detection region corresponding to an object of detection, the method including inputting a first image in which the object of detection is present and a second image in which the object of detection is not present, and determining the detection region from the first image and the second image such that the detection process may be performed on a detection region of a third image.
- FIG. 1 is a software configuration diagram according to an embodiment of the present invention.
- FIG. 2 is a hardware configuration diagram according to an embodiment of the present invention.
- FIG. 3 is a user interface according to an embodiment of the present invention.
- FIG. 4 illustrates processing steps according to an embodiment of the present invention.
- FIG. 5 is an explanatory diagram of process 1 according to an embodiment of the present invention.
- FIG. 6 is an explanatory diagram of process 2 according to an embodiment of the present invention.
- FIG. 7 is an explanatory diagram of process 3 according to an embodiment of the present invention.
- FIGS. 8A to 8C are explanatory diagrams of process 4 according to an embodiment of the present invention.
- FIG. 1 is a software configuration diagram of an object detection system which provides a removal detection function according to a first embodiment of the present invention.
- the object detection system includes a setting apparatus 101 , a network 102 , a network camera 103 , and a client terminal 120 .
- an object detection function setting tool runs which is configured to display a video image based on video data from the network camera 103 and provide a user interface for setting an object detection function for monitoring of removal of an object.
- the network 102 is a network that supports IP-based communication such as a LAN and the Internet.
- the network camera 103 has a video image capturing function and an object detection function and delivers a live image and detected object information regarding a detected object to the setting apparatus 101 and the client terminal 120 through the network 102 .
- the setting apparatus 101 and client terminal 120 display a video image based on video data delivered from the network camera 103 and a detection result based on detected object information.
- the network camera 103 transmits detected object information to the setting apparatus 101 .
- the setting apparatus 101 receives the detected object information
- the setting apparatus 101 displays a detection region frame and indicates that the object has been removed. In this case, the detection region frame is displayed in a different color from the color of the detection region frame while it is being set.
- the setting apparatus 101 is a display control apparatus configured to display a video image and detected object information.
- the network camera 103 has an image capturing unit 104 , an object detecting unit 105 , and a communication unit 106 .
- the object detecting unit 105 may be provided in the setting apparatus 101 instead of in the network camera 103 .
- the image capturing unit 104 acquires a live image by using an image capturing device.
- the image capturing unit 104 may be capable of acquiring a live image of 30 frames per second, for example.
- the image capturing unit 104 encodes a live image acquired from an image capturing device to a JPEG image and converts it to digital video data that may be delivered over a network. Such a conversion function may be provided in the communication unit 106 .
- the object detecting unit 105 analyzes the video data acquired by the image capturing unit 104 and detects whether an object within a detection region identified by the detection setting information set by the setting apparatus 101 has been removed or not. If a video image within the detection region has changed, the object detecting unit 105 detects that the object has been removed.
- the object detecting unit 105 of this embodiment detects removal of an object by analyzing a difference between consecutive video frames.
- Various methods may be applicable as the object removal detection algorithm, such as a method in which a difference between video data acquired by image capturing and a preset reference video image is analyzed.
- the communication unit 106 transmits video data acquired by the image capturing unit 104 and detected object information generated by the object detecting unit 105 to the client terminal 120 and setting apparatus 101 through the network 102 .
- the communication unit 106 further receives detection setting information set by a user by using the setting apparatus 101 and sets it in the object detecting unit 105 .
- the communication unit 110 receives video data and detected object information transmitted from the network camera 103 .
- the communication unit 110 further transmits detection setting information set by a user by using the setting apparatus 101 to the network camera 103 .
- the display control unit 111 displays a video image based on video data delivered from the network camera 103 on a display device such as a display.
- a detection frame setting unit 112 sets a detection region on a video image as detection setting information.
- detection region refers to a region where an object is to be detected within a video image.
- setting a detection region allows setting not all but a part of a video image (of an object to be monitored for removal) as an object of detection. From this, an improvement of accuracy of object detection and reduction of the processing load may be expected.
- the detection frame setting unit 112 causes the display control unit 111 to display a frame indicating a detection region on a video image.
- the display control unit 111 displays a video image from the network camera 103 in a video display region 302 ( FIG. 3 ).
- the display control unit 111 displays a detection region frame 303 within the video display region 302 .
- a display example by the display control unit 111 will be described below with reference to FIG. 3 .
- a detection result display unit 115 causes the display control unit 111 to display the detection region frame 303 based on a detection result from the object detecting unit 105 in the network camera 103 as a result of the determination on whether the object has been removed or not. In this manner, the setting apparatus 101 sets a detection region corresponding to an object of detection within a video image for detection of a change within the detection region.
- the client terminal 120 receives video data and detected object information from the network camera 103 and displays the object detection result and the video image.
- the client terminal 120 has the communication unit 110 , display control unit 111 , and detection result display unit 115 among the components of the setting apparatus 101 and does not have the detection frame setting unit 112 .
- FIG. 2 illustrates hardware configurations of the setting apparatus 101 and the network camera 103 .
- the client terminal 120 has the same hardware configuration as that of the setting apparatus 101 .
- a CPU 201 is a control processing device configured to control the setting apparatus 101 overall.
- a secondary storage device 202 stores a program allowing the CPU 201 to control the setting apparatus 101 .
- a RAM 203 is a memory configured to decompress a program read by the CPU 201 from the secondary storage device 202 and execute a process.
- the RAM 203 is also usable as a temporary memory functioning as a storage area for temporarily storing data to be processed.
- a network interface 204 is a circuit usable for communication through the network 102 .
- the network interface 204 is used for reception of video data and detected object information from the network camera 103 and transmission of detection setting information.
- a display apparatus 205 is a display device such as a display configured to display a video image based on video data, for example.
- the setting apparatus 101 may be integrated with the display apparatus 205 .
- An input device 206 may include a keyboard and a mouse, for example.
- the input device 206 may be a joystick, an audio input device or the like.
- the setting apparatus 101 may be configured as a general personal computer (PC).
- An object detection function setting tool configured to provide a user interface for setting an object detection function for an object of monitoring for removal may be implemented as software capable of running on a general PC. Having described that according to this embodiment the CPU 201 reads out a program from the secondary storage device 202 and executes processes, at least a part of processes of the components in FIG. 1 may be performed by special hardware.
- a CPU 210 is a control unit configured to perform overall control of the network camera 103 overall.
- a ROM 211 stores a program configured to allow the CPU 210 to control the network camera 103 .
- a secondary storage device equivalent to the secondary storage device 202 may be provided instead of the ROM 211 .
- a RAM 212 is a memory usable for decompressing a program read from the ROM 211 and executing a process.
- the RAM 212 is also usable as a temporary memory functioning as a storage area for temporarily storing data to be processed.
- a network interface 213 is a circuit usable for communication through the network 102 .
- the network interface 213 is used for transmission of video data and detected object information to the setting apparatus 101 and for reception of detection setting information.
- An image capturing apparatus 214 has an image capturing device such as a video camera configured to capture a live image as a moving image and still images.
- the network camera 103 and the image capturing apparatus 214 may be provided integrally or may be provided separately.
- the setting apparatus 101 executes an object detection function setting tool that is a computer program configured to display a video image from the network camera 103 on the display apparatus 205 .
- a user interface illustrated in FIG. 3 is displayed on a screen of the display apparatus 205 in FIG. 2 under control of the display control unit 111 in FIG. 1 .
- the setting apparatus 101 and the display apparatus 205 may be provided integrally or separately.
- the setting apparatus 101 is a display control apparatus configured to display a video image on the display apparatus 205 .
- a user interface display region 301 is a display region on a display of the display apparatus 205 .
- the user interface display region 301 may be displayed fully or may be displayed as one window.
- a video display region 302 is a region for displaying a video image (live image) based on video data delivered from the network camera 103 .
- a detection region frame 303 (frame indicated by a thin line in FIG. 3 ) is a frame defining a subject region of an object detection process by the object detecting unit 105 in accordance with a setting by the detection frame setting unit 112 .
- the detection region frame 303 is displayed by the display control unit 111 based on a setting by the detection frame setting unit 112 and the detection result display unit 115 .
- the detection frame setting unit 112 sets a detection region frame for an object of removal detection which is a subject of detection setting information, and the resulting frame is displayed as the detection region frame 303 .
- the size and form may be changed by selecting and dragging a vertex or vertices or a border of the detection region frame 303 by using the input device 206 such as a mouse.
- the detection frame setting unit 112 cancels a setting for the detection region frame being a subject of the detection setting information and stores an image displayed on the video display region 302 as a background in the RAM 203 .
- FIG. 4 illustrates processing steps for explaining processing of creating detection setting information for the detection region frame 303 performed by the setting apparatus 101 according to this embodiment.
- an object detection function setting tool operates which is configured to display a video image from the network camera 103 .
- the processing in FIG. 4 is performed in accordance with operations performed by an operator and the object detection function setting tool program recorded in the secondary storage device 202 and read to the RAM 203 and executed by the CPU 201 .
- at least a part of the processing performed by the CPU 201 and CPU 210 of the processing in FIG. 4 may be performed by special hardware.
- the setting apparatus 101 sets a detection region for an object of detection within a video image for detection of a change within the detection region in the following manner.
- a background acquisition process will be described with reference to FIG. 5 .
- an object such as a PC placed on the right side of a person shown in the video display region 302 in FIG. 6 is an object of removal detection.
- the object of removal detection is saved to a part which is not shown in the video display region 302 , and the detection region clear button 305 is then pressed ( FIG. 5 ).
- the detection frame setting unit 112 acquires an image captured by the image capturing unit 104 and shown in the video display region 302 and stores it as a background image in the RAM 203 .
- frames for a predetermined period of time such as 10 seconds
- a predetermined number of frames such as 300 frames
- the input frames may be simply averaged or may be weight averaged with a predetermined weight.
- This background creation may be performed by the detection frame setting unit 112 , or a background captured by the network camera 103 may be acquired by the detection frame setting unit 112 .
- the currently set detection region is cleared if a detection region is set and the detection region clear button 305 is then pressed.
- a process of acquiring an image including an object of detection will be described with reference to FIG. 6 .
- An object of removal detection is placed at a predetermined position within the video display region 302 , and the detection region set button 304 is then pressed.
- the detection frame setting unit 112 acquires an image captured by the image capturing unit 104 and shown in the video display region 302 and stores it in the RAM 203 .
- frames for a predetermined period of time such as 10 seconds
- a predetermined number of frames such as 300 frames
- an image to be stored may be created based on the input frames.
- the input frames may be simply averaged or may be weight averaged with a predetermined weight.
- This background image creation may be performed by the detection frame setting unit 112 , or a background image captured by the network camera 103 may be acquired by the detection frame setting unit 112 .
- a process of extracting a detection region and vertices will be described with reference to FIG. 7 .
- the detection frame setting unit 112 handles the image stored in process 1 as a background image and extracts a a region in which there is a difference between the background image and the image stored in process 2.
- the images stored in process 1 and process 2 are divided into blocks, and a block having a difference is extracted as a difference region 310 .
- the difference region 310 may be extracted in the form of pixels.
- the detection frame setting unit 112 extracts a difference between a video image in which an object of detection is present and a video image in which it is not present. According to this embodiment, a polygon corresponding to a region determined as having a difference is created.
- Processes 1 to 3 may be executed in the network camera 103 , and a result therefrom may be acquired by the detection frame setting unit 112 .
- a detection region frame may be set by the network camera 103 .
- Thinning of vertices of a detection region will be described with reference to FIG. 8 .
- problems may occur including an increased load on the system, insufficient display capabilities and part of the resulting displayed video image not appearing.
- an inclusion ratio and a coverage are defined as will be described below, and threshold values are defined for the number of vertices, inclusion ratio, and coverage to reduce the number of vertices and loads on the system.
- Those threshold values may be held in a fixed manner in the secondary storage device 202 in the setting apparatus 101 , or an interface may be provided which allows them to be changed externally to the setting apparatus 101 .
- the inclusion ratio and the coverage may be expressed by Expression 1 and Expression 2:
- Inclusion Ratio an area of a region surrounded by the detection region frame 303 within the difference region 310 /an area of the difference region 310 (Expression 1)
- Coverage an area of the difference region 310 within the detection region frame 303 /an area of a region surrounded by the detection region frame 303 (Expression 2).
- the detection frame setting unit 112 calculates thin distances between the vertices 311 .
- the calculation of a thin distance according to this embodiment uses Douglas-Peucker algorithm.
- the detection frame setting unit 112 calculates thin distances between the vertices 311 illustrated in FIG. 8B , sorts the calculated thin distances between the vertices 311 in increasing order, and thins out a vertex having the shortest thin distance to reduce the number of vertices.
- the thin distance is the distance illustrated in FIG. 8B .
- the detection frame setting unit 112 connects the vertices across the reduced vertex and extracts the detection region frame 303 based on the remaining vertices ( FIG. 8A ).
- a thin distance of a vertex is a distance between a straight line connecting two vertices adjacent to the vertex and the vertex.
- the detection frame setting unit 112 reduces the number of vertices of a polygon calculated from a difference between a video image in which an object of detection is present and a video image in which it is not present to a predetermined number of vertices. In other words, the detection frame setting unit 112 deletes vertices in increasing order of distance between a straight line connecting two vertices adjacent to a vertex and the vertex such that the number of vertices may be made equal to or lower than a predetermined number of vertices. Thus, the detection frame setting unit 112 calculates a detection region (frame) which is a polygon having a number of vertices that is equal to or lower than the predetermined number of vertices.
- the detection frame setting unit 112 continues thinning out the vertices 311 until the number of vertices 311 becomes lower than a threshold value for the number of vertices stored in the RAM 203 (S 801 ). This thinning of vertices may be based on Minimum Distance method or Minimum Ratio method which is used for simplifying lines on a map. In this manner, the detection frame setting unit 112 calculates a detection region being a polygon having a predetermined number of vertices or less.
- the detection frame setting unit 112 acquires the inclusion ratio and coverage of the difference region 310 extracted in process 3 and the detection region frame 303 extracted in the step after the number of vertices 311 becomes lower than the threshold value for the number of vertices by using Expressions 1 and 2 above.
- the detection frame setting unit 112 If the inclusion ratio and coverage are lower than their corresponding predetermined threshold values stored in the RAM 203 (S 805 ), the detection frame setting unit 112 outputs a message to the display apparatus 205 and stops the processing (S 802 ).
- the detection frame setting unit 112 again calculates thin distances between the remaining vertices 311 that have not been thinned out and extracts a detection region resulting from the thinning out of the vertex with the shortest thin distance (S 803 ).
- the detection frame setting unit 112 calculates an inclusion ratio and coverage therefor. If the inclusion ratio and coverage are equal to or higher than the threshold values therefor (S 806 ), the detection frame setting unit 112 defines the detection region resulting from the vertex thinning as the detection region frame 303 to reduce the number of vertices 311 in a range where the inclusion ratio and coverage therefor are not lower than the threshold values (S 803 ).
- the detection frame setting unit 112 settles the detection region frame 303 created in step S 803 or S 801 as the final detection region frame 303 in process 4 (S 804 ).
- the display control unit 111 displays the positions of the vertices that have not been thinned out and remain within the extracted detection region frame 303 and lines between the vertices on the display apparatus 205 .
- step S 801 If it is determined in step S 801 that the number of vertices of a polygon corresponding to a region judged as having a difference in process 3 is equal to or lower than the threshold value and if it is determined in step S 806 that the inclusion ratio and coverage are lower than the threshold values when the number of vertices is reduced, the polygon corresponding to a region judged as having a difference in process 3 is defined as a detection region frame.
- the inclusion ratio and coverage may be calculated, and the number of vertices may be reduced in a range where the inclusion ratio and coverage are higher than the threshold values therefor to define the detection region frame 303 .
- the detection frame setting unit 112 calculates a first polygon from a difference between a video image in which an object of detection is present and a video image in which it is not present and calculates a detection region (frame) which is a second polygon acquired by reducing the number of vertices of the first polygon in a range where the second polygon overlaps the first polygon at a predetermined ratio or higher.
- Adjustment of a detection region will be described with reference to FIGS. 8A to 8C and 3 .
- the detection region frame 303 ( FIG. 8A ) extracted in process 4 is displayed on the display apparatus 205 under control of the display control unit 111 .
- the detection frame setting unit 112 changes the detection region frame 303 in response to an instruction given by changing a vertex or a line connecting vertices of the detection region frame 303 by using a pointing device such as a mouse.
- a pointing device such as a mouse.
- lines AB and AC connecting two vertices B and C adjacent to the vertex A and the vertex A are changed to connect the moved vertex A and the vertices B and C.
- the point caught with the mouse is defined as a new vertex F between the vertices D and E, and the line connecting the vertices D and E is changed to a line connecting the vertices D and F and a line connecting the vertices F and E.
- the vertex F is not added. If the number of vertices is higher than the threshold value when a vertex is added, a message notifying that the addition of the vertex is not allowed is displayed on the user interface display region 301 .
- the detection frame setting unit 112 displays the detection region (frame) on a video image as illustrated in FIG. 8A to change the detection region (frame) displayed on the video image.
- the detection frame setting unit 112 settles the detection region frame 303 if the detection region set button 304 is pressed ( FIG. 3 ). In a case where a user judges that adjustment is not necessary for the detection region frame calculated in process 4, the user does not perform the adjustment in process 5.
- the detection frame setting unit 112 transmits detection setting information to the object detecting unit 105 if the detection region frame 303 is settled.
- the object detecting unit 105 extracts an image of the detection region frame 303 set in the detection setting information from the image acquired in process 2 and compares the extracted image and the image captured by the image capturing unit 104 .
- the object detecting unit 105 may be provided in the setting apparatus 101 instead of in the network camera 103 .
- the detection frame setting unit 112 calculates a detection region corresponding to an object of detection from a difference between a video image in which the object of detection extracted in process 3 is present and a video image in which it is not present.
- Use of the function of setting a removal detection range according to this embodiment may allow setting of a removal detection region imposing less load on the system and having predetermined inclusion ratio and coverage or higher, which may eliminate the necessity for complicated operations, as described above.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Closed-Circuit Television Systems (AREA)
- Image Processing (AREA)
- Burglar Alarm Systems (AREA)
- Alarm Systems (AREA)
- Image Analysis (AREA)
Abstract
A setting apparatus which sets a detection region for a detection process of detecting a change of an image within a detection region corresponding to an object of detection inputs a first image in which the object of detection is present and a second image in which the object of detection is not present and determines the detection region from the first image and the second image such that the detection process may be performed on a detection region of a third image.
Description
- This application is a continuation, and claims the benefit, of U.S. patent application Ser. No. 14/290,688, presently pending and filed on May 29, 2014, and claims the benefit of, and priority to, Japanese Patent Application No. 2013-115668, filed May 31, 2013, which applications are hereby incorporated by reference herein in their entireties.
- The present invention relates to a setting apparatus and method which set a detection region corresponding to a detection process of detecting a change of an image within the detection region for an object of detection.
- In recent years, for shop surveillance or building surveillance, for example, a captured image may be analyzed, and when a specific event is detected, an image in which the event is occurring may be stored or may be viewed and checked by a person in charge. Such specific events to be analyzed may include removal detection which is detection of removal of an object of surveillance, for example.
- U.S. Patent Application Publication No. 2008/0018737, for example, discloses a function of detecting removal of an object of surveillance.
- However, in a case where an operator uses a pointing device, for example, to set a detection region corresponding to an object of detection within a video image, complicated operations are involved, imposing loads on the operator. Incorrect setting of such a detection region may result in lower accuracy of detection.
- The present invention may reduce the number of steps of a region-setting procedure by an operator.
- The present invention may further prevent deterioration of accuracy of detection.
- An aspect of the present invention may provide a setting apparatus which sets a detection region for a detection process of detecting a change of an image within a detection region corresponding to an object of detection, the apparatus including an input unit usable for inputting a first image in which the object of detection is present and a second image in which the object of detection is not present, and a determination unit configured to determine the detection region from the first image and the second image such that the detection process may be performed on a detection region of a third image.
- Another aspect of the present invention may further provide a method for determining a detection region in a setting apparatus which sets a detection region for a detection process of detecting a change of an image within a detection region corresponding to an object of detection, the method including inputting a first image in which the object of detection is present and a second image in which the object of detection is not present, and determining the detection region from the first image and the second image such that the detection process may be performed on a detection region of a third image.
- Further features of the present invention will become apparent from the following description of exemplary embodiments with reference to the attached drawings.
-
FIG. 1 is a software configuration diagram according to an embodiment of the present invention. -
FIG. 2 is a hardware configuration diagram according to an embodiment of the present invention. -
FIG. 3 is a user interface according to an embodiment of the present invention. -
FIG. 4 illustrates processing steps according to an embodiment of the present invention. -
FIG. 5 is an explanatory diagram ofprocess 1 according to an embodiment of the present invention. -
FIG. 6 is an explanatory diagram ofprocess 2 according to an embodiment of the present invention. -
FIG. 7 is an explanatory diagram ofprocess 3 according to an embodiment of the present invention. -
FIGS. 8A to 8C are explanatory diagrams ofprocess 4 according to an embodiment of the present invention. -
FIG. 1 is a software configuration diagram of an object detection system which provides a removal detection function according to a first embodiment of the present invention. - Referring to
FIG. 1 , the object detection system includes asetting apparatus 101, anetwork 102, anetwork camera 103, and aclient terminal 120. On thesetting apparatus 101, an object detection function setting tool runs which is configured to display a video image based on video data from thenetwork camera 103 and provide a user interface for setting an object detection function for monitoring of removal of an object. Thenetwork 102 is a network that supports IP-based communication such as a LAN and the Internet. - The
network camera 103 has a video image capturing function and an object detection function and delivers a live image and detected object information regarding a detected object to thesetting apparatus 101 and theclient terminal 120 through thenetwork 102. Thesetting apparatus 101 andclient terminal 120 display a video image based on video data delivered from thenetwork camera 103 and a detection result based on detected object information. - When removal of an object is detected, the
network camera 103 transmits detected object information to thesetting apparatus 101. When thesetting apparatus 101 receives the detected object information, thesetting apparatus 101 displays a detection region frame and indicates that the object has been removed. In this case, the detection region frame is displayed in a different color from the color of the detection region frame while it is being set. Thesetting apparatus 101 is a display control apparatus configured to display a video image and detected object information. Thenetwork camera 103 has animage capturing unit 104, anobject detecting unit 105, and acommunication unit 106. - It should be noted that the
object detecting unit 105 may be provided in thesetting apparatus 101 instead of in thenetwork camera 103. - The
image capturing unit 104 acquires a live image by using an image capturing device. Theimage capturing unit 104 may be capable of acquiring a live image of 30 frames per second, for example. Theimage capturing unit 104 according to this embodiment encodes a live image acquired from an image capturing device to a JPEG image and converts it to digital video data that may be delivered over a network. Such a conversion function may be provided in thecommunication unit 106. - The
object detecting unit 105 analyzes the video data acquired by theimage capturing unit 104 and detects whether an object within a detection region identified by the detection setting information set by thesetting apparatus 101 has been removed or not. If a video image within the detection region has changed, theobject detecting unit 105 detects that the object has been removed. - The
object detecting unit 105 of this embodiment detects removal of an object by analyzing a difference between consecutive video frames. Various methods may be applicable as the object removal detection algorithm, such as a method in which a difference between video data acquired by image capturing and a preset reference video image is analyzed. - The
communication unit 106 transmits video data acquired by theimage capturing unit 104 and detected object information generated by theobject detecting unit 105 to theclient terminal 120 and settingapparatus 101 through thenetwork 102. Thecommunication unit 106 further receives detection setting information set by a user by using thesetting apparatus 101 and sets it in theobject detecting unit 105. - The
communication unit 110 receives video data and detected object information transmitted from thenetwork camera 103. Thecommunication unit 110 further transmits detection setting information set by a user by using thesetting apparatus 101 to thenetwork camera 103. Thedisplay control unit 111 displays a video image based on video data delivered from thenetwork camera 103 on a display device such as a display. - A detection
frame setting unit 112 sets a detection region on a video image as detection setting information. The term “detection region” refers to a region where an object is to be detected within a video image. In other words, setting a detection region allows setting not all but a part of a video image (of an object to be monitored for removal) as an object of detection. From this, an improvement of accuracy of object detection and reduction of the processing load may be expected. - The detection
frame setting unit 112 causes thedisplay control unit 111 to display a frame indicating a detection region on a video image. - The
display control unit 111 displays a video image from thenetwork camera 103 in a video display region 302 (FIG. 3 ). Thedisplay control unit 111 displays adetection region frame 303 within thevideo display region 302. A display example by thedisplay control unit 111 will be described below with reference toFIG. 3 . - A detection
result display unit 115 causes thedisplay control unit 111 to display thedetection region frame 303 based on a detection result from theobject detecting unit 105 in thenetwork camera 103 as a result of the determination on whether the object has been removed or not. In this manner, thesetting apparatus 101 sets a detection region corresponding to an object of detection within a video image for detection of a change within the detection region. - The
client terminal 120 receives video data and detected object information from thenetwork camera 103 and displays the object detection result and the video image. In other words, theclient terminal 120 has thecommunication unit 110,display control unit 111, and detectionresult display unit 115 among the components of thesetting apparatus 101 and does not have the detectionframe setting unit 112. -
FIG. 2 illustrates hardware configurations of thesetting apparatus 101 and thenetwork camera 103. Theclient terminal 120 has the same hardware configuration as that of thesetting apparatus 101. Referring toFIG. 2 , aCPU 201 is a control processing device configured to control thesetting apparatus 101 overall. - A
secondary storage device 202 stores a program allowing theCPU 201 to control thesetting apparatus 101. ARAM 203 is a memory configured to decompress a program read by theCPU 201 from thesecondary storage device 202 and execute a process. TheRAM 203 is also usable as a temporary memory functioning as a storage area for temporarily storing data to be processed. - A
network interface 204 is a circuit usable for communication through thenetwork 102. Thenetwork interface 204 is used for reception of video data and detected object information from thenetwork camera 103 and transmission of detection setting information. - A
display apparatus 205 is a display device such as a display configured to display a video image based on video data, for example. Thesetting apparatus 101 may be integrated with thedisplay apparatus 205. Aninput device 206 may include a keyboard and a mouse, for example. Theinput device 206 may be a joystick, an audio input device or the like. - The
setting apparatus 101 may be configured as a general personal computer (PC). An object detection function setting tool configured to provide a user interface for setting an object detection function for an object of monitoring for removal may be implemented as software capable of running on a general PC. Having described that according to this embodiment theCPU 201 reads out a program from thesecondary storage device 202 and executes processes, at least a part of processes of the components inFIG. 1 may be performed by special hardware. - Next, a hardware configuration of the
network camera 103 will be described. Referring toFIG. 2 , aCPU 210 is a control unit configured to perform overall control of thenetwork camera 103 overall. - A
ROM 211 stores a program configured to allow theCPU 210 to control thenetwork camera 103. A secondary storage device equivalent to thesecondary storage device 202 may be provided instead of theROM 211. ARAM 212 is a memory usable for decompressing a program read from theROM 211 and executing a process. TheRAM 212 is also usable as a temporary memory functioning as a storage area for temporarily storing data to be processed. - A
network interface 213 is a circuit usable for communication through thenetwork 102. Thenetwork interface 213 is used for transmission of video data and detected object information to thesetting apparatus 101 and for reception of detection setting information. - An
image capturing apparatus 214 has an image capturing device such as a video camera configured to capture a live image as a moving image and still images. Thenetwork camera 103 and theimage capturing apparatus 214 may be provided integrally or may be provided separately. - Next, with reference to
FIG. 3 , a user interface of thesetting apparatus 101 will be described. Thesetting apparatus 101 executes an object detection function setting tool that is a computer program configured to display a video image from thenetwork camera 103 on thedisplay apparatus 205. A user interface illustrated inFIG. 3 is displayed on a screen of thedisplay apparatus 205 inFIG. 2 under control of thedisplay control unit 111 inFIG. 1 . Thesetting apparatus 101 and thedisplay apparatus 205 may be provided integrally or separately. Thesetting apparatus 101 is a display control apparatus configured to display a video image on thedisplay apparatus 205. - Referring to
FIG. 3 , a userinterface display region 301 is a display region on a display of thedisplay apparatus 205. The userinterface display region 301 may be displayed fully or may be displayed as one window. - A
video display region 302 is a region for displaying a video image (live image) based on video data delivered from thenetwork camera 103. - A detection region frame 303 (frame indicated by a thin line in
FIG. 3 ) is a frame defining a subject region of an object detection process by theobject detecting unit 105 in accordance with a setting by the detectionframe setting unit 112. Thedetection region frame 303 is displayed by thedisplay control unit 111 based on a setting by the detectionframe setting unit 112 and the detectionresult display unit 115. - When a detection region set
button 304 is pressed or clicked, the detectionframe setting unit 112 sets a detection region frame for an object of removal detection which is a subject of detection setting information, and the resulting frame is displayed as thedetection region frame 303. The size and form may be changed by selecting and dragging a vertex or vertices or a border of thedetection region frame 303 by using theinput device 206 such as a mouse. - When a detection region
clear button 305 is pressed or clicked, the detectionframe setting unit 112 cancels a setting for the detection region frame being a subject of the detection setting information and stores an image displayed on thevideo display region 302 as a background in theRAM 203. -
FIG. 4 illustrates processing steps for explaining processing of creating detection setting information for thedetection region frame 303 performed by thesetting apparatus 101 according to this embodiment. In thesetting apparatus 101, an object detection function setting tool operates which is configured to display a video image from thenetwork camera 103. The processing inFIG. 4 is performed in accordance with operations performed by an operator and the object detection function setting tool program recorded in thesecondary storage device 202 and read to theRAM 203 and executed by theCPU 201. However, at least a part of the processing performed by theCPU 201 andCPU 210 of the processing inFIG. 4 may be performed by special hardware. Thesetting apparatus 101 sets a detection region for an object of detection within a video image for detection of a change within the detection region in the following manner. - A background acquisition process will be described with reference to
FIG. 5 . In this example, an object such as a PC placed on the right side of a person shown in thevideo display region 302 inFIG. 6 is an object of removal detection. The object of removal detection is saved to a part which is not shown in thevideo display region 302, and the detection regionclear button 305 is then pressed (FIG. 5 ). In response to pressing of the detection regionclear button 305, the detectionframe setting unit 112 acquires an image captured by theimage capturing unit 104 and shown in thevideo display region 302 and stores it as a background image in theRAM 203. - In this process, frames for a predetermined period of time (such as 10 seconds) or a predetermined number of frames (such as 300 frames) may be input, and a background image may be created based on the input frames. The input frames may be simply averaged or may be weight averaged with a predetermined weight. This background creation may be performed by the detection
frame setting unit 112, or a background captured by thenetwork camera 103 may be acquired by the detectionframe setting unit 112. The currently set detection region is cleared if a detection region is set and the detection regionclear button 305 is then pressed. - A process of acquiring an image including an object of detection will be described with reference to
FIG. 6 . - An object of removal detection is placed at a predetermined position within the
video display region 302, and the detection region setbutton 304 is then pressed. In response to pressing of the detection region setbutton 304, the detectionframe setting unit 112 acquires an image captured by theimage capturing unit 104 and shown in thevideo display region 302 and stores it in theRAM 203. Also in this process, frames for a predetermined period of time (such as 10 seconds) or a predetermined number of frames (such as 300 frames) may be input, and an image to be stored may be created based on the input frames. The input frames may be simply averaged or may be weight averaged with a predetermined weight. This background image creation may be performed by the detectionframe setting unit 112, or a background image captured by thenetwork camera 103 may be acquired by the detectionframe setting unit 112. - In this manner, a video image in which an object of detection is present and a video image in which the object of detection is not present are input through the
communication unit 110 inprocesses - A process of extracting a detection region and vertices will be described with reference to
FIG. 7 . - The detection
frame setting unit 112 handles the image stored inprocess 1 as a background image and extracts a a region in which there is a difference between the background image and the image stored inprocess 2. InFIG. 7 , the images stored inprocess 1 andprocess 2 are divided into blocks, and a block having a difference is extracted as adifference region 310. Thedifference region 310 may be extracted in the form of pixels. In this manner, the detectionframe setting unit 112 extracts a difference between a video image in which an object of detection is present and a video image in which it is not present. According to this embodiment, a polygon corresponding to a region determined as having a difference is created. -
Processes 1 to 3 may be executed in thenetwork camera 103, and a result therefrom may be acquired by the detectionframe setting unit 112. A detection region frame may be set by thenetwork camera 103. - When the
difference region 310 is extracted, aline connecting vertices 311 surrounding the region is extracted. - Thinning of vertices of a detection region will be described with reference to
FIG. 8 . - In a case where the
difference region 310 extracted inprocess 3 has a complicated form, problems may occur including an increased load on the system, insufficient display capabilities and part of the resulting displayed video image not appearing. In order to solve such problems, an inclusion ratio and a coverage are defined as will be described below, and threshold values are defined for the number of vertices, inclusion ratio, and coverage to reduce the number of vertices and loads on the system. - Those threshold values may be held in a fixed manner in the
secondary storage device 202 in thesetting apparatus 101, or an interface may be provided which allows them to be changed externally to thesetting apparatus 101. - In this case, the inclusion ratio and the coverage may be expressed by
Expression 1 and Expression 2: -
Inclusion Ratio=an area of a region surrounded by thedetection region frame 303 within thedifference region 310/an area of the difference region 310 (Expression 1) -
Coverage=an area of thedifference region 310 within thedetection region frame 303/an area of a region surrounded by the detection region frame 303 (Expression 2). - The detection
frame setting unit 112 calculates thin distances between thevertices 311. The calculation of a thin distance according to this embodiment uses Douglas-Peucker algorithm. The detectionframe setting unit 112 calculates thin distances between thevertices 311 illustrated inFIG. 8B , sorts the calculated thin distances between thevertices 311 in increasing order, and thins out a vertex having the shortest thin distance to reduce the number of vertices. The thin distance is the distance illustrated inFIG. 8B . The detectionframe setting unit 112 connects the vertices across the reduced vertex and extracts thedetection region frame 303 based on the remaining vertices (FIG. 8A ). A thin distance of a vertex is a distance between a straight line connecting two vertices adjacent to the vertex and the vertex. - Thus, the detection
frame setting unit 112 reduces the number of vertices of a polygon calculated from a difference between a video image in which an object of detection is present and a video image in which it is not present to a predetermined number of vertices. In other words, the detectionframe setting unit 112 deletes vertices in increasing order of distance between a straight line connecting two vertices adjacent to a vertex and the vertex such that the number of vertices may be made equal to or lower than a predetermined number of vertices. Thus, the detectionframe setting unit 112 calculates a detection region (frame) which is a polygon having a number of vertices that is equal to or lower than the predetermined number of vertices. - A method of reducing the number of vertices will be described below with reference to
FIG. 8C . - The detection
frame setting unit 112 continues thinning out thevertices 311 until the number ofvertices 311 becomes lower than a threshold value for the number of vertices stored in the RAM 203 (S801). This thinning of vertices may be based on Minimum Distance method or Minimum Ratio method which is used for simplifying lines on a map. In this manner, the detectionframe setting unit 112 calculates a detection region being a polygon having a predetermined number of vertices or less. - The detection
frame setting unit 112 acquires the inclusion ratio and coverage of thedifference region 310 extracted inprocess 3 and thedetection region frame 303 extracted in the step after the number ofvertices 311 becomes lower than the threshold value for the number of vertices by usingExpressions - If the inclusion ratio and coverage are lower than their corresponding predetermined threshold values stored in the RAM 203 (S805), the detection
frame setting unit 112 outputs a message to thedisplay apparatus 205 and stops the processing (S802). - If the inclusion ratio and coverage are both equal to or higher than the threshold values (S806), the detection
frame setting unit 112 again calculates thin distances between the remainingvertices 311 that have not been thinned out and extracts a detection region resulting from the thinning out of the vertex with the shortest thin distance (S803). The detectionframe setting unit 112 calculates an inclusion ratio and coverage therefor. If the inclusion ratio and coverage are equal to or higher than the threshold values therefor (S806), the detectionframe setting unit 112 defines the detection region resulting from the vertex thinning as thedetection region frame 303 to reduce the number ofvertices 311 in a range where the inclusion ratio and coverage therefor are not lower than the threshold values (S803). - If one of the inclusion ratio and coverage is lower than the corresponding threshold value (S806), the detection
frame setting unit 112 settles thedetection region frame 303 created in step S803 or S801 as the finaldetection region frame 303 in process 4 (S804). - The
display control unit 111 displays the positions of the vertices that have not been thinned out and remain within the extracteddetection region frame 303 and lines between the vertices on thedisplay apparatus 205. - If it is determined in step S801 that the number of vertices of a polygon corresponding to a region judged as having a difference in
process 3 is equal to or lower than the threshold value and if it is determined in step S806 that the inclusion ratio and coverage are lower than the threshold values when the number of vertices is reduced, the polygon corresponding to a region judged as having a difference inprocess 3 is defined as a detection region frame. - Even when the number of
vertices 311 is not lower than the threshold value for the number of vertices, the inclusion ratio and coverage may be calculated, and the number of vertices may be reduced in a range where the inclusion ratio and coverage are higher than the threshold values therefor to define thedetection region frame 303. - In this manner, the detection
frame setting unit 112 calculates a first polygon from a difference between a video image in which an object of detection is present and a video image in which it is not present and calculates a detection region (frame) which is a second polygon acquired by reducing the number of vertices of the first polygon in a range where the second polygon overlaps the first polygon at a predetermined ratio or higher. - Adjustment of a detection region will be described with reference to
FIGS. 8A to 8C and 3 . - The detection region frame 303 (
FIG. 8A ) extracted inprocess 4 is displayed on thedisplay apparatus 205 under control of thedisplay control unit 111. - The detection
frame setting unit 112 changes thedetection region frame 303 in response to an instruction given by changing a vertex or a line connecting vertices of thedetection region frame 303 by using a pointing device such as a mouse. In a case where a vertex A is moved, lines AB and AC connecting two vertices B and C adjacent to the vertex A and the vertex A are changed to connect the moved vertex A and the vertices B and C. - In a case where one point F of a line connecting two vertices D and E is caught and moved with a mouse, the point caught with the mouse is defined as a new vertex F between the vertices D and E, and the line connecting the vertices D and E is changed to a line connecting the vertices D and F and a line connecting the vertices F and E. However, if the number of vertices is higher than the threshold value in step S801 in
FIG. 8C when the vertex F is added, the vertex F is not added. If the number of vertices is higher than the threshold value when a vertex is added, a message notifying that the addition of the vertex is not allowed is displayed on the userinterface display region 301. - The detection
frame setting unit 112 displays the detection region (frame) on a video image as illustrated inFIG. 8A to change the detection region (frame) displayed on the video image. - The detection
frame setting unit 112 settles thedetection region frame 303 if the detection region setbutton 304 is pressed (FIG. 3 ). In a case where a user judges that adjustment is not necessary for the detection region frame calculated inprocess 4, the user does not perform the adjustment in process 5. - The detection
frame setting unit 112 transmits detection setting information to theobject detecting unit 105 if thedetection region frame 303 is settled. Theobject detecting unit 105 extracts an image of thedetection region frame 303 set in the detection setting information from the image acquired inprocess 2 and compares the extracted image and the image captured by theimage capturing unit 104. - This allows detection of removal of an object of detection within the
detection region frame 303. Theobject detecting unit 105 may be provided in thesetting apparatus 101 instead of in thenetwork camera 103. - In this manner, the detection
frame setting unit 112 calculates a detection region corresponding to an object of detection from a difference between a video image in which the object of detection extracted inprocess 3 is present and a video image in which it is not present. - Use of the function of setting a removal detection range according to this embodiment may allow setting of a removal detection region imposing less load on the system and having predetermined inclusion ratio and coverage or higher, which may eliminate the necessity for complicated operations, as described above.
- While the present invention has been described with reference to exemplary embodiments, it is to be understood that the invention is not limited to the disclosed exemplary embodiments. The scope of the following claims is to be accorded the broadest interpretation so as to encompass all such modifications and equivalent structures and functions.
Claims (12)
1. A setting apparatus which sets a detection region for a detection process of detecting a change of an image within the detection region corresponding to an object to be detected, the apparatus comprising:
an interface configured to connect an image capture unit from which a first image in which the object to be detected is present and a second image in which the object to be detected is not present are input; and
a processor, operatively coupled to a memory, and configured to display a guide in a captured image captured by the image capture unit for setting the detection region, based on the first image and the second image, wherein the guide is changeable by user operation.
2. The setting apparatus according to claim 1 , wherein the processor displays the guide as a polygon with a number of vertices not more than a predetermined number.
3. The setting apparatus according to claim 1 , wherein the processor is configured to allow a user to set the displayed guide as the detection region.
4. The setting apparatus according to claim 1 , wherein the detection process detects that the object to be detected has been removed.
5. A method for determining a region in a setting apparatus which sets a detection region for detecting a change in the detection region of an image caused by a removal of an object which has existed in the detection region, the method comprising:
inputting a first image in which the object does not exist and a second image in which the object exists, from an image capturing unit; and
displaying a guide in a captured image captured by the image capture unit for setting the detection region based on the first image and the second image, wherein the guide is changeable by user operation.
6. The method according to claim 5 , wherein the guide is displayed as a polygon with a number of vertices not more than a predetermined number.
7. The method according to claim 5 , further comprising allowing a user to set the guide as the detection region.
8. The method according to claim 5 , wherein removal of the object to be detected is detected.
9. A non-transitory storage medium storing a computer program for setting a detection region for detecting a change in the detection region of an image caused by a removal of an object which has existed in the detection region, the computer program comprising:
inputting a first image does not exist and a second image in which the object exists, from an image capturing unit; and
displaying a guide in a captured image captured by the image capture unit for setting the detection region based on the first image and the second image, wherein the guide is changeable by user operation.
10. The non-transitory medium according to claim 9 , wherein the guide polygon is displayed with a number of vertices not more than a predetermined number.
11. The medium according to claim 9 , further comprising allowing a user to set the guide as the detection region.
12. The medium according to claim 9 , wherein removal of the object to be detected is detected.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/872,835 US20180144206A1 (en) | 2013-05-31 | 2018-01-16 | Setting apparatus, setting method, and storage medium |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2013-115668 | 2013-05-31 | ||
JP2013115668A JP6261190B2 (en) | 2013-05-31 | 2013-05-31 | Setting device and setting method |
US14/290,688 US9904865B2 (en) | 2013-05-31 | 2014-05-29 | Setting apparatus which sets a detection region for a detection process |
US15/872,835 US20180144206A1 (en) | 2013-05-31 | 2018-01-16 | Setting apparatus, setting method, and storage medium |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/290,688 Continuation US9904865B2 (en) | 2013-05-31 | 2014-05-29 | Setting apparatus which sets a detection region for a detection process |
Publications (1)
Publication Number | Publication Date |
---|---|
US20180144206A1 true US20180144206A1 (en) | 2018-05-24 |
Family
ID=51985152
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/290,688 Active 2035-05-02 US9904865B2 (en) | 2013-05-31 | 2014-05-29 | Setting apparatus which sets a detection region for a detection process |
US15/872,835 Abandoned US20180144206A1 (en) | 2013-05-31 | 2018-01-16 | Setting apparatus, setting method, and storage medium |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/290,688 Active 2035-05-02 US9904865B2 (en) | 2013-05-31 | 2014-05-29 | Setting apparatus which sets a detection region for a detection process |
Country Status (2)
Country | Link |
---|---|
US (2) | US9904865B2 (en) |
JP (1) | JP6261190B2 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11245850B2 (en) | 2018-10-18 | 2022-02-08 | Canon Kabushiki Kaisha | Image processing apparatus, image processing method, and non-transitory computer-readable storage medium |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6261190B2 (en) * | 2013-05-31 | 2018-01-17 | キヤノン株式会社 | Setting device and setting method |
US20160026898A1 (en) * | 2014-07-24 | 2016-01-28 | Agt International Gmbh | Method and system for object detection with multi-scale single pass sliding window hog linear svm classifiers |
JP6650677B2 (en) | 2015-02-26 | 2020-02-19 | キヤノン株式会社 | Video processing apparatus, video processing method, and program |
JP2020109582A (en) * | 2019-01-07 | 2020-07-16 | エイアイビューライフ株式会社 | Information processing device |
JP7472482B2 (en) * | 2019-12-11 | 2024-04-23 | オムロン株式会社 | Information processing device, range determination method, and program |
CN116563521B (en) * | 2023-04-14 | 2024-04-23 | 依未科技(北京)有限公司 | Detection frame processing method and device for target detection and electronic equipment |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030112240A1 (en) * | 2001-10-10 | 2003-06-19 | Cerny Mark Evan | System and method for point pushing to render polygons in environments with changing levels of detail |
US7558403B2 (en) * | 2005-04-01 | 2009-07-07 | Canon Kabushiki Kaisha | Information processing apparatus and information processing method |
US20100182433A1 (en) * | 2007-10-17 | 2010-07-22 | Hitachi Kokusai Electric, Inc. | Object detection system |
US7903141B1 (en) * | 2005-02-15 | 2011-03-08 | Videomining Corporation | Method and system for event detection by multi-scale image invariant analysis |
US20120170902A1 (en) * | 2010-12-30 | 2012-07-05 | Pelco Inc. | Inference Engine for Video Analytics Metadata-Based Event Detection and Forensic Search |
US20130058537A1 (en) * | 2011-09-07 | 2013-03-07 | Michael Chertok | System and method for identifying a region of interest in a digital image |
US20140015782A1 (en) * | 2012-07-13 | 2014-01-16 | Samsung Electronics Co. Ltd. | Method for transmitting and receiving data between memo layer and application and electronic device using the same |
US20140355828A1 (en) * | 2013-05-31 | 2014-12-04 | Canon Kabushiki Kaisha | Setting apparatus, setting method, and storage medium |
US20160012302A1 (en) * | 2013-03-21 | 2016-01-14 | Fuji Xerox Co., Ltd. | Image processing apparatus, image processing method and non-transitory computer readable medium |
Family Cites Families (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5590248A (en) * | 1992-01-02 | 1996-12-31 | General Electric Company | Method for reducing the complexity of a polygonal mesh |
JP3574170B2 (en) * | 1994-03-17 | 2004-10-06 | 富士通株式会社 | Distributed image processing device |
JP4020982B2 (en) * | 1995-03-10 | 2007-12-12 | 松下電器産業株式会社 | Moving image processing device |
JP2746214B2 (en) * | 1995-07-25 | 1998-05-06 | 日本電気株式会社 | Polygon mesh connection device |
JP3341549B2 (en) * | 1995-11-14 | 2002-11-05 | ミノルタ株式会社 | 3D shape data processing device |
JP2001266159A (en) * | 2000-03-17 | 2001-09-28 | Toshiba Corp | Method and device for generating object domain information, and method and device for generating approximate polygon |
JP3810657B2 (en) * | 2001-07-31 | 2006-08-16 | 松下電器産業株式会社 | Moving object detection method and apparatus |
US7400344B2 (en) * | 2002-12-19 | 2008-07-15 | Hitachi Kokusai Electric Inc. | Object tracking method and object tracking apparatus |
JP4208596B2 (en) * | 2003-02-14 | 2009-01-14 | キヤノン株式会社 | Operation terminal device, camera setting method thereof, and program |
US8675059B2 (en) * | 2010-07-29 | 2014-03-18 | Careview Communications, Inc. | System and method for using a video monitoring system to prevent and manage decubitus ulcers in patients |
JP4646668B2 (en) * | 2005-03-29 | 2011-03-09 | キヤノン株式会社 | Imaging apparatus and control method thereof |
US9077882B2 (en) * | 2005-04-05 | 2015-07-07 | Honeywell International Inc. | Relevant image detection in a camera, recorder, or video streaming device |
EP2479990A3 (en) * | 2005-06-23 | 2013-01-23 | Israel Aerospace Industries Ltd. | A system and method for tracking moving objects |
CN101410875B (en) * | 2006-04-24 | 2012-07-04 | 松下电器产业株式会社 | Drawing device and drawing method |
JP4201025B2 (en) | 2006-06-30 | 2008-12-24 | ソニー株式会社 | Monitoring device, monitoring system, filter setting method, and monitoring program |
JP2008182456A (en) * | 2007-01-24 | 2008-08-07 | Sony Corp | Monitoring device, detection setting method, and monitoring camera system |
JP5495930B2 (en) * | 2010-05-10 | 2014-05-21 | キヤノン株式会社 | Image processing apparatus, method, and program |
JP5697369B2 (en) * | 2010-06-28 | 2015-04-08 | キヤノン株式会社 | Image processing apparatus, image processing system, and image processing method |
WO2012016374A1 (en) * | 2010-08-03 | 2012-02-09 | Empire Technology Development Llc | Method for identifying objects in video |
JP5653119B2 (en) * | 2010-08-06 | 2015-01-14 | キヤノン株式会社 | SETTING DEVICE, SETTING METHOD, AND PROGRAM |
KR101355974B1 (en) * | 2010-08-24 | 2014-01-29 | 한국전자통신연구원 | Method and devices for tracking multiple object |
JP5938631B2 (en) * | 2011-12-19 | 2016-06-22 | パナソニックIpマネジメント株式会社 | Object detection apparatus and object detection method |
US9171213B2 (en) * | 2013-03-15 | 2015-10-27 | Xerox Corporation | Two-dimensional and three-dimensional sliding window-based methods and systems for detecting vehicles |
-
2013
- 2013-05-31 JP JP2013115668A patent/JP6261190B2/en active Active
-
2014
- 2014-05-29 US US14/290,688 patent/US9904865B2/en active Active
-
2018
- 2018-01-16 US US15/872,835 patent/US20180144206A1/en not_active Abandoned
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030112240A1 (en) * | 2001-10-10 | 2003-06-19 | Cerny Mark Evan | System and method for point pushing to render polygons in environments with changing levels of detail |
US7903141B1 (en) * | 2005-02-15 | 2011-03-08 | Videomining Corporation | Method and system for event detection by multi-scale image invariant analysis |
US7558403B2 (en) * | 2005-04-01 | 2009-07-07 | Canon Kabushiki Kaisha | Information processing apparatus and information processing method |
US20100182433A1 (en) * | 2007-10-17 | 2010-07-22 | Hitachi Kokusai Electric, Inc. | Object detection system |
US20120170902A1 (en) * | 2010-12-30 | 2012-07-05 | Pelco Inc. | Inference Engine for Video Analytics Metadata-Based Event Detection and Forensic Search |
US20130058537A1 (en) * | 2011-09-07 | 2013-03-07 | Michael Chertok | System and method for identifying a region of interest in a digital image |
US20140015782A1 (en) * | 2012-07-13 | 2014-01-16 | Samsung Electronics Co. Ltd. | Method for transmitting and receiving data between memo layer and application and electronic device using the same |
US20160012302A1 (en) * | 2013-03-21 | 2016-01-14 | Fuji Xerox Co., Ltd. | Image processing apparatus, image processing method and non-transitory computer readable medium |
US20140355828A1 (en) * | 2013-05-31 | 2014-12-04 | Canon Kabushiki Kaisha | Setting apparatus, setting method, and storage medium |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11245850B2 (en) | 2018-10-18 | 2022-02-08 | Canon Kabushiki Kaisha | Image processing apparatus, image processing method, and non-transitory computer-readable storage medium |
Also Published As
Publication number | Publication date |
---|---|
JP6261190B2 (en) | 2018-01-17 |
US9904865B2 (en) | 2018-02-27 |
JP2014236312A (en) | 2014-12-15 |
US20140355828A1 (en) | 2014-12-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20180144206A1 (en) | Setting apparatus, setting method, and storage medium | |
JP5398341B2 (en) | Object recognition apparatus and object recognition method | |
KR101687530B1 (en) | Control method in image capture system, control apparatus and a computer-readable storage medium | |
JP6525545B2 (en) | INFORMATION PROCESSING APPARATUS, INFORMATION PROCESSING METHOD, AND COMPUTER PROGRAM | |
US8907989B2 (en) | Display control apparatus and display control method | |
EP3573022A1 (en) | Method for tracking pedestrian and electronic device | |
EP3442219B1 (en) | Information processing apparatus, information processing method, and storage medium | |
JP5693094B2 (en) | Image processing apparatus, image processing method, and computer program | |
US11134226B2 (en) | Surveillance system, surveillance method, and program | |
US20150334299A1 (en) | Monitoring system | |
JP5460793B2 (en) | Display device, display method, television receiver, and display control device | |
EP3477544A1 (en) | Information processing apparatus, information processing method, imaging apparatus, network camera system, and program | |
US11263759B2 (en) | Image processing apparatus, image processing method, and storage medium | |
JP2008009938A (en) | Moving image data processor, moving image data processing method, moving image data processing program and storage medium recording the program | |
JP5769468B2 (en) | Object detection system and object detection method | |
CN115937730A (en) | Estimation device, estimation method, and storage medium | |
US20190080179A1 (en) | Monitoring system and terminal device | |
EP3496390B1 (en) | Information processing device, information processing method, and storage medium | |
JP7140583B2 (en) | Image processing device, image processing method and program | |
JP7004786B1 (en) | Detection device and detection method | |
CN117499778B (en) | Character image shooting interface interaction method and device and electronic equipment | |
WO2023276050A1 (en) | Image processing system and imaging device | |
JP2023013064A (en) | Information processing apparatus, information processing method, and program | |
JP4944720B2 (en) | Image processing apparatus and method, and storage medium | |
JP2019128791A (en) | Image processing device and control method thereof |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |