KR101756916B1 - Device and method for action recognition - Google Patents
Device and method for action recognition Download PDFInfo
- Publication number
- KR101756916B1 KR101756916B1 KR1020160013087A KR20160013087A KR101756916B1 KR 101756916 B1 KR101756916 B1 KR 101756916B1 KR 1020160013087 A KR1020160013087 A KR 1020160013087A KR 20160013087 A KR20160013087 A KR 20160013087A KR 101756916 B1 KR101756916 B1 KR 101756916B1
- Authority
- KR
- South Korea
- Prior art keywords
- image
- target
- information
- visible light
- recognizing
- Prior art date
Links
Images
Classifications
-
- G06K9/00335—
-
- G06K9/00342—
-
- G06K9/00348—
-
- G06K9/20—
-
- G06K9/2054—
Landscapes
- Image Analysis (AREA)
Abstract
BACKGROUND OF THE INVENTION 1. Field of the Invention The present invention relates to an apparatus and method for recognizing an action, and more particularly to an apparatus and method for detecting an object using an object detection and tracking information. The present invention is capable of recognizing the behavior of the object stably in various lighting and weather environments and real-time behavior recognition because it does not use complicated features.
Description
BACKGROUND OF THE INVENTION 1. Field of the Invention The present invention relates to an apparatus and method for recognizing an action, and more particularly to an apparatus and method for detecting an object using an object detection and tracking information.
Behavior recognition technology, which automatically understands human behavior in video, has been used for intelligent video surveillance and human - object interaction, and research is proceeding with much interest. Most existing methods of recognizing human behavior in video use motion information because motion information provides important information for classifying the behavior. However, the process of extracting motion information requires a lot of computation and it is difficult to recognize human actions in real time because of the large number of dimensions of extracted features.
The background art of the present invention is disclosed in Korean Patent Publication No. 2008-0051013 (published on Jun. 10, 2008).
The present invention provides an apparatus and method for recognizing an action of an object using detection and tracking information of an object in a visible ray and a thermal image.
The objects of the present invention are not limited to the above-mentioned objects, and other objects not mentioned can be clearly understood from the following description.
According to an aspect of the present invention, an action recognition apparatus is provided.
A behavior recognition apparatus according to an embodiment of the present invention includes a visible light target region detection unit for generating a visible light background image and detecting a visible light target candidate region using a difference image of a visible light background image and an input visible light ray image, A thermal image target region detection unit for generating a thermal image background image and detecting a thermal image target candidate region by using a difference image between the generated thermal image background image and the input thermal image, A target region combining unit for combining regions to generate combined target region information, a target tracking unit for generating target tracking information by tracking the target based on the combining target region information, and a target recognition unit for recognizing the target target using the target tracking information And an action recognition unit.
According to another aspect of the present invention, there is provided a method for recognizing an action and a computer program for executing the method.
A behavior recognition method and a computer program for implementing the method of the present invention generate a visible light background image and detect a visible light target candidate area using a difference image of a visible light background image and an input visible light ray image A step of generating a thermal image background image, detecting a thermal image candidate region by using a difference image between the generated thermal image background image and the input thermal image, detecting the detected visible light region and the thermal image region Generating combining target area information, generating target tracking information by tracking the target based on the combining target area information, and recognizing an action of the target using the target tracking information.
The present invention is capable of recognizing the behavior of the object stably in various lighting and weather environments and real-time behavior recognition because it does not use complicated features.
1 is a view for explaining an action recognition system according to an embodiment of the present invention;
2 is a view for explaining an action recognition apparatus according to an embodiment of the present invention.
3 is a view for explaining an object behavior recognizing unit according to an embodiment of the present invention;
4 is a view for explaining an action recognition method according to an embodiment of the present invention;
5 is a view for explaining a single action recognition method among the action recognition methods according to an embodiment of the present invention.
6 to 8 are diagrams for explaining a hand waving, a kicking, and a fist striking method in an action recognition method according to an embodiment of the present invention.
9 is a diagram for explaining a mutual action recognition method between a plurality of objects according to an embodiment of the present invention;
Hereinafter, embodiments of the present invention will be described in detail with reference to the accompanying drawings so that those skilled in the art can easily carry out the present invention. The present invention may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein. Also, when a part is referred to as "including " an element, it is to be understood that it may include other elements as well, without departing from the other elements unless specifically stated otherwise.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS Reference will now be made in detail to embodiments of the present invention, examples of which are illustrated in the accompanying drawings.
1 is a view for explaining an action recognition system according to an embodiment of the present invention.
Referring to FIG. 1, an action recognition system includes a visible light
The visible light
The thermal
The
2 is a view for explaining an action recognition apparatus according to an embodiment of the present invention.
2, the
The visible light target
The thermal image target
The target
The
The target
The target mutual
FIG. 3 is a diagram for explaining an object behavior recognizing unit according to an embodiment of the present invention.
3, the target
The horizontal and vertical
The horizontal
The center point position change
The center point
The aspect
The hand and foot
The hand and foot
The hand and foot
4 is a diagram for explaining an action recognition method according to an embodiment of the present invention.
Referring to FIG. 4, in step S410, the
In step S420, the
In step S430, the
Steps S410 to S430 may be performed in units of a specific image interval or in a case where target tracking fails in the next step S440.
In step S440, the
In step S450, the
5 is a diagram for explaining a single action recognition method among the action recognition methods according to an embodiment of the present invention.
Referring to FIG. 5, in step S510, the
In step S520, the
In step S530, the
In step S540, the
In step S550, the
In step S560, the
In step S570, the
In step S580, the
6 to 8 are views for explaining a hand waving, a kicking, and a fist striking method among the methods of recognizing an action according to an embodiment of the present invention.
Referring to FIG. 6, in step S610, the
Referring to FIG. 7, the
The
When the ratio of the first distance L1 to the second distance L2 is larger than the predetermined distance ratio threshold, the
In step S620, when the foot position is found in step S610, the
In step S630, the
In step S640, the
9 is a diagram for explaining a mutual action recognition method between a plurality of objects according to an embodiment of the present invention.
In step S910, when the
In step S920, the
The behavior recognition method according to various embodiments of the present invention can be implemented in the form of a program command that can be executed through various means such as a server. Further, the program for executing the behavior recognition method according to the present invention may be installed in a computer means and recorded in a computer-readable medium. The computer readable medium may include program instructions, data files, data structures, and the like, alone or in combination. Program instructions to be recorded on a computer-readable medium may be those specially designed and constructed for the present invention or may be available to those skilled in the computer software arts. Examples of computer-readable media include magnetic media such as hard disks, floppy disks and magnetic tape; optical media such as CD-ROMs and DVDs; magnetic media such as floppy disks; Includes hardware devices specifically configured to store and execute program instructions such as magneto-optical media and ROM, RAM, flash memory, and the like.
The embodiments of the present invention have been described above. It will be understood by those skilled in the art that the foregoing description of the present invention has been presented for illustrative purposes and that those skilled in the art will readily understand that various changes in form and details may be made therein without departing from the spirit and scope of the invention. It will be possible. It is therefore to be understood that the above-described embodiments are illustrative in all aspects and not restrictive. The scope of the present invention is defined by the appended claims rather than the detailed description, and all changes or modifications derived from the meaning and scope of the claims and their equivalents should be construed as being included in the scope of the present invention.
100: visible light image input unit
200: thermal image input unit
300:
Claims (20)
A visible light target area detecting unit for generating a visible light background image and detecting a visible light target candidate area using a difference image between the visible light background image and the input visible light image;
A thermal image target area detecting unit for generating a thermal image background image and detecting a thermal image target candidate area by using a difference image between the generated thermal image background image and the input thermal image;
An object area combining unit for combining the detected visible light target area and the thermal image target area to generate the combined object area information;
A target tracking unit that generates target tracking information by tracking the target based on the combining target area information;
An object behavior recognition unit for recognizing an object behavior using object tracking information; And
And a target mutual action recognition unit for determining presence or absence of an mutual action among a plurality of objects when there is a target action recognition result for a plurality of objects,
Wherein the target mutual action recognizing unit judges that an action of the target has a mutual action when there is a walk or a run among a plurality of target action recognition results and if the distance change information between the plurality of targets is larger than a predetermined distance change threshold, The device is judged to be farther away, and the device is judged to be approached in a small case.
A visible light target area detecting unit for generating a visible light background image and detecting a visible light target candidate area using a difference image between the visible light background image and the input visible light image;
A thermal image target area detecting unit for generating a thermal image background image and detecting a thermal image target candidate area by using a difference image between the generated thermal image background image and the input thermal image;
An object area combining unit for combining the detected visible light target area and the thermal image target area to generate the combined object area information;
A target tracking unit that generates target tracking information by tracking the target based on the combining target area information; And
And a target behavior recognition unit for recognizing an action of the target using the target tracking information,
The target behavior recognition unit
A horizontal and vertical information generator for generating horizontal length information and vertical length information of an input target area;
A horizontal size change comparator for recognizing any one of an action of hand waving, kicking, punching, punching, walking, and walking by using information of cumulative horizontal size change of N input regions (where N is a natural number) And
And a center point position change comparing unit for recognizing the run based on the center point velocity in the X axis direction among the N pieces of the object area information inputted or determining cumulative change information of the center point Y axis coordinate position as walking or running Action recognition device.
The target behavior recognition unit
And an aspect ratio comparing unit for recognizing the inputted N pieces of the target area information as one of a run, a clerk and a sitting by using an average value of the ratio of the length to the length of the object.
The target behavior recognition unit
The hand or foot position is determined in the input image, and the kicking direction is recognized using the X-axis coordinate position of the foot in some images of the inputted N target images or the X-axis coordinate position change information of the hand is calculated, And a hand / foot position comparing unit for recognizing the hand / foot position comparing unit.
Generating a visible light background image and detecting a visible light target candidate region using a difference image of the visible light background image and the input visible light image;
Generating a thermal background image, detecting a thermal image candidate region using a difference image between the generated thermal background image and the input thermal image;
Combining the detected visible light target area and the thermal image target area to generate combined object area information;
Generating object tracking information by tracking an object based on the combining object area information; And
And recognizing an action of the object using the object tracking information,
The step of recognizing the behavior of the object using the object tracking information
Accumulating at least N (where N is a natural number) tracked object image and object area information;
Comparing the horizontal size change information among the target area information with a predetermined horizontal size change threshold; And
Recognizing the action of the object as a jump if the horizontal size change information is smaller than the horizontal size change threshold and the center point X axis coordinate position change information of the target area information is larger than a predefined center point X axis coordinate position change threshold Including an action recognition method.
The step of generating the visible light background image and detecting the visible light target candidate area using the difference image of the visible light background image and the input visible light ray image
A method of recognizing a visible light target area by using a difference image between input images or using a difference image between a background image and an input image, or by performing background modeling of GMM (Gaussian Mixture Models) or MOG (Model of Gaussian).
The step of combining the detected visible light target area and the thermal image target area to generate the combined object area information
A method for recognizing an action that combines a visible light target area and a thermal image target area by selectively using an object area detected in each image or by changing a weight, by checking the input time of the input image or the quality of each image.
If the center point X axis coordinate position change information is not larger than the predefined center point X axis coordinate position change threshold and the average value of the horizontal length to vertical length ratio in the object area information is smaller than the predefined first aspect ratio threshold, And recognizing that the user is walking.
When the horizontal size change information among the inputted N pieces of the object area information is not larger than the predetermined horizontal size change threshold and the change information of the center point Y axis coordinate position is larger than the center point Y axis coordinate position change threshold, And recognizing the action as one of the actions.
And recognizing that the target behavior is a walk when the velocity of the center point Y of the N pieces of the target area information is not greater than the center point Y velocity change threshold value.
When the change information of the center point Y axis coordinate position among the inputted N pieces of target area information is not larger than the center point Y axis coordinate position change threshold and the average value of the ratio of width to height ratio is smaller than the predefined second aspect ratio threshold, A step of recognizing the action as a clerk.
Generating a visible light background image and detecting a visible light target candidate region using a difference image of the visible light background image and the input visible light image;
Generating a thermal background image, detecting a thermal image candidate region using a difference image between the generated thermal background image and the input thermal image;
Combining the detected visible light target area and the thermal image target area to generate combined object area information;
Generating object tracking information by tracking an object based on the combining object area information; And
And recognizing an action of the object using the object tracking information,
The step of recognizing the behavior of the object using the object tracking information
Determining a position of the hand or foot in the target image information; And
And recognizing an action of the subject as a kick-out when the X-axis coordinate position of the foot in the tracked target image is larger than the predefined foot X-axis coordinate position threshold.
Axis coordinate position of the foot is not larger than the predefined foot X-axis coordinate position threshold, and the X-axis coordinate and Y-axis coordinate position of the hand change in the images are changed. And recognizing an action of the target as a hand shake when the threshold value is larger than the threshold value.
Axis coordinate position and a Y-axis coordinate position change information of the hand are not larger than a predetermined hand X-axis coordinate position variation threshold and a hand Y-axis coordinate position variation threshold, and cumulative change information of an X-axis coordinate position of the hand is not greater than a predetermined second hand And recognizing an action of the object as a punching if it is larger than the X axis coordinate position change threshold.
Generating a visible light background image and detecting a visible light target candidate region using a difference image of the visible light background image and the input visible light image;
Generating a thermal background image, detecting a thermal image candidate region using a difference image between the generated thermal background image and the input thermal image;
Combining the detected visible light target area and the thermal image target area to generate combined object area information;
Generating object tracking information by tracking an object based on the combining object area information; And
And recognizing an action of the object using the object tracking information,
The step of recognizing the behavior of the object using the object tracking information
If there is a single target action recognition result for two objects and the distance change information between two objects of the input N area information is larger than the predetermined distance change threshold value, Recognizing an action; And
And recognizing an action of an object when the distance change information between the two objects among the input N pieces of object area information is smaller than a predetermined distance change threshold.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020160013087A KR101756916B1 (en) | 2016-02-02 | 2016-02-02 | Device and method for action recognition |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020160013087A KR101756916B1 (en) | 2016-02-02 | 2016-02-02 | Device and method for action recognition |
Publications (1)
Publication Number | Publication Date |
---|---|
KR101756916B1 true KR101756916B1 (en) | 2017-07-12 |
Family
ID=59352941
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
KR1020160013087A KR101756916B1 (en) | 2016-02-02 | 2016-02-02 | Device and method for action recognition |
Country Status (1)
Country | Link |
---|---|
KR (1) | KR101756916B1 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR102127946B1 (en) * | 2019-02-21 | 2020-06-30 | 동국대학교 산학협력단 | Device and method for person re-identification |
CN112070016A (en) * | 2020-09-08 | 2020-12-11 | 安徽兰臣信息科技有限公司 | Detection method for identifying child behavior and action |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101583773B1 (en) * | 2014-10-29 | 2016-01-20 | 동국대학교 산학협력단 | Apparatus and method for object detection |
-
2016
- 2016-02-02 KR KR1020160013087A patent/KR101756916B1/en active IP Right Grant
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101583773B1 (en) * | 2014-10-29 | 2016-01-20 | 동국대학교 산학협력단 | Apparatus and method for object detection |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR102127946B1 (en) * | 2019-02-21 | 2020-06-30 | 동국대학교 산학협력단 | Device and method for person re-identification |
CN112070016A (en) * | 2020-09-08 | 2020-12-11 | 安徽兰臣信息科技有限公司 | Detection method for identifying child behavior and action |
CN112070016B (en) * | 2020-09-08 | 2023-12-26 | 浙江铂视科技有限公司 | Detection method for identifying child behavior and action |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP7272417B2 (en) | Object tracking system, object tracking device, and object tracking method | |
US11789545B2 (en) | Information processing device and method, program and recording medium for identifying a gesture of a person from captured image data | |
JP6625220B2 (en) | Method and system for detecting the action of an object in a scene | |
EP1330128B1 (en) | Automatic detection and tracking of multiple individuals' faces using multiple cues | |
JP6419830B2 (en) | System, method and apparatus for image retrieval | |
US7613325B2 (en) | Human detection device and human detection method | |
US9600760B2 (en) | System and method for utilizing motion fields to predict evolution in dynamic scenes | |
JP5077164B2 (en) | Tracking device and tracking method | |
JP5754990B2 (en) | Information processing apparatus, information processing method, and program | |
JP5634517B2 (en) | Video image monitoring | |
JP2009075868A (en) | Apparatus, method, and program for detecting object from image | |
JP2013501289A (en) | Image analysis method | |
JP6649231B2 (en) | Search device, search method and program | |
Manikandan et al. | Human object detection and tracking using background subtraction for sports applications | |
KR101756916B1 (en) | Device and method for action recognition | |
Okamoto et al. | Summarization of egocentric moving videos for generating walking route guidance | |
US20210235012A1 (en) | Electronic device and method for controlling electronic device | |
Zhang et al. | Robust multi-view multi-camera face detection inside smart rooms using spatio-temporal dynamic programming | |
JP2019096062A (en) | Object tracking device, object tracking method, and object tracking program | |
Farazi et al. | Real-time visual tracking and identification for a team of homogeneous humanoid robots | |
KR101771655B1 (en) | Device and method for action recognition | |
JP6040825B2 (en) | Object tracking program, object tracking method, and object tracking apparatus | |
JP5241687B2 (en) | Object detection apparatus and object detection program | |
Zhang et al. | Long-term face tracking in the wild using deep learning | |
JP7240940B2 (en) | Object image extraction device, method, and software program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
E701 | Decision to grant or registration of patent right | ||
GRNT | Written decision to grant |