US20120136890A1 - Behavior pattern recognition method, system and computer application program thereof - Google Patents
Behavior pattern recognition method, system and computer application program thereof Download PDFInfo
- Publication number
- US20120136890A1 US20120136890A1 US12/969,254 US96925410A US2012136890A1 US 20120136890 A1 US20120136890 A1 US 20120136890A1 US 96925410 A US96925410 A US 96925410A US 2012136890 A1 US2012136890 A1 US 2012136890A1
- Authority
- US
- United States
- Prior art keywords
- behavior
- detecting unit
- pattern recognition
- information
- feature information
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
- G06V40/173—Classification, e.g. identification face re-identification, e.g. recognising unknown faces across different face tracks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/94—Hardware or software architectures specially adapted for image or video understanding
- G06V10/95—Hardware or software architectures specially adapted for image or video understanding structured as a network, e.g. client-server architectures
Definitions
- the present invention relates to a behavior pattern recognition method, system and a computer application program thereof, and more particularly to a behavior pattern recognition method and system using sensor information and behavior feature information of relevant persons to deduce a user's behavior and a computer application program thereof.
- the manual management and control is one of the most important parts. Since each person has a different identity, level, authority and nature of work, the region that can be reached by the person in the office or business area differs.
- multiple monitors are disposed at different spots in a building to send back the captured image frame to a control center at any time, and display devices disposed in the control center switch the frames at a specific time or play the frames of several monitors at the same time for the management staff to observe if there are persons without authorization enter the spots under surveillance.
- this surveillance method does not need to arrange a guard at the spot under control, this surveillance method still needs manpower dedicated to observe, and in practice, oversights inevitably occur. Therefore, there is a need for a means which is more effective and capable of automatically informing the abnormal situations properly.
- the present invention is directed to a behavior pattern recognition method, system and a computer application program thereof, thereby providing the convenience for a person to manage and control.
- the present invention provides a behavior pattern recognition method, which is applicable to an electronic device having a storage unit for storing multiple sets of behavior record information, and the method comprises the following steps. Firstly, a first detecting unit acquires first behavior feature information and a collaboration network module acquires at least one second detecting unit having coherence with the first detecting unit. Then, at least one second detecting unit acquires at least one second behavior feature information and a processing unit compares at least one second behavior feature information and the behavior record information to generate at least one comparison result. Finally, a behavior definition represented by the first behavior feature information is determined according to the comparison result.
- a plurality of detecting units acquires a plurality of sample behavior information and analyzes the sample behavior information to generate behavior record information.
- the method further comprises comparing coherence of the detecting units and the first detecting unit; and screening out the detecting units having the coherence with the first detecting unit exceeding a preset value to serve as at least one second detecting unit according to the coherence.
- the method further comprises disposing a behavior analysis module to analyze sample behavior information to generate a plurality of behavior record information.
- the first detecting unit and the second detecting unit are non-invasive detectors, and each comprise an electrical detector, a sound sensor, an infrared sensor, a video/audio recorder, an electromagnetic sensor, and a mobile phone having detecting and sensing functions.
- the method further comprises the following steps. First behavior feature information and at least one second behavior feature information are acquired respectively by the first detecting unit and the at least one second detecting unit according to a preset time interval.
- the present invention provides a behavior pattern recognition system, which comprises a storage unit, a first detecting unit, at least one second detecting unit, and a processing unit.
- the storage unit stores multiple sets of behavior record information and the first detecting unit acquires first behavior feature information.
- At least one second detecting unit has coherence with the first detecting unit and acquires at least one second behavior feature information.
- the processing unit compares at least one second behavior feature information and the behavior record information to generate at least one comparison result, and determines a behavior definition represented by the first behavior feature information according to the comparison result.
- the system further comprises a plurality of detecting units to acquire a plurality of sample behavior information.
- the system further comprises a behavior analysis module for analyzing the sample behavior information to generate a plurality of behavior record information.
- system further comprises a collaboration network module for acquiring at least one second detecting unit having coherence with the first detecting unit.
- the first detecting unit and the second detecting unit are non-invasive detectors.
- the non-invasive detector comprises an electrical detector, a sound sensor, an infrared sensor, a video/audio recorder, an electromagnetic sensor, and a mobile phone having detecting and sensing functions.
- a time interval is preset in the first detecting unit and the first behavior feature information is acquired according to the time interval.
- a time interval is preset in the at least one second detecting unit and at least one second behavior feature information is acquired according to the time interval.
- the coherence of the first detecting unit and at least one second detecting unit includes position information.
- the present invention further provides a computer program product, which is provided for an electronic equipment to execute the above behavior pattern recognition method, the process flow is as described above, and the details will not be repeated herein again.
- the present invention adopts a group interaction structure and utilizes a feature capturing method and a machine learning method in cooperation with the group interaction model to deduce the user behavior, thereby acquiring a more accurate user behavior, such that the overall recognition rate is greatly improved.
- FIG. 1 is a flow chart of steps of a behavior pattern recognition method according to the present invention
- FIG. 2 is a flow chart of steps of behavior pattern recognition method and preparation works thereof according to the present invention.
- FIG. 3 is a schematic block diagram of elements of a behavior pattern recognition system according to the present invention.
- FIG. 4 is a schematic view of a behavior pattern recognition system according to another embodiment of the present invention.
- FIG. 1 is a flow chart of steps of a behavior pattern recognition method according to the present invention.
- the present invention provides a behavior pattern recognition method, which is applicable to an electronic device having a storage unit for storing multiple sets of behavior record information.
- the method includes the following step.
- Step S 110 firstly, a first detecting unit acquires first behavior feature information.
- a plurality of detecting units must be used to acquire a plurality of sample behavior information and analyze the sample behavior information to generate behavior record information. Coherence of the detecting unit and the first detecting unit is compared, the detecting units having the coherence with the first detecting unit exceeding a preset value are screened out to serve as the at least one second detecting unit according to the coherence.
- the method further includes disposing a behavior analysis module to analyze the sample behavior information to generate a plurality of behavior record information.
- Step S 120 a collaboration network module acquires at least one second detecting unit having coherence with the first detecting unit.
- the first detecting unit and the second detecting unit are non-invasive detectors
- the non-invasive detector includes an electrical detector, a sound sensor, an infrared sensor, a video/audio recorder, an electromagnetic sensor, and a mobile phone having detecting and sensing functions.
- Step S 130 the at least one second detecting unit acquires at least one second behavior feature information.
- the first behavior feature information and the at least one second behavior feature information are acquired respectively by the first detecting unit and the at least one second detecting unit according to a preset time interval.
- Step S 140 a processing unit compares the at least one second behavior feature information and the behavior record information to generate at least one comparison result.
- Step S 150 a behavior definition represented by the first behavior feature information is determined according to the comparison result.
- FIG. 2 is a flow chart of steps of a behavior pattern recognition method according to the present invention and preparation works thereof. The method includes the following steps.
- Step S 210 a first detecting unit acquires first behavior feature information.
- Step S 220 a plurality of detecting units acquires a plurality of sample behavior information.
- Step S 230 the sample behavior information is analyzed to generate behavior record information.
- Step S 240 coherence of the detecting units and the first detecting unit is compared.
- Step S 250 the detecting units having coherence with the first detecting unit exceeding a preset value are screened out to serve as the at least one second detecting unit according to the coherence.
- Step S 260 a collaboration network module acquires at least one second detecting unit having coherence with the first detecting unit.
- Step S 270 at least one second detecting unit acquires at least one second behavior feature information.
- Step S 280 a processing unit compares the at least one second behavior feature information and the behavior record information to generate at least one comparison result.
- Step S 290 a behavior definition represented by the first behavior feature information is determined according to the comparison result.
- FIG. 3 is a schematic block diagram of elements of a behavior pattern recognition system according to the present invention.
- a behavior pattern recognition system of the present invention which includes a storage unit 310 , a first detecting unit 320 , at least one second detecting unit 330 , and a processing unit 340 .
- the storage unit 310 stores multiple sets of behavior record information 311
- the first detecting unit 320 acquire first behavior feature information 321 .
- the at least one second detecting unit 330 has coherence with the first detecting unit 320 , and acquires at least one second behavior feature information 331 .
- the processing unit 340 compares the at least one second behavior feature information 331 with the behavior record information 311 to generate at least one comparison result, and a behavior definition represented by the first behavior feature information 321 is determined according to the comparison result.
- the system further includes a plurality of detecting units to acquire a plurality of sample behavior information.
- system further includes a behavior analysis module to analyze the sample behavior information to generate a plurality of behavior record information.
- system further includes a collaboration network module to acquire at least one second detecting unit having coherence with the first detecting unit.
- the first detecting unit and the second detecting unit are non-invasive detectors.
- the non-invasive detector includes an electrical detector, a sound sensor, an infrared sensor, a video/audio recorder, an electromagnetic sensor, and a mobile phone having detecting and sensing functions.
- a time interval is preset in the first detecting unit and the first behavior feature information is acquired according to the time interval.
- a time interval is preset in the at least one second detecting unit and at least one second behavior feature information is acquired according to the time interval.
- the coherence of the first detecting unit and at least one second detecting unit includes position information, for example, at the position of an office or a classroom.
- FIG. 4 is a schematic view of a behavior pattern recognition system according to another embodiment of the present invention.
- a first detecting unit 410 detects the behavior feature information of the user, and firstly the behavior feature information of the first detecting unit 410 , a second detecting unit 420 and a third detecting unit 430 are input to a feature capturing device 440 .
- the second detecting unit 420 is preset to be other users coherent to the user.
- the feature capturing device 440 captures feature by observation, and different sensors may be applicable to different feature capturing methods, but the binary signals like PIR or Reed Switch do not need this method.
- the behavior feature information of the second detecting unit 420 and the third detecting unit 430 passes through the feature capturing device 440 and is input in a sorting module 450 and a collaboration network module 460 .
- the sorting module 450 mainly identifies status of other users, such as at work, leave the seat, and off work.
- Group information 470 is added in the collaboration network module 460 to screen out the behavior feature information of the second detecting unit 420 and the third detecting unit 430 . Then, the collaboration network inputs the user behavior feature information of the second detecting unit 420 related to the user of the first detecting unit 410 to a behavior recognition module 480 , thereby determining a behavior definition represented by the user behavior feature information of the first detecting unit 410 .
- the user status values of the first detecting unit 410 and the second detecting unit 420 are taken as the input values, and the output values are the user behavior, such as, at the seat (operating the computer), at the seat (other behaviors), leaving the seat (at the meeting inside), leaving the seat (at the meeting outside), leaving the seat (others), or off work.
- the present invention adopts a group interaction structure and utilizes a feature capturing method and a machine learning method in cooperation with the group interaction model to deduce the user behavior, and thus the present invention may be schemed in the following fields.
- the features of the present invention may be used to assist the enterprises or consultancy companies to clearly know the social network status and the working status of the staff, thereby providing a solution for the enterprises, improving the working efficiency, the innovation of the enterprises and the working satisfaction. Different from collecting the poll, the working status of the staff are more clearly, so suggestion and assistance may be made to those having a low working efficiency.
- the family members of a senior may know the living status of the senior by using this system, and this system may record detailed living behaviors and can also be used to identify the health conditions of the senior.
- the parents may know the child's behaviors and the activities of the child by using this system.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Human Computer Interaction (AREA)
- Software Systems (AREA)
- Telephonic Communication Services (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
A behavior pattern recognition method, system and a computer application program thereof are presented. The method is applicable to an electronic device which has a storage unit for storing multiple sets of behavior record information, and the method includes the following step. Firstly, a first detecting unit acquires first behavior feature information, a collaboration network module acquires at least one second detecting unit having coherence with the first detecting unit. Then, the at least one second detecting unit acquires at least one second behavior feature information, and a processing unit compares the at least one second behavior feature information and the behavior record information to generate at least one comparison result. Finally, a behavior definition represented by the first behavior feature information is determined according to the comparison result.
Description
- This application claims the benefit of Taiwan Patent Application No. 099141005, filed on Nov. 26, 2010, which is hereby incorporated by reference for all purposes as if fully set forth herein.
- 1. Field of Invention
- The present invention relates to a behavior pattern recognition method, system and a computer application program thereof, and more particularly to a behavior pattern recognition method and system using sensor information and behavior feature information of relevant persons to deduce a user's behavior and a computer application program thereof.
- 2. Related Art
- Due to the quick progress of technology, the processing of events becomes more complicated. In accord with the changes, the management of human, event, environmental, and object resources gradually changes from the manual supervision inspection into the automated management and control.
- In the office or business areas, the manual management and control is one of the most important parts. Since each person has a different identity, level, authority and nature of work, the region that can be reached by the person in the office or business area differs. In a common manner, multiple monitors are disposed at different spots in a building to send back the captured image frame to a control center at any time, and display devices disposed in the control center switch the frames at a specific time or play the frames of several monitors at the same time for the management staff to observe if there are persons without authorization enter the spots under surveillance.
- However, although this surveillance method does not need to arrange a guard at the spot under control, this surveillance method still needs manpower dedicated to observe, and in practice, oversights inevitably occur. Therefore, there is a need for a means which is more effective and capable of automatically informing the abnormal situations properly.
- The present invention is directed to a behavior pattern recognition method, system and a computer application program thereof, thereby providing the convenience for a person to manage and control.
- The present invention provides a behavior pattern recognition method, which is applicable to an electronic device having a storage unit for storing multiple sets of behavior record information, and the method comprises the following steps. Firstly, a first detecting unit acquires first behavior feature information and a collaboration network module acquires at least one second detecting unit having coherence with the first detecting unit. Then, at least one second detecting unit acquires at least one second behavior feature information and a processing unit compares at least one second behavior feature information and the behavior record information to generate at least one comparison result. Finally, a behavior definition represented by the first behavior feature information is determined according to the comparison result.
- In an embodiment of the present invention, before the step of acquiring at least one second detecting unit having coherence with the first detecting unit by the collaboration network module, a plurality of detecting units acquires a plurality of sample behavior information and analyzes the sample behavior information to generate behavior record information.
- In an embodiment of the present invention, the method further comprises comparing coherence of the detecting units and the first detecting unit; and screening out the detecting units having the coherence with the first detecting unit exceeding a preset value to serve as at least one second detecting unit according to the coherence.
- In an embodiment of the present invention, the method further comprises disposing a behavior analysis module to analyze sample behavior information to generate a plurality of behavior record information.
- In an embodiment of the present invention, the first detecting unit and the second detecting unit are non-invasive detectors, and each comprise an electrical detector, a sound sensor, an infrared sensor, a video/audio recorder, an electromagnetic sensor, and a mobile phone having detecting and sensing functions.
- In an embodiment of the present invention, the method further comprises the following steps. First behavior feature information and at least one second behavior feature information are acquired respectively by the first detecting unit and the at least one second detecting unit according to a preset time interval.
- The present invention provides a behavior pattern recognition system, which comprises a storage unit, a first detecting unit, at least one second detecting unit, and a processing unit. The storage unit stores multiple sets of behavior record information and the first detecting unit acquires first behavior feature information. At least one second detecting unit has coherence with the first detecting unit and acquires at least one second behavior feature information. The processing unit compares at least one second behavior feature information and the behavior record information to generate at least one comparison result, and determines a behavior definition represented by the first behavior feature information according to the comparison result.
- In an embodiment of the present invention, the system further comprises a plurality of detecting units to acquire a plurality of sample behavior information.
- In an embodiment of the present invention, the system further comprises a behavior analysis module for analyzing the sample behavior information to generate a plurality of behavior record information.
- In an embodiment of the present invention, the system further comprises a collaboration network module for acquiring at least one second detecting unit having coherence with the first detecting unit.
- In an embodiment of the present invention, the first detecting unit and the second detecting unit are non-invasive detectors. The non-invasive detector comprises an electrical detector, a sound sensor, an infrared sensor, a video/audio recorder, an electromagnetic sensor, and a mobile phone having detecting and sensing functions.
- In an embodiment of the present invention, a time interval is preset in the first detecting unit and the first behavior feature information is acquired according to the time interval.
- In an embodiment of the present invention, a time interval is preset in the at least one second detecting unit and at least one second behavior feature information is acquired according to the time interval.
- In an embodiment of the present invention, the coherence of the first detecting unit and at least one second detecting unit includes position information.
- The present invention further provides a computer program product, which is provided for an electronic equipment to execute the above behavior pattern recognition method, the process flow is as described above, and the details will not be repeated herein again.
- The present invention adopts a group interaction structure and utilizes a feature capturing method and a machine learning method in cooperation with the group interaction model to deduce the user behavior, thereby acquiring a more accurate user behavior, such that the overall recognition rate is greatly improved.
- The present invention will become more fully understood from the detailed description given herein below for illustration only, and thus are not limitative of the present invention, and wherein:
-
FIG. 1 is a flow chart of steps of a behavior pattern recognition method according to the present invention; -
FIG. 2 is a flow chart of steps of behavior pattern recognition method and preparation works thereof according to the present invention; -
FIG. 3 is a schematic block diagram of elements of a behavior pattern recognition system according to the present invention; and -
FIG. 4 is a schematic view of a behavior pattern recognition system according to another embodiment of the present invention. - Hereinafter, the details of the embodiments of the present invention will be illustrated with reference to the drawings to make features and advantages of the present invention more comprehensible.
-
FIG. 1 is a flow chart of steps of a behavior pattern recognition method according to the present invention. The present invention provides a behavior pattern recognition method, which is applicable to an electronic device having a storage unit for storing multiple sets of behavior record information. The method includes the following step. - In Step S110, firstly, a first detecting unit acquires first behavior feature information.
- In this embodiment, a plurality of detecting units must be used to acquire a plurality of sample behavior information and analyze the sample behavior information to generate behavior record information. Coherence of the detecting unit and the first detecting unit is compared, the detecting units having the coherence with the first detecting unit exceeding a preset value are screened out to serve as the at least one second detecting unit according to the coherence.
- In this embodiment, the method further includes disposing a behavior analysis module to analyze the sample behavior information to generate a plurality of behavior record information.
- In Step S120, a collaboration network module acquires at least one second detecting unit having coherence with the first detecting unit.
- In this embodiment, the first detecting unit and the second detecting unit are non-invasive detectors, and the non-invasive detector includes an electrical detector, a sound sensor, an infrared sensor, a video/audio recorder, an electromagnetic sensor, and a mobile phone having detecting and sensing functions.
- In Step S130, the at least one second detecting unit acquires at least one second behavior feature information.
- In this embodiment, the first behavior feature information and the at least one second behavior feature information are acquired respectively by the first detecting unit and the at least one second detecting unit according to a preset time interval.
- In Step S140, a processing unit compares the at least one second behavior feature information and the behavior record information to generate at least one comparison result.
- In Step S150, a behavior definition represented by the first behavior feature information is determined according to the comparison result.
-
FIG. 2 is a flow chart of steps of a behavior pattern recognition method according to the present invention and preparation works thereof. The method includes the following steps. - In Step S210, a first detecting unit acquires first behavior feature information.
- In Step S220, a plurality of detecting units acquires a plurality of sample behavior information.
- In Step S230, the sample behavior information is analyzed to generate behavior record information.
- In Step S240, coherence of the detecting units and the first detecting unit is compared.
- In Step S250, the detecting units having coherence with the first detecting unit exceeding a preset value are screened out to serve as the at least one second detecting unit according to the coherence.
- In Step S260, a collaboration network module acquires at least one second detecting unit having coherence with the first detecting unit.
- In Step S270, at least one second detecting unit acquires at least one second behavior feature information.
- In Step S280, a processing unit compares the at least one second behavior feature information and the behavior record information to generate at least one comparison result.
- In Step S290, a behavior definition represented by the first behavior feature information is determined according to the comparison result.
-
FIG. 3 is a schematic block diagram of elements of a behavior pattern recognition system according to the present invention. In this figure, a behavior pattern recognition system of the present invention is shown, which includes astorage unit 310, a first detectingunit 320, at least one second detectingunit 330, and aprocessing unit 340. Thestorage unit 310 stores multiple sets ofbehavior record information 311, and the first detectingunit 320 acquire firstbehavior feature information 321. The at least one second detectingunit 330 has coherence with the first detectingunit 320, and acquires at least one secondbehavior feature information 331. Theprocessing unit 340 compares the at least one secondbehavior feature information 331 with thebehavior record information 311 to generate at least one comparison result, and a behavior definition represented by the firstbehavior feature information 321 is determined according to the comparison result. - In this embodiment, the system further includes a plurality of detecting units to acquire a plurality of sample behavior information.
- In this embodiment, the system further includes a behavior analysis module to analyze the sample behavior information to generate a plurality of behavior record information.
- In this embodiment, the system further includes a collaboration network module to acquire at least one second detecting unit having coherence with the first detecting unit.
- In this embodiment, the first detecting unit and the second detecting unit are non-invasive detectors. The non-invasive detector includes an electrical detector, a sound sensor, an infrared sensor, a video/audio recorder, an electromagnetic sensor, and a mobile phone having detecting and sensing functions.
- In this embodiment, a time interval is preset in the first detecting unit and the first behavior feature information is acquired according to the time interval.
- In this embodiment, a time interval is preset in the at least one second detecting unit and at least one second behavior feature information is acquired according to the time interval.
- In this embodiment, the coherence of the first detecting unit and at least one second detecting unit includes position information, for example, at the position of an office or a classroom.
-
FIG. 4 is a schematic view of a behavior pattern recognition system according to another embodiment of the present invention. In this embodiment, a first detectingunit 410 detects the behavior feature information of the user, and firstly the behavior feature information of the first detectingunit 410, a second detectingunit 420 and a third detectingunit 430 are input to afeature capturing device 440. For the ease of illustration, the second detectingunit 420 is preset to be other users coherent to the user. - In this embodiment, the
feature capturing device 440 captures feature by observation, and different sensors may be applicable to different feature capturing methods, but the binary signals like PIR or Reed Switch do not need this method. - Then, the behavior feature information of the second detecting
unit 420 and the third detectingunit 430 passes through thefeature capturing device 440 and is input in asorting module 450 and acollaboration network module 460. In this embodiment, thesorting module 450 mainly identifies status of other users, such as at work, leave the seat, and off work. -
Group information 470 is added in thecollaboration network module 460 to screen out the behavior feature information of the second detectingunit 420 and the third detectingunit 430. Then, the collaboration network inputs the user behavior feature information of the second detectingunit 420 related to the user of the first detectingunit 410 to abehavior recognition module 480, thereby determining a behavior definition represented by the user behavior feature information of the first detectingunit 410. - In this embodiment, the user status values of the first detecting
unit 410 and the second detectingunit 420 are taken as the input values, and the output values are the user behavior, such as, at the seat (operating the computer), at the seat (other behaviors), leaving the seat (at the meeting inside), leaving the seat (at the meeting outside), leaving the seat (others), or off work. - In summary, the present invention adopts a group interaction structure and utilizes a feature capturing method and a machine learning method in cooperation with the group interaction model to deduce the user behavior, and thus the present invention may be schemed in the following fields.
- (1) The features of the present invention may be used to assist the enterprises or consultancy companies to clearly know the social network status and the working status of the staff, thereby providing a solution for the enterprises, improving the working efficiency, the innovation of the enterprises and the working satisfaction. Different from collecting the poll, the working status of the staff are more clearly, so suggestion and assistance may be made to those having a low working efficiency.
- (2) In regard with the aged care, the family members of a senior may know the living status of the senior by using this system, and this system may record detailed living behaviors and can also be used to identify the health conditions of the senior.
- (3) In regard with the surveillance on the behaviors of the kindergarten children, the parents may know the child's behaviors and the activities of the child by using this system.
- It will be apparent to those skilled in the art that various modifications and variations can be made to the structure of the present invention without departing from the scope or spirit of the invention. In view of the foregoing, it is intended that the present invention cover modifications and variations of this invention provided they fall within the scope of the following claims.
Claims (18)
1. A behavior pattern recognition method, applicable to an electronic device having a storage unit for storing multiple sets of behavior record information, comprising:
acquiring first behavior feature information by a first detecting unit;
acquiring at least one second detecting unit having coherence with the first detecting unit by a collaboration network module;
acquiring at least one second behavior feature information by at least one second detecting unit;
comparing the at least one second behavior feature information with the behavior record information by a processing unit to generate at least one comparison result; and
determining a behavior definition represented by the first behavior feature information according to the comparison result.
2. The behavior pattern recognition method according to claim 1 , wherein before the step of acquiring at least one second detecting unit having coherence with the first detecting unit by the collaboration network module, the method further comprises:
acquiring a plurality of sample behavior information by a plurality of detecting units; and
analyzing the sample behavior information to generate the behavior record information.
3. The behavior pattern recognition method according to claim 2 , further comprising:
comparing coherence of the detecting units with the first detecting unit; and
screening out the detecting units having coherence with the first detecting unit exceeding a preset value to serve as the at least one second detecting unit according to the coherence.
4. The behavior pattern recognition method according to claim 2 , further comprising disposing a behavior analysis module to analyze the sample behavior information to generate a plurality of behavior record information.
5. The behavior pattern recognition method according to claim 1 , wherein the first detecting unit and the second detecting unit are non-invasive detectors.
6. The behavior pattern recognition method according to claim 5 , wherein the non-invasive detector comprises an electrical detector, a sound sensor, an infrared sensor, a video/audio recorder, an electromagnetic sensor, and a mobile phone having detecting and sensing functions.
7. The behavior pattern recognition method according to claim 1 , further comprising acquiring and storing coherence information of time and the behavior feature information, wherein the step comprises:
acquiring the first behavior feature information and the at least one second behavior feature information respectively by the first detecting unit and the at least one second detecting unit according to a preset time interval.
8. The behavior pattern recognition method according to claim 1 , wherein the coherence of the first detecting unit and the at least one second detecting unit comprises position information.
9. A behavior pattern recognition system, comprising:
a storage unit, for storing multiple sets of behavior record information;
a first detecting unit, for acquiring first behavior feature information;
at least one second detecting unit, having coherence with the first detecting unit and acquiring at least one second behavior feature information; and
a processing unit, for comparing the at least one second behavior feature information and the behavior record information to generate at least one comparison result and determining a behavior definition represented by the first behavior feature information according to the comparison result.
10. The behavior pattern recognition system according to claim 9 , further comprising a plurality of detecting units for acquiring a plurality of sample behavior information.
11. The behavior pattern recognition system according to claim 9 , further comprising a behavior analysis module, for analyzing the sample behavior information to generate a plurality of behavior record information.
12. The behavior pattern recognition system according to claim 9 , further comprising a collaboration network module, for acquiring at least one second detecting unit having coherence with the first detecting unit.
13. The behavior pattern recognition system according to claim 9 , wherein the first detecting unit and the second detecting unit are non-invasive detectors.
14. The behavior pattern recognition system according to claim 13 , wherein the non-invasive detector comprises an electrical detector, a sound sensor, an infrared sensor, a video/audio recorder, an electromagnetic sensor, and a mobile phone having detecting and sensing functions.
15. The behavior pattern recognition system according to claim 9 , wherein a time interval is preset in the first detecting unit, and the first behavior feature information is acquired according to the time interval.
16. The behavior pattern recognition system according to claim 9 , wherein a time interval is preset in the at least one second detecting unit, and the at least one second behavior feature information is acquired according to the time interval.
17. The behavior pattern recognition system according to claim 9 , wherein the coherence of the first detecting unit and the at least one second detecting unit comprises position information.
18. A computer application program for behavior pattern recognition, applicable to an electronic device which carries out the behavior pattern recognition method and comprises a storage unit for storing multiple sets of behavior record information, and the method comprises:
acquiring first behavior feature information by a first detecting unit;
acquiring at least one second detecting unit having coherence with the first detecting unit by a collaboration network module;
acquiring at least one second behavior feature information by the at least one second detecting unit;
comparing the at least one second behavior feature information with the behavior record information by a processing unit to generate at least one comparison result; and
determining a behavior definition represented by the first behavior feature information according to the comparison result.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
TW099141005A TW201222431A (en) | 2010-11-26 | 2010-11-26 | Behavior pattern recognition method, system and computer application program thereof |
TW099141005 | 2010-11-26 |
Publications (1)
Publication Number | Publication Date |
---|---|
US20120136890A1 true US20120136890A1 (en) | 2012-05-31 |
Family
ID=46127335
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/969,254 Abandoned US20120136890A1 (en) | 2010-11-26 | 2010-12-15 | Behavior pattern recognition method, system and computer application program thereof |
Country Status (2)
Country | Link |
---|---|
US (1) | US20120136890A1 (en) |
TW (1) | TW201222431A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105094305A (en) * | 2014-05-22 | 2015-11-25 | 华为技术有限公司 | Method for identifying user behavior, user equipment and behavior identification server |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5517429A (en) * | 1992-05-08 | 1996-05-14 | Harrison; Dana C. | Intelligent area monitoring system |
US6049749A (en) * | 1996-12-13 | 2000-04-11 | Koito Manufacturing Co., Ltd. | Lighting device for a vehicle |
US20040143398A1 (en) * | 2003-01-03 | 2004-07-22 | Nelson Mitchell C. | Method and system for monitoring vibration and/or mechanical waves in mechanical systems |
US7006866B1 (en) * | 1997-11-07 | 2006-02-28 | Siemens Aktiengesellschaft | Arrangement for predicting an abnormality of a system and for carrying out an action which counteracts the abnormality |
US20070194979A1 (en) * | 2006-02-14 | 2007-08-23 | Furuno Electric Company, Ltd. | Navigational aid and carrier sense technique |
US20110081634A1 (en) * | 2009-10-02 | 2011-04-07 | Masatomo Kurata | Behaviour Pattern Analysis System, Mobile Terminal, Behaviour Pattern Analysis Method, and Program |
-
2010
- 2010-11-26 TW TW099141005A patent/TW201222431A/en unknown
- 2010-12-15 US US12/969,254 patent/US20120136890A1/en not_active Abandoned
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5517429A (en) * | 1992-05-08 | 1996-05-14 | Harrison; Dana C. | Intelligent area monitoring system |
US6049749A (en) * | 1996-12-13 | 2000-04-11 | Koito Manufacturing Co., Ltd. | Lighting device for a vehicle |
US7006866B1 (en) * | 1997-11-07 | 2006-02-28 | Siemens Aktiengesellschaft | Arrangement for predicting an abnormality of a system and for carrying out an action which counteracts the abnormality |
US20040143398A1 (en) * | 2003-01-03 | 2004-07-22 | Nelson Mitchell C. | Method and system for monitoring vibration and/or mechanical waves in mechanical systems |
US20070194979A1 (en) * | 2006-02-14 | 2007-08-23 | Furuno Electric Company, Ltd. | Navigational aid and carrier sense technique |
US20110081634A1 (en) * | 2009-10-02 | 2011-04-07 | Masatomo Kurata | Behaviour Pattern Analysis System, Mobile Terminal, Behaviour Pattern Analysis Method, and Program |
Non-Patent Citations (2)
Title |
---|
Andrew Liu, Deural Computation: Modeling and Prediction of Human Behavior, 01/1999, Massachusetts Institute of Technology, vol.11, 229-242 * |
Nuria Oliver et al, A Bayesian computer vision system for modeling human interactions, IEEE, 08/2000, vol.22 issue 8, pp.831-843 * |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105094305A (en) * | 2014-05-22 | 2015-11-25 | 华为技术有限公司 | Method for identifying user behavior, user equipment and behavior identification server |
EP3139287A4 (en) * | 2014-05-22 | 2017-05-17 | Huawei Technologies Co., Ltd. | User behavior recognition method, user equipment, and behavior recognition server |
US10409841B2 (en) | 2014-05-22 | 2019-09-10 | Huawei Technologies Co., Ltd. | User behavior recognition method, user equipment, and behavior recognition server |
EP3796134A1 (en) * | 2014-05-22 | 2021-03-24 | Huawei Technologies Co., Ltd. | User behavior recognition method, user equipment and behavior recognition server |
EP4339810A3 (en) * | 2014-05-22 | 2024-06-19 | Huawei Technologies Co., Ltd. | User behavior recognition method, user equipment, and behavior recognition server |
Also Published As
Publication number | Publication date |
---|---|
TW201222431A (en) | 2012-06-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7840515B2 (en) | System architecture and process for automating intelligent surveillance center operations | |
EP3279700A1 (en) | Security inspection centralized management system | |
JP5669082B2 (en) | Verification device | |
CN109040693B (en) | Intelligent alarm system and method | |
CN110089104A (en) | Event storage, event searching device and event alarms device | |
CN110705482A (en) | Personnel behavior alarm prompt system based on video AI intelligent analysis | |
CN105404849B (en) | Using associative memory sorted pictures to obtain a measure of pose | |
JP2021114338A (en) | Business activity analysis device, business activity analysis method, and program | |
WO2022041484A1 (en) | Human body fall detection method, apparatus and device, and storage medium | |
CN105940434A (en) | Information processing device, information processing method, and program | |
CN110516568B (en) | College multi-scene data management method and system based on face recognition | |
CN112567400A (en) | Information processing apparatus, information processing method, and job evaluation system | |
US10965916B2 (en) | Video file processing method, video file processing device and monitoring system | |
CN117238508B (en) | Information screening system based on artificial intelligence | |
CN207966119U (en) | The identifying system of equipment user's destruction in a kind of shop | |
US20120136890A1 (en) | Behavior pattern recognition method, system and computer application program thereof | |
US9111237B2 (en) | Evaluating an effectiveness of a monitoring system | |
US11676439B2 (en) | Face authentication system and face authentication method | |
CN112949619B (en) | Guest room sanitation monitoring method, electronic equipment and storage medium | |
JP2011150425A (en) | Research device and research method | |
JP2020095651A (en) | Productivity evaluation system, productivity evaluation device, productivity evaluation method, and program | |
CN110135744B (en) | Construction worker safety behavior habit evaluation method | |
CN115063752A (en) | Video tracking early warning method and system based on UWB positioning | |
Schlenke et al. | Towards activity recognition in smart homes using multimodal data | |
Rafferty et al. | NFC based dataset annotation within a behavioral alerting platform |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: INSTITUTE FOR INFORMATION INDUSTRY, TAIWAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:WEN, YUNG-CHUAN;LIANG, MIN-SIONG;REEL/FRAME:025513/0042 Effective date: 20101127 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |