WO2016016277A1 - A computer-implemented method and a system for remotely monitoring a user, and a computer program product implementing the method - Google Patents

A computer-implemented method and a system for remotely monitoring a user, and a computer program product implementing the method Download PDF

Info

Publication number
WO2016016277A1
WO2016016277A1 PCT/EP2015/067324 EP2015067324W WO2016016277A1 WO 2016016277 A1 WO2016016277 A1 WO 2016016277A1 EP 2015067324 W EP2015067324 W EP 2015067324W WO 2016016277 A1 WO2016016277 A1 WO 2016016277A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
biometrics
status
data
monitored user
Prior art date
Application number
PCT/EP2015/067324
Other languages
French (fr)
Inventor
Xavier BARÓ SOLÉ
Sergio Escalera Guerrero
Jordi GONZÁLEZ SABATÉ
Martha MACKAY JARQUE
Original Assignee
Universitat Autonoma De Barcelona
Centro De Visión Por Computador (Cvc)
Universitat De Barcelona
Acceplan Accesibilidad S.L.
Fundació Per A La Universitat Oberta De Catalunya (Uoc)
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Universitat Autonoma De Barcelona, Centro De Visión Por Computador (Cvc), Universitat De Barcelona, Acceplan Accesibilidad S.L., Fundació Per A La Universitat Oberta De Catalunya (Uoc) filed Critical Universitat Autonoma De Barcelona
Publication of WO2016016277A1 publication Critical patent/WO2016016277A1/en

Links

Classifications

    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H40/00ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices
    • G16H40/60ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices
    • G16H40/67ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices for remote operation
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/20ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems

Definitions

  • a computer-implemented method and a system for remotely monitoring a user and a computer program product implementing the method
  • the present invention generally relates, in a first aspect, to a computer- implemented method for remotely monitoring a user, comprising acquiring and analysing multi-modal user data, including depth data, of a monitored user, and more particularly to a method applied to a plurality of unidentified users.
  • a second aspect of the invention concerns to a system adapted to implement the method of the first aspect.
  • a third aspect of the invention relates to a computer program product adapted to implement the processing steps of the method of the first aspect of the invention.
  • US2014/0052464 A1 discloses a method and a system for remotely monitoring a patient by acquiring information corresponding to a condition of a patient from a plurality of input sources, selecting an analytical model from a database which also stores several analytical models and medical data, based on the information acquired and on the medical data for said patient, and determining a state of the patient and formulate a health prediction by performing another algorithm on the acquired information and medical data with the selected analytical model. Based on the determined state and on the health prediction, a recommendation is determined and transmitted to a remote entity.
  • EP2560141 A1 describes an interactive virtual care system which includes sensors to acquire multi-modal user data related to user movement, including depth image data, and analyses the acquired multi- modal user data to identify an anomaly in the user movement. Said analysis is performed by comparing the acquired multi-modal user data to predetermined historical user data (such as medical history) and/or statistical norm data for users to identify an anomaly in the user movement.
  • predetermined historical user data such as medical history
  • statistical norm data for users to identify an anomaly in the user movement.
  • the user movement, a real version and a computer- generated version, and the identified anomaly can be displayed in a responder interface which can be placed at a remote location.
  • EP2560141 A1 shares some of the drawbacks of the proposal made by US2014/0052464 A1 , as it also needs accessing historic data for the user, such as medical history, and thus can only work with previously identified users.
  • the present invention relates, in a first aspect, to a computer- implemented method for remotely monitoring a user, the method comprising:
  • multi-modal user data related to a monitored user including at least depth data, using at least a depth camera in a local surveillance area;
  • the method of the first aspect of the present invention is applied to a plurality of unidentified users and comprises:
  • Said information associated to the identified status of the monitored user presented to the remote computing means includes at least, for a preferred embodiment, information indicative of an alarm.
  • the information associated to the identified status of the monitored user presented to the remote computing means includes at least, for a preferred embodiment, information indicative of an alarm
  • the method further comprises presenting part or all of the acquired multi-modal user data to the remote computing means.
  • said multi-modal user data comprises at least audio data, RGB video data and infrared depth video data.
  • the mentioned biometrics are extracted from multi-modal user data including at least depth video o depth images data acquired using at least a depth camera, whether said depth camera and/or other depth cameras depending on the local surveillance area in which the user which biometrics are being extracted is placed.
  • the biometrics are extracted also from audio data and/or RGB video data of said multi-modal user data and/or from other suitable data sources.
  • the method of the first aspect of the invention comprises analysing, by said and/or another processing means, the acquired multimodal user data to perform the next steps:
  • step iv) performing, using said processing means, said identifying of the status of the monitored user analysing at least the detected skeleton point joints of step iii) and also movement and 3D geometry, regarding the monitored user, using at least depth information of said acquired multi-modal user data.
  • step iv) is performed, for an embodiment, by analysing also data from sensors carried by the monitored user (for example in the form of a bracelet, helmet, chest band, etc.), including an accelerometer and/or a gyroscope and/or a temperature sensor and or a blood pressure sensor and/or an oximeter sensor and/or an ECG sensor assembly and/or an EEG sensor assembly and/or an EMG sensor assembly.
  • sensors carried by the monitored user for example in the form of a bracelet, helmet, chest band, etc.
  • said processing means are implemented in local computing means comprising and/or connected with said depth camera, all of said steps i) to iv) being performed by said local computing means, the local computing means being bidirectionally connected to said remote computing means and delivering the latter with at least said identified status of the monitored user and/or said information associated thereto.
  • a local portion of said processing means are implemented in local computing means and a remote portion of the processing means are implemented in said remote computing means, some of said steps i) to iv) being performed by said local computing means and the rest of the steps i) to iv) being performed by said remote portion of the processing means implemented the remote computing means.
  • processing means are implemented only in said remote computing means, all of said steps i) to iv) being performed by the remote computing means.
  • step i) is performed, for a preferred embodiment, using colour image data and depth information acquired by the depth camera, applying HOG features in RGBD space and learning those features on a predefined database of users and nonusers, and, afterwards, extracting body pixels by depth thresholding and surface normal analysing in an area within a defined bounding box around the monitored user.
  • skeleton joint points they are extracted at step iii), for a preferred embodiment, applying a model-based Random forest algorithm, using at least depth information acquired by said depth camera, comprising dividing the depth image data corresponding to the body of the monitored user into body parts, attaching an identification pixel label to each body part, and building a gestures database with at least part of the attached identification pixel labels corresponding to different gestures, such as gestures associated to getting up, bending over, falling, picking up an object, etc.
  • the model-based Random forest algorithm can be applied to colour image data to complement the results obtained from the depth image data.
  • the method further comprises teaching the model-based Random forest algorithm through feature vectors on body pixels of said body parts and their identification pixel labels, and, later, using the model-based Random forest algorithm to predict each identification pixel label of a new detected body using its associated feature vector, wherein said feature vectors are build using relative depth information between different voxels of said body parts, such that the feature vectors comprise depth relations between said voxels.
  • the method of the first aspect of the invention comprises adding the detected skeleton joint points to a continuous stream, including at least a depth video or a sequence of depth images of the acquired multi-modal user data, and performing the identifying of the status of the monitored user at step iv) analysing said continuous stream, regarding the movement and 3D geometry of the monitored user, on temporal sequences of the continuous stream, together with the detected skeleton point joints.
  • the method further comprises presenting said continuous stream to the remote computing means.
  • the method comprises performing the biometrics extraction used for the recognizing of step ii) after the skeleton joints points detection of step iii) has been performed, and performing said biometrics extraction by at least analysing the spatial relations between the detected skeleton joint points (for example by correlating body limbs).
  • the method of the first aspect of the invention comprises building a status database with information regarding different styles of status of users, and using the information included in said status database to teach an algorithm in charge of performing step iv) to define the action the monitored user is doing based on the continuous stream to identify the status of the monitored user, which, generally, is a behavioural status, such as: walking, sitting, standing, laying down, watching TV, sleeping and/or falling.
  • the method comprises performing in real time at least the identification of the status of the monitored user.
  • the method of eh first aspect of the invention comprises using a single user multi camera system comprising a plurality of depth cameras aiming to different local surveillance areas through which the user can circulate, and performing the acquiring of the multi-modal user data and processing thereof by using only the depth camera aimed to the area where the user is placed, and stop processing for other cameras until the user is available within their respective surveillance areas, in order to reduce processing such that the method operates in real time.
  • the method comprises, for an embodiment, performing the biometrics extraction from the above mentioned continuous stream, generally, regarding the monitored user, from an initial portion of the continuous stream and performing the analysis to identify a status of the monitored user on next portions of the same continuous stream.
  • a second aspect of the present invention relates to a system for remotely monitoring a user, comprising:
  • - local acquiring means for acquiring multi-modal user data related to a monitored user, said local acquiring means comprising at least one depth camera to acquire the multi-modal user data in a local surveillance area; - processing means for analysing, by means of data processing, the acquired multi-modal user data to identify a status of the monitored user;
  • system of the second aspect of the invention is a multi-user system and comprises:
  • biometrics database correlated to identity information of the correspondent user
  • the system o the second aspect of the invention is adapted for performing the method of the first aspect, where all the steps of the method, except for the step of acquiring multi-modal user data (which is performed by at least a depth camera), are performed by data processing by means of said processing means, the processing means being implemented in the remote computing means and/or in local computing means included in the system.
  • the remote computing means are adapted for establishing bidirectional communication with the local computing means, to allow, to an authorized person, to control, configure and upgrade operation settings of the local computing means.
  • the remote computing means are implemented in a portable computing device, such as a smartphone or a tablet, and has installed therein a software application for performing at least part of the above mentioned authorized control, configuration and upgrade of operation settings of the local computing means, and/or to manage received identified status and alarms and/or to perform the identification of said status and the generation of said alarms, and/or to receive or build the continuous stream and show the latter in a display of the portable computing device and/or to present part or all of the acquired multi-modal user data.
  • a portable computing device such as a smartphone or a tablet
  • a software application for performing at least part of the above mentioned authorized control, configuration and upgrade of operation settings of the local computing means, and/or to manage received identified status and alarms and/or to perform the identification of said status and the generation of said alarms, and/or to receive or build the continuous stream and show the latter in a display of the portable computing device and/or to present part or all of the acquired multi-modal user data.
  • a third aspect of the invention relates to a computer program product, which includes code instructions that when executed in a computer implement all the steps of the method of the first aspect of the invention, except for the step of acquiring multimodal user data.
  • the computer program product comprises a first software application running on the above mentioned local computing means and a second software application constituting the above mentioned software application installed in the portable computing device, said first and second software application having capabilities for exchanging data with each other.
  • FIG. 1 schematical ly shows the system of the present invention, for an embodiment
  • FIG. 2 depicts different examples of use case scenarios of the system of the present invention, showing some possible identified user status of the monitored user.
  • the system of the present invention can be used for remotely monitoring different kind of users, it has been particularly thought for elders monitoring, to help people protecting elders easier and in a more comfortable way besides high safety in indoor applications. It provides the tools, both hardware and software, in a package needed for such a monitoring. For a preferred embodiment, it is a portable system that can be put in any place in the room where it can see the user.
  • the system uses a single camera for monitoring. While this is a client-side specification, for another embodiment, a server can be involved for a multi-camera system in more elaborated versions to bring the framework to the client in a cheaper, robust and wider usage. This multi-camera system allows monitoring users automatically everywhere in indoor environments.
  • the system uses a depth camera, such as the Microsoft Kinect® camera, to capture the environment in colour and depth, along with a specific usage computer to process the raw information from the camera.
  • a depth camera such as the Microsoft Kinect® camera
  • the client interacts with the system through an internet line using a Smartphone/PC based application.
  • the client can see the user instantly through the depth camera (for personal rights, just the depth image can be shown to the client) or view the messages of the system sent to his/her Smartphone/PC based application.
  • the messages are sent based on the predefined user status.
  • a ringtone alarm would occur on the client Smartphone/PC if the user status was urgent.
  • the client puts the depth camera where the user can be seen from the camera viewpoint, connects the usb cable of the depth camera to the processor computer (called in a previous section as local computing means), connects the power cables of the depth camera and computer, and finally runs the system.
  • the client can install the Smartphone/PC based application on his/her device (called in a previous section as remote computing means) easily and communicate with the system, set the settings, upgrade the software, etc.
  • the security protocols do not allow anonymous persons reach the system illegally. Therefore, the system is easy to work and safe enough.
  • FIG. 1 a technical implementation of the system of the present invention is shown, including the next elements:
  • a depth camera C to acquire the multi-modal user data in a local surveillance area, including sound, depth images and colour images which are conditioned in a suitable format;
  • a local processor computer L implementing a core, for analysing the acquired multi-modal user data to identify a status of the monitored user
  • a core implementation to analyse the raw information achieved by the depth camera C which is installed on the local processor computer L and provides the feedbacks to the client R. This part is responsible to detect the user and monitor it.
  • the architecture of this part can be seen in the Figure 1 .
  • a GUI to interact with the core L.
  • This is the client-side software and is responsible to receive messages (such as the results of the recognizing of users) and alarms (and any result of the identified status, if needed, even if not associated to an alarm status) from the core and show them to the client R, and upgrade new versions of the core and set new user biometrics (if needed) on the processor computer L.
  • messages such as the results of the recognizing of users
  • alarms and any result of the identified status, if needed, even if not associated to an alarm status
  • local processor computer L implements the steps of the method already described in a previous section, by means of the illustrated functional modules, regarding the detection of users using a Human features Database (DB), the recognizing of the detected users using an users biometrics database (DB), the detection of skeleton joint points using a pixels labels Database (DB), the generation of a continuous stream including the detected skeleton joint points and also a depth video and the analysis of said continuous stream to identify the status of the patient.
  • DB Human features Database
  • DB users biometrics database
  • DB pixels labels Database
  • the depth camera C gives the scene data in 30 fps.
  • This data must be processed by the core in real time which is a critical task.
  • the first task is detecting users pixels and separate them from the background. This can be done in real time using background subtraction (see T. Bouwmans, F. Porikli, B. Horferlin, A. Vacavant, Handbook on "Background Modeling and Foreground Detection for Video Surveillance: Traditional and Recent Approaches, Implementations, Benchmarking and Evaluation", CRC Press, Taylor and Francis Group, June 2014) and depth thresholding method then tracking them (see Lyudmila Mihaylova, Paul Brasnett, Nishan Canagarajan and David Bull (2007).
  • the functional module "Recognize users” has been placed before the functional module "Detect skeleton joint points", the recognizing of users is preferably done after detecting skeleton joint points, because then biometrics are easier to extract and there is a correlation among body limbs.
  • Skeleton joint points are extracted applying a superfast (using GPU programming) model-based Random forest approach (see Jamie Shotton, Andrew Fitzgibbon, Mat Cook, Toby Sharp, Mark Finocchio, Richard Moore, Alex Kipman, and Andrew Blake, Real-Time Human Pose Recognition in Parts from a Single Depth Image, in CVPR, IEEE, June 201 1 ).
  • body is divided into parts attached a label to each part, and then a database of pixel labels of very different gestures is composed.
  • Random forest is a learning algorithm and it is taught, according to the present invention, through feature vectors on body pixels and their labels. Later, the algorithm predicts each pixel label of the new detected body using its associated feature vector.
  • Feature vectors are using depth information of other pixels to be computed, particularly using relative depth information between different voxels of the body parts.
  • Body parts are defined such that mean point of each part shows the joint point of it.
  • Working with depth enables us to continue without light. In working with RGB images in presence of light, some works extracted joint points instantly from body bounding box but the accuracy is less than working with depth.
  • Skeleton joint points are added to the stream and using this continuous stream the status of the user can be analysed, i.e. identified from analysis the continuous stream.
  • different status can be defined, walking, sitting, standing, laying down, watching TV, sleeping, falling, etc.
  • a database for different styles of each status is made and an algorithm to define the action the user is doing based on the current stream is taught.
  • a solution can be hierarchical aligned cluster analysis (see F. Zhou, F. De la Torre and J. K. Hodgins. Hierarchical Aligned Cluster Analysis for Temporal Clustering of Human Motion. IEEE Transactions Pattern Analysis and Machine Intelligence (PAMI), vol. 35, no. 3, pp. 582-596, 2013).
  • the system can be tuned to reduce processing and bring the system into real time. For instance, using the core for initial recognition and applying tracking in the next frames, or in a single user multi camera system, focusing the system on the camera sees the user and stop processing for other cameras until he is available.
  • FIG. 2 shows several depth cameras C connected to their respective local computing means L in three different situations of monitored users: - Scenario 1 (User walking at home): Users with a certain degree of mobility but who can move freely in their homes, for example a dining room.
  • Scenario 1 User walking at home
  • - Scenario 2 (User watching TV): Many carers leave home at those moments when the person they are taking care of is performing an activity that does not involve any movement, for example watching TV or also performing resting activities (reading, eating, etc.). In these cases, the user is monitored in a small space like a room or a hall.
  • each local computing means L communicate with the remote computing means R, sending to the latter the above described data, including biometric information about the user and, generally, the above described continuous stream and identified user status (although depending on the embodiment, the building of said stream and/or the identification of the user status can be performed whether by the local computing means L or by the remote computing means R).
  • the remote computing means R analyses the received data and is responsible for communicating with the devices of authorized persons, like caregivers to the elderly or relatives, through the network with information indicative of an alarm.

Abstract

The method is applied to a plurality of unidentified users and comprises: -automatically extracting biometrics for at least some of the plurality of unidentified users and storing the extracted biometrics in a biometrics database correlated to identity information of the correspondent user, -acquiring multi-modal user data for a monitored user, including at least depth data; -recognizing the identity of the monitored user by automatically extracting biometrics for the monitored user and searching the biometrics database to look for a match; -analysing the acquired multi-modal user data to identify a status of the monitored user; and -presenting the identified status and/or information associated thereto, to a remote computing means. The system and the computer program product are both adapted to implement the method of the invention.

Description

A computer-implemented method and a system for remotely monitoring a user, and a computer program product implementing the method
Field of the Invention
The present invention generally relates, in a first aspect, to a computer- implemented method for remotely monitoring a user, comprising acquiring and analysing multi-modal user data, including depth data, of a monitored user, and more particularly to a method applied to a plurality of unidentified users.
A second aspect of the invention concerns to a system adapted to implement the method of the first aspect.
A third aspect of the invention relates to a computer program product adapted to implement the processing steps of the method of the first aspect of the invention.
Background of the Invention
Remote monitoring of users in indoor applications, such as patients of a hospital or elders of a residence, nowadays is a need which has not yet been completely satisfied.
However, there are some known proposals focused on such remote user monitoring, although they are clearly susceptible of improving. Next, some patent documents disclosing such proposals are identified and their relevant background briefly described.
US2014/0052464 A1 discloses a method and a system for remotely monitoring a patient by acquiring information corresponding to a condition of a patient from a plurality of input sources, selecting an analytical model from a database which also stores several analytical models and medical data, based on the information acquired and on the medical data for said patient, and determining a state of the patient and formulate a health prediction by performing another algorithm on the acquired information and medical data with the selected analytical model. Based on the determined state and on the health prediction, a recommendation is determined and transmitted to a remote entity.
The proposal made by US2014/0052464 A1 has several drawbacks, including the needing of working only with identified patients and of having access to the medical data of the patients.
Another of said proposals is disclosed by EP2560141 A1 , which describes an interactive virtual care system which includes sensors to acquire multi-modal user data related to user movement, including depth image data, and analyses the acquired multi- modal user data to identify an anomaly in the user movement. Said analysis is performed by comparing the acquired multi-modal user data to predetermined historical user data (such as medical history) and/or statistical norm data for users to identify an anomaly in the user movement. The user movement, a real version and a computer- generated version, and the identified anomaly can be displayed in a responder interface which can be placed at a remote location.
The system disclosed by EP2560141 A1 shares some of the drawbacks of the proposal made by US2014/0052464 A1 , as it also needs accessing historic data for the user, such as medical history, and thus can only work with previously identified users.
Summary of the invention
It is an object of the present invention to offer an alternative to the prior state of the art, with the purpose of providing a method and a system for remotely monitoring a user which lacks of the drawbacks of the above cited proposals, and is able of working with unidentified users, and thus without the need of using users historic data.
To that end, the present invention relates, in a first aspect, to a computer- implemented method for remotely monitoring a user, the method comprising:
- acquiring multi-modal user data related to a monitored user, including at least depth data, using at least a depth camera in a local surveillance area;
- analysing, by processing means, the acquired multi-modal user data to identify a status of the monitored user; and
- presenting said identified status of the monitored user and/or information associated thereto, to a remote computing means.
Contrary to the known methods, the method of the first aspect of the present invention is applied to a plurality of unidentified users and comprises:
- automatically extracting biometrics for at least some of said plurality of unidentified users and storing the extracted biometrics in a biometrics database correlated to identity information of the correspondent user, and
- recognizing the identity of the monitored user by automatically extracting biometrics for the monitored user and searching the biometrics database to look for a match.
Said information associated to the identified status of the monitored user presented to the remote computing means includes at least, for a preferred embodiment, information indicative of an alarm. For another embodiment, the information associated to the identified status of the monitored user presented to the remote computing means includes at least, for a preferred embodiment, information indicative of an alarm
For an embodiment, the method further comprises presenting part or all of the acquired multi-modal user data to the remote computing means.
For an embodiment, said multi-modal user data comprises at least audio data, RGB video data and infrared depth video data.
According to a preferred embodiment, the mentioned biometrics are extracted from multi-modal user data including at least depth video o depth images data acquired using at least a depth camera, whether said depth camera and/or other depth cameras depending on the local surveillance area in which the user which biometrics are being extracted is placed.
Optionally, the biometrics are extracted also from audio data and/or RGB video data of said multi-modal user data and/or from other suitable data sources.
According to an embodiment, the method of the first aspect of the invention comprises analysing, by said and/or another processing means, the acquired multimodal user data to perform the next steps:
i) detecting the monitored user,
ii) performing said recognizing of the monitored user,
iii) detecting skeleton point joints of the monitored user, and
iv) performing, using said processing means, said identifying of the status of the monitored user analysing at least the detected skeleton point joints of step iii) and also movement and 3D geometry, regarding the monitored user, using at least depth information of said acquired multi-modal user data.
The above mentioned step iv) is performed, for an embodiment, by analysing also data from sensors carried by the monitored user (for example in the form of a bracelet, helmet, chest band, etc.), including an accelerometer and/or a gyroscope and/or a temperature sensor and or a blood pressure sensor and/or an oximeter sensor and/or an ECG sensor assembly and/or an EEG sensor assembly and/or an EMG sensor assembly.
For an embodiment, said processing means are implemented in local computing means comprising and/or connected with said depth camera, all of said steps i) to iv) being performed by said local computing means, the local computing means being bidirectionally connected to said remote computing means and delivering the latter with at least said identified status of the monitored user and/or said information associated thereto. For an alternative embodiment, a local portion of said processing means are implemented in local computing means and a remote portion of the processing means are implemented in said remote computing means, some of said steps i) to iv) being performed by said local computing means and the rest of the steps i) to iv) being performed by said remote portion of the processing means implemented the remote computing means.
For another alternative embodiment, the processing means are implemented only in said remote computing means, all of said steps i) to iv) being performed by the remote computing means.
The detection of step i) is performed, for a preferred embodiment, using colour image data and depth information acquired by the depth camera, applying HOG features in RGBD space and learning those features on a predefined database of users and nonusers, and, afterwards, extracting body pixels by depth thresholding and surface normal analysing in an area within a defined bounding box around the monitored user.
Regarding the skeleton joint points, they are extracted at step iii), for a preferred embodiment, applying a model-based Random forest algorithm, using at least depth information acquired by said depth camera, comprising dividing the depth image data corresponding to the body of the monitored user into body parts, attaching an identification pixel label to each body part, and building a gestures database with at least part of the attached identification pixel labels corresponding to different gestures, such as gestures associated to getting up, bending over, falling, picking up an object, etc. Optionally, if necessary, the model-based Random forest algorithm can be applied to colour image data to complement the results obtained from the depth image data.
Preferably, the method further comprises teaching the model-based Random forest algorithm through feature vectors on body pixels of said body parts and their identification pixel labels, and, later, using the model-based Random forest algorithm to predict each identification pixel label of a new detected body using its associated feature vector, wherein said feature vectors are build using relative depth information between different voxels of said body parts, such that the feature vectors comprise depth relations between said voxels.
According to an embodiment, the method of the first aspect of the invention comprises adding the detected skeleton joint points to a continuous stream, including at least a depth video or a sequence of depth images of the acquired multi-modal user data, and performing the identifying of the status of the monitored user at step iv) analysing said continuous stream, regarding the movement and 3D geometry of the monitored user, on temporal sequences of the continuous stream, together with the detected skeleton point joints.
For an embodiment, the method further comprises presenting said continuous stream to the remote computing means.
Preferably, the method comprises performing the biometrics extraction used for the recognizing of step ii) after the skeleton joints points detection of step iii) has been performed, and performing said biometrics extraction by at least analysing the spatial relations between the detected skeleton joint points (for example by correlating body limbs).
For an embodiment, the method of the first aspect of the invention comprises building a status database with information regarding different styles of status of users, and using the information included in said status database to teach an algorithm in charge of performing step iv) to define the action the monitored user is doing based on the continuous stream to identify the status of the monitored user, which, generally, is a behavioural status, such as: walking, sitting, standing, laying down, watching TV, sleeping and/or falling.
Preferably, the method comprises performing in real time at least the identification of the status of the monitored user.
According to an embodiment, the method of eh first aspect of the invention comprises using a single user multi camera system comprising a plurality of depth cameras aiming to different local surveillance areas through which the user can circulate, and performing the acquiring of the multi-modal user data and processing thereof by using only the depth camera aimed to the area where the user is placed, and stop processing for other cameras until the user is available within their respective surveillance areas, in order to reduce processing such that the method operates in real time.
In order also to further reduce processing, the method comprises, for an embodiment, performing the biometrics extraction from the above mentioned continuous stream, generally, regarding the monitored user, from an initial portion of the continuous stream and performing the analysis to identify a status of the monitored user on next portions of the same continuous stream.
A second aspect of the present invention relates to a system for remotely monitoring a user, comprising:
- local acquiring means for acquiring multi-modal user data related to a monitored user, said local acquiring means comprising at least one depth camera to acquire the multi-modal user data in a local surveillance area; - processing means for analysing, by means of data processing, the acquired multi-modal user data to identify a status of the monitored user;
- remote computing means; and
- means for presenting said identified status of the monitored user and/or information associated thereto, to said remote computing means.
Contrary to the known systems, the system of the second aspect of the invention is a multi-user system and comprises:
- means for automatically extracting biometrics for at least some of a plurality of unidentified users and storing the extracted biometrics in a biometrics database correlated to identity information of the correspondent user, said biometrics database being included in the system, and
- means for recognizing the identity of the monitored user by automatically extracting biometrics for the monitored user and searching the biometrics database to look for a match.
The system o the second aspect of the invention is adapted for performing the method of the first aspect, where all the steps of the method, except for the step of acquiring multi-modal user data (which is performed by at least a depth camera), are performed by data processing by means of said processing means, the processing means being implemented in the remote computing means and/or in local computing means included in the system.
According to an embodiment, the remote computing means are adapted for establishing bidirectional communication with the local computing means, to allow, to an authorized person, to control, configure and upgrade operation settings of the local computing means.
For an embodiment, the remote computing means are implemented in a portable computing device, such as a smartphone or a tablet, and has installed therein a software application for performing at least part of the above mentioned authorized control, configuration and upgrade of operation settings of the local computing means, and/or to manage received identified status and alarms and/or to perform the identification of said status and the generation of said alarms, and/or to receive or build the continuous stream and show the latter in a display of the portable computing device and/or to present part or all of the acquired multi-modal user data.
A third aspect of the invention relates to a computer program product, which includes code instructions that when executed in a computer implement all the steps of the method of the first aspect of the invention, except for the step of acquiring multimodal user data. According to an embodiment, the computer program product comprises a first software application running on the above mentioned local computing means and a second software application constituting the above mentioned software application installed in the portable computing device, said first and second software application having capabilities for exchanging data with each other.
The previous and other advantages and features will be better understood from the following detailed description of embodiments, with reference to the attached drawings, which must be considered in an illustrative and non-limiting manner. Brief description of the drawings
FIG. 1 schematical ly shows the system of the present invention, for an embodiment; and
FIG. 2 depicts different examples of use case scenarios of the system of the present invention, showing some possible identified user status of the monitored user.
Detailed description of particular embodiments
Although the system of the present invention can be used for remotely monitoring different kind of users, it has been particularly thought for elders monitoring, to help people protecting elders easier and in a more comfortable way besides high safety in indoor applications. It provides the tools, both hardware and software, in a package needed for such a monitoring. For a preferred embodiment, it is a portable system that can be put in any place in the room where it can see the user.
For an embodiment or basic version of the system, it uses a single camera for monitoring. While this is a client-side specification, for another embodiment, a server can be involved for a multi-camera system in more elaborated versions to bring the framework to the client in a cheaper, robust and wider usage. This multi-camera system allows monitoring users automatically everywhere in indoor environments.
In the basic version, the system uses a depth camera, such as the Microsoft Kinect® camera, to capture the environment in colour and depth, along with a specific usage computer to process the raw information from the camera. This is the basics of the system. Then, based on the client usages, an alarm system is needed. Here, the client interacts with the system through an internet line using a Smartphone/PC based application. The client can see the user instantly through the depth camera (for personal rights, just the depth image can be shown to the client) or view the messages of the system sent to his/her Smartphone/PC based application. The messages are sent based on the predefined user status. A ringtone alarm would occur on the client Smartphone/PC if the user status was urgent.
At the first time, the client puts the depth camera where the user can be seen from the camera viewpoint, connects the usb cable of the depth camera to the processor computer (called in a previous section as local computing means), connects the power cables of the depth camera and computer, and finally runs the system. Now, the client can install the Smartphone/PC based application on his/her device (called in a previous section as remote computing means) easily and communicate with the system, set the settings, upgrade the software, etc. The security protocols do not allow anonymous persons reach the system illegally. Therefore, the system is easy to work and safe enough.
In FIG. 1 a technical implementation of the system of the present invention is shown, including the next elements:
- a depth camera C to acquire the multi-modal user data in a local surveillance area, including sound, depth images and colour images which are conditioned in a suitable format;
- a local processor computer L, implementing a core, for analysing the acquired multi-modal user data to identify a status of the monitored user; and
- remote computing means R, or client (two of them are illustrated in Figure 1 ), bidirectionally communicated with the local computing means by a user interface Ul.
Two kinds of software applications are used:
1 . A core implementation to analyse the raw information achieved by the depth camera C, which is installed on the local processor computer L and provides the feedbacks to the client R. This part is responsible to detect the user and monitor it. The architecture of this part can be seen in the Figure 1 .
2. And a GUI to interact with the core L. This is the client-side software and is responsible to receive messages (such as the results of the recognizing of users) and alarms (and any result of the identified status, if needed, even if not associated to an alarm status) from the core and show them to the client R, and upgrade new versions of the core and set new user biometrics (if needed) on the processor computer L. Although two clients R (i.e. remote computing means) have been depicted in Figure 1 for performing different functions, each client R can perform the described functions involving the shown and described bidirectional data flow.
Particularly, according to the illustrated embodiment, local processor computer L implements the steps of the method already described in a previous section, by means of the illustrated functional modules, regarding the detection of users using a Human features Database (DB), the recognizing of the detected users using an users biometrics database (DB), the detection of skeleton joint points using a pixels labels Database (DB), the generation of a continuous stream including the detected skeleton joint points and also a depth video and the analysis of said continuous stream to identify the status of the patient.
For an implementation of the system/method of the present invention, the depth camera C gives the scene data in 30 fps. This data must be processed by the core in real time which is a critical task. The first task is detecting users pixels and separate them from the background. This can be done in real time using background subtraction (see T. Bouwmans, F. Porikli, B. Horferlin, A. Vacavant, Handbook on "Background Modeling and Foreground Detection for Video Surveillance: Traditional and Recent Approaches, Implementations, Benchmarking and Evaluation", CRC Press, Taylor and Francis Group, June 2014) and depth thresholding method then tracking them (see Lyudmila Mihaylova, Paul Brasnett, Nishan Canagarajan and David Bull (2007). Object Tracking by Particle Filtering Techniques in Video Sequences; In: Advances and Challenges in Multisensor Data and Information. NATO Security Through Science Series, 8. Netherlands: IOS Press, pp. 260-268), which is inaccurate mostly when the user is near to other stuffs in the room, and needs an initial gesture or movement. Instead, a superfast approach is used according to the present invention, thanks to the GPU programming, applying HOG features (see V. Prisacariu, I. Reid, "fastHOG - a real-time GPU implementation of HOG", Technical report, University of Oxford, 2012) in RGBD space and learning those features on a predefined database of users and nonusers. Afterwards, while a bounding box around the user is placed, body pixels are extracted by depth thresholding and surface normal analysing (see http://en.wikipedia.org/wiki/Normal_(geometry)).
In multi-user systems, it is important to know who the detected user is. Each individual has different biometrics like hair colour, face shape, height, weight, body size, etc., that combination of them makes a unique feature allows us to recognize him/her (see http://en.wikipedia.org/wiki/Surveillance). The system and method of the present invention automatically extracts these biometrics for each user for the first time and keep them in a database which can be upgraded. Users are recognized by searching this database and matching the parameters. Then the user can be tracked after recognizing him/her and do this process whenever the system lost the user from tracking. Although in FIG. 1 the functional module "Recognize users" has been placed before the functional module "Detect skeleton joint points", the recognizing of users is preferably done after detecting skeleton joint points, because then biometrics are easier to extract and there is a correlation among body limbs.
Skeleton joint points are extracted applying a superfast (using GPU programming) model-based Random forest approach (see Jamie Shotton, Andrew Fitzgibbon, Mat Cook, Toby Sharp, Mark Finocchio, Richard Moore, Alex Kipman, and Andrew Blake, Real-Time Human Pose Recognition in Parts from a Single Depth Image, in CVPR, IEEE, June 201 1 ). Here, body is divided into parts attached a label to each part, and then a database of pixel labels of very different gestures is composed. Random forest is a learning algorithm and it is taught, according to the present invention, through feature vectors on body pixels and their labels. Later, the algorithm predicts each pixel label of the new detected body using its associated feature vector. Feature vectors are using depth information of other pixels to be computed, particularly using relative depth information between different voxels of the body parts. Body parts are defined such that mean point of each part shows the joint point of it. Working with depth enables us to continue without light. In working with RGB images in presence of light, some works extracted joint points instantly from body bounding box but the accuracy is less than working with depth.
Skeleton joint points are added to the stream and using this continuous stream the status of the user can be analysed, i.e. identified from analysis the continuous stream. Depending on the client, different status can be defined, walking, sitting, standing, laying down, watching TV, sleeping, falling, etc. A database for different styles of each status is made and an algorithm to define the action the user is doing based on the current stream is taught. A solution can be hierarchical aligned cluster analysis (see F. Zhou, F. De la Torre and J. K. Hodgins. Hierarchical Aligned Cluster Analysis for Temporal Clustering of Human Motion. IEEE Transactions Pattern Analysis and Machine Intelligence (PAMI), vol. 35, no. 3, pp. 582-596, 2013).
Each part of the core needs heavy processing. The system can be tuned to reduce processing and bring the system into real time. For instance, using the core for initial recognition and applying tracking in the next frames, or in a single user multi camera system, focusing the system on the camera sees the user and stop processing for other cameras until he is available.
Examples of Use Case Scenarios:
Examples of use case scenarios of the system of the present invention are depicted in Figure 2, which shows several depth cameras C connected to their respective local computing means L in three different situations of monitored users: - Scenario 1 (User walking at home): Users with a certain degree of mobility but who can move freely in their homes, for example a dining room.
- Scenario 2 (User watching TV): Many carers leave home at those moments when the person they are taking care of is performing an activity that does not involve any movement, for example watching TV or also performing resting activities (reading, eating, etc.). In these cases, the user is monitored in a small space like a room or a hall.
- Scenario 3 (User sleeping): Another situation that is considered is to remotely monitor users while they are in bed because of a disease or because they are sleeping.
In these three scenarios, each local computing means L communicate with the remote computing means R, sending to the latter the above described data, including biometric information about the user and, generally, the above described continuous stream and identified user status (although depending on the embodiment, the building of said stream and/or the identification of the user status can be performed whether by the local computing means L or by the remote computing means R). The remote computing means R analyses the received data and is responsible for communicating with the devices of authorized persons, like caregivers to the elderly or relatives, through the network with information indicative of an alarm.
In these scenarios, such authorized persons can perform other activities, since they know that if something happens they will receive a warning, they will see what the user is doing, and they will be able to communicate at a distance with his/her voice with the user.
A person skilled in the art could introduce changes and modifications in the embodiments described without departing from the scope of the invention as it is defined in the attached claims.

Claims

Claims
1 . - A computer-implemented method for remotely monitoring a user, comprising:
- acquiring multi-modal user data related to a monitored user, including at least depth data, using at least a depth camera in a local surveillance area;
- analysing, by processing means, the acquired multi-modal user data to identify a status of the monitored user; and
- presenting said identified status of the monitored user and/or information associated thereto, to a remote computing means;
characterised in that the method comprises:
- automatically extracting biometrics for at least some of a plurality of unidentified users, and storing the extracted biometrics in a biometrics database correlated to identity information of the correspondent user; and
- recognizing the identity of the monitored user by automatically extracting biometrics for the monitored user and searching the biometrics database to look for a match.
2. - The method according to claim 1 , wherein said multi-modal user data comprises at least audio data, RGB video data and infrared depth video data.
3. - The method according to any of the claims 1 or 2, wherein said biometrics are extracted from multi-modal user data including depth video or depth images data acquired using at least a depth camera.
4. - The method according to claim 3, wherein said biometrics are extracted also from audio data and/or RGB video data of said multi-modal user data.
5. - The method according to any of the claims 1 to 4, comprising analysing, by said processing means, the acquired multi-modal user data to perform the next steps: i) detecting the monitored user,
ii) performing said recognizing of the monitored user,
iii) detecting skeleton point joints of the monitored user, and
iv) performing, using said processing means, said identifying of the status of the monitored user analysing at least the detected skeleton point joints of step iii) and also movement and 3D geometry, regarding the monitored user, using at least depth information of said acquired multi-modal user data.
6. - The method according to claim 5, wherein said step iv) is performed by analysing also data from sensors carried by the monitored user, said sensors comprising an accelerometer and/or a gyroscope and/or a temperature sensor and/or a blood pressure sensor and/or an oximeter sensor and/or an ECG sensor assembly and/or an EEG sensor assembly and/or an EMG sensor assembly.
7. - The method according to any of the claims 5 or 6, wherein said processing means are implemented in local computing means comprising and/or connected with said depth camera, all of said steps i) to iv) being performed by said local computing means, the local computing means being bidirectionally connected to said remote computing means and delivering the latter with at least said identified status of the monitored user and/or said information associated thereto.
8. - The method according to any of the claims 5 or 6, wherein a local portion of said processing means are implemented in local computing means and a remote portion of the processing means are implemented in said remote computing means, some of said steps i) to iv) being performed by said local computing means and the rest of the steps i) to iv) being performed by said remote portion of the processing means implemented the remote computing means.
9.- The method according to any of the claims claims 5 or 6, wherein said processing means are implemented in said remote computing means, all of said steps i) to iv) being performed by said remote computing means.
10. - The method according to any of the claims 5 to 9, wherein said detection of step i) is performed, using colour image data and depth information acquired by said depth camera, applying HOG features in RGBD space and learning those features on a predefined database of users and nonusers, and, afterwards, extracting body pixels by depth thresholding and surface normal analysing in an area within a defined bounding box around the monitored user.
1 1 . - The method according to any of the claims 5 to 10, wherein said skeleton joint points are extracted, at step iii), applying a model-based Random forest algorithm, using at least depth information acquired by said depth camera, comprising dividing the depth image data corresponding to the body of the monitored user into body parts, attaching an identification pixel label to each body part, and building a gestures database with at least part of the attached identification pixel labels corresponding to different gestures.
12. - The method according to claim 1 1 , further comprising teaching said model- based random forest algorithm through feature vectors on body pixels of said body parts and their identification pixel labels, and, later, using the model-based random forest algorithm to predict each identification pixel label of a new detected body using its associated feature vector, wherein said feature vectors are build using relative depth information between different voxels of said body parts, such that the feature vectors comprise depth relations between said voxels.
13. - The method according to claim 12, comprising adding the detected skeleton joint points to a continuous stream, including at least a depth video or a sequence of depth images of the acquired multi-modal user data, and performing the identifying of the status of the monitored user at step iv) analysing said continuous stream, regarding the movement and 3D geometry of the monitored user, on temporal sequences of said continuous stream, together with the detected skeleton point joints.
14. - The method according to claim 5 or according to any of claims 6 to 13 when depending on claim 5, comprising performing the biometrics extraction used for the recognizing of step ii) after the skeleton joints points detection of step iii) has been performed, and performing said biometrics extraction by at least analysing the spatial relations between the detected skeleton joint points.
15. - The method of claim 14 when depending on claim 13, comprising performing the biometrics extraction from said continuous stream.
16. - The method according to claim 15, comprising performing the biometrics extraction for the monitored user from an initial portion of the continuous stream and performing the analysis to identify a status of the monitored user on next portions of the continuous stream.
17.- The method according to claim 13, comprising building a status database with information regarding different styles of status of users, and using the information included in said status database to teach an algorithm in charge of performing step iv) to define the action the monitored user is doing based on the continuous stream to identify the status of the monitored user.
18.- The method according to any of the previous claims, wherein said status of the monitored user is a behavioural status.
19.- The method according to any of the previous claims, wherein the behavioural status is at least one of walking, sitting, standing, laying down, watching TV, sleeping and falling.
20.- The method according to any of the previous claims, comprising using a single user multi camera system comprising a plurality of depth cameras aiming to different local surveillance areas through which the user can circulate, and performing the acquiring of the multi-modal user data by using only the depth camera aimed to the area where the user is placed.
21 .- The method according to any of the previous claims, comprising performing in real time at least the identification of the status of the monitored user.
22. - The method according to any of the previous claims, wherein said information associated to the identified status of the monitored user presented to the remote computing means is indicative of an alarm.
23. - A system for remotely monitoring a user, comprising:
- local acquiring means for acquiring multi-modal user data related to a monitored user, said local acquiring means comprising at least one depth camera (C) to acquire the multi-modal user data in a local surveillance area;
- processing means for analysing, by means of data processing, the acquired multi-modal user data to identify a status of the monitored user;
- remote computing means (R); and
- means for presenting said identified status of the monitored user and/or information associated thereto, to said remote computing means (R);
characterised in that the system is a multi-user system and comprises:
- means for automatically extracting biometrics for at least some of a plurality of unidentified users and storing the extracted biometrics in a biometrics database correlated to identity information of the correspondent user, said biometrics database being included in the system, and
- means for recognizing the identity of the monitored user by automatically extracting biometrics for the monitored user and searching the biometrics database to look for a match.
24. - The system according to claim 23, adapted for performing the method of any of claims 1 to 22, where all the steps of the method, except for the step of acquiring multi-modal user data, are performed by data processing by means of said processing means, the processing means being implemented in the remote computing means (R) and/or in local computing means (L) included in the system.
25. - The system according to claim 24, wherein said remote computing means (R) are adapted for establishing bidirectional communication with the local computing means (L), to allow, to an authorized person, to control, configure and upgrade operation settings of the local computing means (L).
26.- The system according to claim 25, wherein said remote computing means
(R) are implemented in a portable computing device.
27.- A computer program product, which includes code instructions that when executed in a computer implement all the steps of the method according to any of the claims 1 to 22, except for the step of acquiring multi-modal user data.
PCT/EP2015/067324 2014-07-30 2015-07-29 A computer-implemented method and a system for remotely monitoring a user, and a computer program product implementing the method WO2016016277A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP14382294 2014-07-30
EP14382294.8 2014-07-30

Publications (1)

Publication Number Publication Date
WO2016016277A1 true WO2016016277A1 (en) 2016-02-04

Family

ID=51266257

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2015/067324 WO2016016277A1 (en) 2014-07-30 2015-07-29 A computer-implemented method and a system for remotely monitoring a user, and a computer program product implementing the method

Country Status (1)

Country Link
WO (1) WO2016016277A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220076365A1 (en) * 2019-11-26 2022-03-10 Ncr Corporation Frictionless security monitoring and management

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120075464A1 (en) * 2010-09-23 2012-03-29 Stryker Corporation Video monitoring system
EP2560141A1 (en) 2011-08-19 2013-02-20 Accenture Global Services Limited Interactive virtual care
US20140052464A1 (en) 2012-08-16 2014-02-20 Abhijit Ray Method and system for remote patient monitoring

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120075464A1 (en) * 2010-09-23 2012-03-29 Stryker Corporation Video monitoring system
EP2560141A1 (en) 2011-08-19 2013-02-20 Accenture Global Services Limited Interactive virtual care
US20140052464A1 (en) 2012-08-16 2014-02-20 Abhijit Ray Method and system for remote patient monitoring

Non-Patent Citations (10)

* Cited by examiner, † Cited by third party
Title
F. ZHOU; F. DE LA TORRE; J. K. HODGINS: "Hierarchical Aligned Cluster Analysis for Temporal Clustering of Human Motion", IEEE TRANSACTIONS PATTERN ANALYSIS AND MACHINE INTELLIGENCE (PAMI, vol. 35, no. 3, 2013, pages 582 - 596, XP011494042, DOI: doi:10.1109/TPAMI.2012.137
FENG ZHOU ET AL: "Hierarchical Aligned Cluster Analysis for Temporal Clustering of Human Motion", IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, IEEE COMPUTER SOCIETY, USA, vol. 35, no. 3, 1 March 2013 (2013-03-01), pages 582 - 596, XP011494042, ISSN: 0162-8828, DOI: 10.1109/TPAMI.2012.137 *
JAMIE SHOTTON ET AL: "Real-time human pose recognition in parts from single depth images", COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2011 IEEE CONFERENCE ON, IEEE, 20 June 2011 (2011-06-20), pages 1297 - 1304, XP032037818, ISBN: 978-1-4577-0394-2, DOI: 10.1109/CVPR.2011.5995316 *
JAMIE SHOTTON; ANDREW FITZGIBBON; MAT COOK; TOBY SHARP; MARK FINOCCHIO; RICHARD MOORE; ALEX KIPMAN; ANDREW BLAKE: "Real-Time Human Pose Recognition in Parts from a Single Depth Image", CVPR, IEEE, June 2011 (2011-06-01)
LYUDMILA MIHAYLOVA; PAUL BRASNETT; NISHAN CANAGARAJAN; DAVID BULL: "NATO Security Through Science", 2007, IOS PRESS, article "Object Tracking by Particle Filtering Techniques in Video Sequences; In: Advances and Challenges in Multisensor Data and Information", pages: 260 - 268
MATTHEW FITZPATRICK ET AL: "WORCESTER POLYTECHNIC INSTITUTE Real Time Person Tracking and Identification using the Kinect sensor Major Qualifying Project in Electrical & Computer Engineering", 25 April 2013 (2013-04-25), XP055218198, Retrieved from the Internet <URL:https://www.wpi.edu/Pubs/E-project/Available/E-project-042513-081838/unrestricted/Real_Time_Person_Tracking_and_Identification_using_the_Kinect_sensor.pdf> [retrieved on 20151005] *
T D RÄTY: "Survey on Contemporary Remote Surveillance Systems for Public Safety", IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS: PART C:APPLICATIONS AND REVIEWS., vol. 40, no. 5, 1 September 2010 (2010-09-01), US, pages 493 - 515, XP055219318, ISSN: 1094-6977, DOI: 10.1109/TSMCC.2010.2042446 *
T. BOUWMANS; F. PORIKLI; B. HORFERLIN; A. VACAVANT: "Background Modeling and Foreground Detection for Video Surveillance: Traditional and Recent Approaches, Implementations, Benchmarking and Evaluation", June 2014, CRC PRESS, TAYLOR AND FRANCIS GROUP
V. PRISACARIU; I. REID: "fastHOG - a real-time GPU implementation of HOG", TECHNICAL REPORT, 2012
V. PRISACARIU; I. REID: "fastHOG - a real-time GPU implementation of HOG", TECHNICAL REPORT, 4 July 2012 (2012-07-04), XP055219207 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220076365A1 (en) * 2019-11-26 2022-03-10 Ncr Corporation Frictionless security monitoring and management
US11727520B2 (en) * 2019-11-26 2023-08-15 Ncr Corporation Frictionless security monitoring and management

Similar Documents

Publication Publication Date Title
Mastorakis et al. Fall detection system using Kinect’s infrared sensor
Ann et al. Human activity recognition: A review
Auvinet et al. Fall detection with multiple cameras: An occlusion-resistant method based on 3-d silhouette vertical distribution
Foroughi et al. An eigenspace-based approach for human fall detection using integrated time motion image and neural network
US11282367B1 (en) System and methods for safety, security, and well-being of individuals
US20150302310A1 (en) Methods for data collection and analysis for event detection
US20170206664A1 (en) Method for identifying, tracking persons and objects of interest
JP7346019B2 (en) Systems and methods in object history association
Nar et al. Abnormal activity detection for bank ATM surveillance
KR102397248B1 (en) Image analysis-based patient motion monitoring system and method for providing the same
Joshi et al. A fall detection and alert system for an elderly using computer vision and Internet of Things
Ghose et al. Unobtrusive indoor surveillance of patients at home using multiple kinect sensors
Bhattacharya et al. Arrays of single pixel time-of-flight sensors for privacy preserving tracking and coarse pose estimation
Bouachir et al. Automated video surveillance for preventing suicide attempts
WO2020144835A1 (en) Information processing device and information processing method
CN114972727A (en) System and method for multi-modal neural symbol scene understanding
JP2020194493A (en) Monitoring system for nursing-care apparatus or hospital and monitoring method
KR102234995B1 (en) Method, device and system for performing rehabilitation training of cognitive function using virtual object model
Ezatzadeh et al. Fall detection for elderly in assisted environments: Video surveillance systems and challenges
Yoon et al. Tracking System for mobile user Based on CCTV
WO2016016277A1 (en) A computer-implemented method and a system for remotely monitoring a user, and a computer program product implementing the method
US20230326318A1 (en) Environment sensing for care systems
Oumaima et al. Vision-based fall detection and prevention for the elderly people: A review & ongoing research
Avgerinakis et al. Activities of daily living recognition using optimal trajectories from motion boundaries
Youm et al. Development of a methodology to predict and monitor emergency situations of the elderly based on object detection

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15750270

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 15750270

Country of ref document: EP

Kind code of ref document: A1