WO2016150750A1 - Restraint management - Google Patents

Restraint management Download PDF

Info

Publication number
WO2016150750A1
WO2016150750A1 PCT/EP2016/055483 EP2016055483W WO2016150750A1 WO 2016150750 A1 WO2016150750 A1 WO 2016150750A1 EP 2016055483 W EP2016055483 W EP 2016055483W WO 2016150750 A1 WO2016150750 A1 WO 2016150750A1
Authority
WO
WIPO (PCT)
Prior art keywords
restraint
subject
parameter
status
management apparatus
Prior art date
Application number
PCT/EP2016/055483
Other languages
French (fr)
Inventor
Teun Van Den Heuvel
Yingrong Xie
Adrienne Heinrich
Thomas Falck
Esther Marjan VAN DER HEIDE
Harald Greiner
Original Assignee
Koninklijke Philips N.V.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Koninklijke Philips N.V. filed Critical Koninklijke Philips N.V.
Priority to EP16711556.7A priority Critical patent/EP3274884B1/en
Priority to JP2017549428A priority patent/JP6730304B2/en
Priority to CN201680017575.1A priority patent/CN107408159A/en
Priority to US15/552,906 priority patent/US10945667B2/en
Publication of WO2016150750A1 publication Critical patent/WO2016150750A1/en

Links

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/68Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
    • A61B5/6801Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
    • A61B5/683Means for maintaining contact with the body
    • A61B5/6831Straps, bands or harnesses
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/103Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
    • A61B5/11Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
    • A61B5/1113Local tracking of patients, e.g. in a hospital or private home
    • A61B5/1114Tracking parts of the body
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/103Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
    • A61B5/11Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
    • A61B5/1116Determining posture transitions
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/103Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
    • A61B5/11Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
    • A61B5/1126Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb using a particular sensing technique
    • A61B5/1128Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb using a particular sensing technique using image analysis
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/48Other medical applications
    • A61B5/4806Sleep evaluation
    • A61B5/4809Sleep detection, i.e. determining whether a subject is asleep or not
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/48Other medical applications
    • A61B5/4848Monitoring or testing the effects of treatment, e.g. of medication
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/74Details of notification to user or communication with user or patient ; user input means
    • A61B5/746Alarms related to a physiological condition, e.g. details of setting alarm thresholds or avoiding false alarms
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61GTRANSPORT, PERSONAL CONVEYANCES, OR ACCOMMODATION SPECIALLY ADAPTED FOR PATIENTS OR DISABLED PERSONS; OPERATING TABLES OR CHAIRS; CHAIRS FOR DENTISTRY; FUNERAL DEVICES
    • A61G7/00Beds specially adapted for nursing; Devices for lifting patients or disabled persons
    • A61G7/05Parts, details or accessories of beds
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61GTRANSPORT, PERSONAL CONVEYANCES, OR ACCOMMODATION SPECIALLY ADAPTED FOR PATIENTS OR DISABLED PERSONS; OPERATING TABLES OR CHAIRS; CHAIRS FOR DENTISTRY; FUNERAL DEVICES
    • A61G7/00Beds specially adapted for nursing; Devices for lifting patients or disabled persons
    • A61G7/05Parts, details or accessories of beds
    • A61G7/0507Side-rails
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/04Inference or reasoning models
    • G06N5/046Forward inferencing; Production systems
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H40/00ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices
    • G16H40/60ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B2505/00Evaluating, monitoring or diagnosing in the context of a particular type of medical care
    • A61B2505/03Intensive care
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B2505/00Evaluating, monitoring or diagnosing in the context of a particular type of medical care
    • A61B2505/07Home care
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B2562/00Details of sensors; Constructional details of sensor housings or probes; Accessories for sensors
    • A61B2562/02Details of sensors specially adapted for in-vivo measurements
    • A61B2562/0204Acoustic sensors
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B2562/00Details of sensors; Constructional details of sensor housings or probes; Accessories for sensors
    • A61B2562/02Details of sensors specially adapted for in-vivo measurements
    • A61B2562/0219Inertial sensors, e.g. accelerometers, gyroscopes, tilt switches
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B2562/00Details of sensors; Constructional details of sensor housings or probes; Accessories for sensors
    • A61B2562/02Details of sensors specially adapted for in-vivo measurements
    • A61B2562/0247Pressure sensors
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B2562/00Details of sensors; Constructional details of sensor housings or probes; Accessories for sensors
    • A61B2562/02Details of sensors specially adapted for in-vivo measurements
    • A61B2562/0261Strain gauges
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/02Detecting, measuring or recording pulse, heart rate, blood pressure or blood flow; Combined pulse/heart-rate/blood pressure determination; Evaluating a cardiovascular condition not otherwise provided for, e.g. using combinations of techniques provided for in this group with electrocardiography or electroauscultation; Heart catheters for measuring blood pressure
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/08Detecting, measuring or recording devices for evaluating the respiratory organs
    • A61B5/0816Measuring devices for examining respiratory frequency
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/16Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
    • A61B5/165Evaluating the state of mind, e.g. depression, anxiety
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61GTRANSPORT, PERSONAL CONVEYANCES, OR ACCOMMODATION SPECIALLY ADAPTED FOR PATIENTS OR DISABLED PERSONS; OPERATING TABLES OR CHAIRS; CHAIRS FOR DENTISTRY; FUNERAL DEVICES
    • A61G2203/00General characteristics of devices
    • A61G2203/30General characteristics of devices characterised by sensor means
    • A61G2203/34General characteristics of devices characterised by sensor means for pressure
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H10/00ICT specially adapted for the handling or processing of patient-related medical or healthcare data
    • G16H10/60ICT specially adapted for the handling or processing of patient-related medical or healthcare data for patient-specific data, e.g. for electronic patient records
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/20ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems

Definitions

  • the invention relates to a restraint management apparatus and method for use in restraining a subject, and in particular relates to a restraint management apparatus and method for use in conjunction with a medical restraint system.
  • a restraint management apparatus comprises a processing unit arranged to: receive one or more types of sensor data; determine a status of a subject based on the received sensor data; determine, based on the determined subject status, a restraint parameter for a restraint device configured to restrain a body part of the subject; and output a signal based on the determined restraint parameter.
  • embodiments of the invention advantageously provide an automatic way to predict when a given subject should be restrained, released, or allowed more or less freedom of movement, based on the detected condition of the subject. This creates the possibility of automatically updating restraint parameters in real time, in response to changes in a subject's condition.
  • an optimum balance between the clinical need to ensure the safety of the subject and the subject's need to be comfortable and free to move can be achieved.
  • the processing unit is further arranged to receive one or more types of non-sensor data relating to the subject. In some such embodiments the processing unit is arranged to determine a status of a subject based additionally on the received non-sensor data. In some embodiments the one or more types of non-sensor data comprises data originating from a caregiver. In some such embodiments the one or more types of non-sensor data comprises data manually input by a caregiver to the restraint management apparatus or to a device in communication with the restraint management apparatus.
  • the processing unit comprises a memory containing one or more predefined signatures, each predefined signature relating to a particular type of feature. In some such embodiments the processing unit is arranged to determine a status of a subject by comparing the received data to the one or more predefined signatures to detect one or more features in the received sensor data.
  • the processing unit is arranged to determine a status of a subject by determining one or more of: a physiological condition of the subject, a position of the subject; a position of a body part of the subject; a movement level of a body part of the subject; an activity level of the subject; whether the subject is asleep or awake; whether the subject's eyes are open; whether the subject is delirious; whether movement of a body part of the subject meets a predefined criterion; whether movement of the subject meets a predefined criterion.
  • the processing unit is arranged to determine a status of a subject such that the determined status comprises a status value for one or more factors.
  • status value comprises one or more of: a descriptive indication; a numerical score; a non-numerical level indication.
  • the processing unit comprises a memory containing rules relating subject status values to restraint parameters. In some such embodiments the processing unit is arranged to determine a restraint parameter by applying one or more of the rules to the determined subject status value. In some such embodiments the processing unit comprises a machine learning module and the rules have been generated by the machine learning module based on historical data relating to the subject. In some embodiments the processing unit is arranged to receive a restraint parameter which has been manually- input to a restraint device, and the machine learning module is arranged to update the generated rules based on the received manually- input restraint parameter.
  • the processing unit is arranged to determine a restraint parameter such that the determined restraint parameter comprises one of: activation of a restraint device; inactivation of a restraint device; activation of a given restraining component of a restraint device; inactivation of a given restraining component of a restraint device; tightness of a given restraining component of a restraint device; length of a given restraining component of a restraint device; duration of activation of a restraint device; duration of activation of a given restraining component of a restraint device.
  • the processing unit is arranged to output a signal comprising one or more of: a control signal to an automated restraint device arranged to cause the automated restraint device to apply the determined restraint parameter; a signal to a device associated with a caregiver, the signal comprising an instruction to the caregiver to implement the determined restraint parameter; a control signal to an alarm module of the restraint management apparatus arranged to cause the alarm module to generate an alarm in dependence on the determined restraint parameter; a control signal to a display module of the restraint management apparatus arranged to cause the display module to display information in dependence on the determined restraint parameter; a control signal to a subject feedback module arranged to cause the subject feedback module to generate a message to the subject; a control signal to a lighting module arranged to cause the lighting module to activate a light.
  • the processing unit is arranged to output the signal after a predefined amount of time has passed since the processing unit last output a signal of the same type as the signal.
  • the invention also provides, according to a second aspect, a restraint system.
  • the restraint system comprises a restraint management apparatus according to the first aspect; and a restraint device configured to restrain a body part of a subject.
  • the restraint device is arranged to: receive a signal output by the restraint management apparatus, and in response to a signal received from the restraint management apparatus, apply a restraint to the subject or alter a parameter of a restraint applied to the subject.
  • the restraint device is further arranged to: receive a manually- input restraint parameter; in response in response to a signal received from the restraint management apparatus, apply a restraint to the subject or alter a parameter of a restraint applied to the subject; and transmit a signal containing the manually- input restraint parameter to the restraint management apparatus.
  • the invention also provides, according to a third aspect, a method for use in restraining a subject.
  • the method comprises: receiving one or more types of sensor data;
  • the received sensor data comprises one or more of: an image including a subject; an image including a subject and one or more objects in the vicinity of the subject; audio data including sounds generated by a subject; accelerometer data obtained by an accelerometer worn by a subject; measurements of one or more physiological characteristics of a subject; restraint-related data obtained by a sensor associated with a restraint device; pressure data obtained by a bed sensor; bed rail position data obtained by a bed rail sensor.
  • the received sensor data comprises an image including a subject and one or more objects.
  • determining a status of a subject comprises: determining a position of a body part of the subject; determining a position of an object in the image; calculating a distance between the body part and the object; and comparing the calculated distance to a predetermined threshold.
  • Figure 1 is an illustration of a restraint management apparatus according to a general embodiment of the invention
  • Figure 2 is a flow chart illustrating a method for use in restraining a subject according to a general embodiment of the invention.
  • Figure 3 is an illustration of an example restraint device for use with a restraint management apparatus according to the invention.
  • FIG. 1 shows a restraint management apparatus 10 according to an embodiment of the invention.
  • the restraint management apparatus 10 is for use with a restraint device 14 configured to restrain a body part of a subject.
  • the restraint device 14 is configured to provide a controllable resistance to movement of a body part.
  • the restraint management apparatus 10 comprises a processing unit.
  • the restraint management apparatus also comprises one or more functional modules, e.g. an alarm module, a display module, a communications module, a user input module, a subject feedback module to present messages to the subject (e.g. audio or text messages), a lighting module to illuminate the subject and/or their surroundings, etc.
  • the processing unit of the restraint management apparatus 10 is arranged to receive one or more types of sensor data 16.
  • the processing unit is arranged to receive sensor data 16 from one or more sensor devices 18.
  • Each sensor device 18 can be, e.g., a camera, microphone, accelerometer, medical monitoring device, or any other type of sensor device arranged to collect sensor data relating to a subject.
  • the processing unit is arranged to receive sensor data from the restraint device 14.
  • the processing unit is arranged to receive sensor data from a user input module of the restraint management apparatus (e.g. where a caregiver has manually input readings obtained by a sensor), and/or a communications module of the restraint management apparatus (e.g.
  • sensor data is sent to the restraint management apparatus from a remote electronic device, such as a laptop or smartphone of a caregiver).
  • the processing unit is arranged to receive sensor data in real time.
  • the processing unit is arranged to receive sensor data at periodic intervals, e.g. in response to a request sent from the processing unit to a sensor device 18 or the restraint device 14.
  • the processing unit is further arranged to receive non- sensor data, which may or may not be related to the subject.
  • Subject-related non-sensor data can comprise, for example, historical subject information, medical test results, diagnoses, etc.
  • Non-subject related non-sensor data can comprise, for example, software updates, information relating to a restraint device, or any other type of information/data that may be used by the restraint management apparatus.
  • the non-sensor data comprises a manually- input restraint parameter which has been input to a restraint device by a caregiver. It will be appreciated that manually- input data comprises any data received by a device which originates from a caregiver or other human.
  • Manually-input data can, but need not, be input via an input interface of the device receiving the data.
  • Manually-input data can also be input via a remote device in communication with the device receiving the data.
  • Non- sensor data may be received in any of the manners described above in relation to the sensor data.
  • the processing unit is further arranged to determine a status of a subject based on the received sensor data (and, in some embodiments, received non-sensor data).
  • the determined status should be related to the restraint requirements of the subject, and will generally therefore be indicative of a current or expected amount/type of activity/movement of the subject.
  • the processing unit is arranged to determine a status of a subject by determining one or more of: a physiological condition of the subject, a position of the subject; a position of a body part of the subject; a movement level of a body part of the subject; an activity level of the subject; whether the subject is asleep or awake; whether the subject's eyes are open; whether the subject is delirious; whether movement of a body part of the subject meets a predefined criterion (e.g. defined such that only unusual movement will meet the criterion); whether movement of the subject meets a predefined criterion (e.g.
  • Various techniques for extracting physiological, position and/or activity information from sensor data are known in the art. For example, data from an accelerometer mounted on the subject, and/or from a pressure sensor mounted on their hospital bed can be analyzed to determine the position of a subject and/or whether and to what extent the subject is moving. Subject movement can also be determined from image data acquired by a camera, using video analysis techniques. Image data can be used to determine the proximity of a body part of a subject to a medical device or other object (e.g.
  • Audio data collected by a microphone can detect vocalizations, movement noises and/or impact noises generated by the patient, which can be used to determine whether they are in distress, likely to attempt to get out of bed, and/or are at risk of harming them self.
  • Physiological data e.g. heart rate, respiration rate, breathing pattern, ECG data
  • Physiological data can be used in the determination of the status of a subject, e.g. by indicating whether the subject is asleep or awake, indicating which sleep stage the subject is in, and/or indicating whether they are experiencing increased circulatory effort (which can result from increased activity).
  • the processing unit is arranged to determine the status of a subject by detecting patterns and/or features in the received sensor data. For example, in some embodiments signatures and/or patterns associated with features of interest (such as a particular body part in an image of the subject; a particular object in an image of the subject; a vocalization in an audio recording of the subject; an impact sound in an audio recording of the subject; a peak in accelerometer data, etc.) are stored in a memory associated with the processing unit, and the processing unit is arranged to determine the status of a subject by comparing detected patterns/features with the stored signatures/patterns. In some embodiments signatures and/or patterns associated with features of interest (such as a particular body part in an image of the subject; a particular object in an image of the subject; a vocalization in an audio recording of the subject; an impact sound in an audio recording of the subject; a peak in accelerometer data, etc.) are stored in a memory associated with the processing unit, and the processing unit is arranged to determine the status of a subject by
  • the determination is performed based on received sensor data from a particular instant in time (i.e. the determination uses a static analysis). In some embodiments the determination is performed using received sensor data covering a time period (i.e. the determination uses a dynamic analysis). In some embodiments determining the status of a subject comprises tracking a detected feature over a time period. In some embodiments determining the status of a subject comprises detecting movement of a body part of the subject. In some such embodiments, detected movements are classified as being one or more of several different predefined movement types. Any suitable known signal analysis and feature extraction techniques may be used in the performance of step 202.
  • the processing unit is arranged to determine the status of the subject by detecting, using image analysis techniques, the relative positions of body parts of the subjects and objects.
  • the detected body parts can comprise, for example, one or more of a hand, a foot, the head, etc.
  • the detected objects can comprise, for example, one or more of a patient monitor, an IV stand, a bed rail, a bedside table, etc.
  • the processing unit is arranged to calculate distances between the detected body parts and the detected objects.
  • the processing unit is arranged to compare the calculated distances to a predefined threshold. The subject status may then be determined in dependence on the result(s) of the comparing.
  • a status indicating that restraint of a body part is required will be determined only if the distance between that body part and an object is less than the predefined threshold. In some embodiments a status indicating that restraint of multiple or all body parts is required will be determined if the distance between at least one body part and an object is less than the predefined threshold.
  • the processing unit is arranged to calculate the velocity of a detected body part towards a detected object (for example from a series of measurements of the relative positions of body parts and objects obtained from a time series of images). In some such embodiments the processing unit is arranged to compare the calculated velocity to a predefined velocity threshold. The subject status may then be determined in dependence on the result(s) of the comparing.
  • a status indicating that restraint of a body part is required will be determined only if the velocity of the body part towards the object is more than the predefined velocity threshold. In some embodiments a status indicating that restraint of multiple or all body parts is required will be determined if the velocity of at least one body part towards an object is more than the predefined velocity threshold.
  • An exemplary velocity threshold could be 2m/s.
  • the processing unit is arranged to receive sensor data in real time
  • the processing unit is arranged to determine a status of a subject in real time or near real time.
  • the determined subject status comprises a status value in respect of each of multiple factors, where a factor could be, e.g., activity level, medical condition, sleep/wake state, lucidity, etc.
  • the determined subject status is indicated, e.g. by the restraint management apparatus or a device in communication with the restraint management apparatus.
  • indicating the determined subject status comprises displaying a message or visual status indication.
  • indicating the determined subject status comprises emitting light of a particular color.
  • the processing unit is further arranged to determine, based on the determined subject status, a restraint parameter for a restraint device configured to restrain a body part of the subject, e.g. the restraint device 14.
  • the determined restraint parameter can comprise, for example, one or more of: activation/inactivation of the restraint device (i.e. as a whole), activation/inactivation of a given restraining component of the restraint device,
  • the processing unit is provided with a set of rules/relationships (e.g. stored in a memory of the restraint
  • the restraint parameter can indicate a level of resistance to movement to be provided by the restraint device.
  • the processing unit is arranged to determine a restraint parameter by applying at least one predefined criterion to the determined subject status.
  • the predefined criterion comprises a set of conditions.
  • the set of conditions comprises a time-based condition and a value-based condition.
  • a value-based condition is defined such that whether or not a status value meets the condition depends on the value itself.
  • a time-based condition specifies a condition (e.g. a minimum) relating to the amount of time for which the status information has consistently met a value-based condition.
  • the set of conditions comprises a plurality of value-based conditions.
  • the processing unit is arranged to determine a restraint parameter using a function of the distance from a body part to an object in the environment, and/or of the velocity of a body part towards an object in the environment.
  • An exemplary function based on distance x is set out below:
  • F r F max + (BT - x)/(RT - BT) ⁇ (F max - F min )
  • BT is a blocking threshold distance at which the movement of the body part is to be blocked (i.e. a maximum restraining force, F max , is to be applied)
  • RT is a restraining threshold distance beyond which a minimum restraining force (F m i n ) is to be applied.
  • F max a blocking threshold distance at which the movement of the body part is to be blocked (i.e. a maximum restraining force, F max , is to be applied)
  • RT is a restraining threshold distance beyond which a minimum restraining force (F m i n ) is to be applied.
  • F max can be 1000 Newton (N), although any force that would prevent movement of the body part can be used.
  • An exemplary value for F m i n can be 0.1 N, although any force that would allow free or reasonably free movement of the body part can be used.
  • An exemplary value for BT can be 10 centimeters, and an exemplary value for RT can be 25 cm.
  • the restraining force provided by the above function can be related to a restraint parameter. It will be appreciated that the above function is merely exemplary, and those skilled in the art will be aware of other types or forms of functions that can be used to relate distance and/or velocity to a restraining force.
  • the at least one predefined criterion may be factor-specific.
  • the determined subject status comprises a status value for the factor "sleep/wake state” and a status value for the factor "activity level”. In some such
  • the processing unit is arranged to determine a restraint parameter based on both whether the sleep/wake state value is equal to a predefined value and whether the activity level value exceeds a predefined threshold. For example, in some such embodiments if the sleep/wake status value is "asleep" and the activity level is “medium” then the processing unit will determine that the restraining device should be used, but if the sleep/wake status value is "awake” and the activity level is "medium” then the processing unit will determine that the restraining device should not be used.
  • the processing unit includes a machine learning module, configured to use standard machine learning techniques to identify or generate rules, relationships, etc. relating restraint parameters to subject status.
  • the machine learning module is configured to apply machine learning techniques to historical data, e.g. historical subject status data, historical physiological measurement data, and/or historical restraint information (e.g. information about restraint parameters used and/or their appropriateness).
  • the machine learning module is configured to generate rules/relationships relating restraint parameters to subject status based on manually- input restraint parameters received from the restraint apparatus.
  • the machine learning module is arranged to update existing rules/relationships based on manually- input restraint parameters received from the restraint apparatus.
  • the processing unit 12 is arranged to determine a restraint parameter based on rules or relationships generated or identified by the machine learning module.
  • the processing unit is arranged to determine a restraint parameter in real time or near real time.
  • this feature can prevent inappropriate restraint of a subject, and minimize the amount of time for which a subject is restrained. Conversely, it can ensure that restraint is applied rapidly when required, reducing the risk of harm to the subject.
  • the processing unit is further arranged to output a signal 12 based on the determined restraint parameter.
  • the signal 12 comprises a control signal, and is output to the restraint device 14, e.g. by a communications module of the restraint management apparatus 10.
  • the signal 12 causes the restraint device to apply the determined restraint parameter.
  • the output signal 12 comprises a control signal which causes the restraint device 14 to set the length of the given restraint component to be the specified length.
  • the output signal 12 comprises a control signal which causes the restraint device 14 to set the resistance to movement to the specified level of resistance.
  • the signal 12 is arranged to cause the restraint device 14 to display the determined restraint parameter, e.g. on a display associated with the restraint device 14.
  • the signal comprises an instruction to a caregiver to implement the determined restraint parameter.
  • the signal comprises a control signal arranged to cause a message to be presented to the subject.
  • a message could provide the subject with instructions and/or reassurance.
  • a suitable message could say, for example, "Please calm down. A nurse will be with you shortly".
  • the message to the subject comprises an audio message, e.g. generated by a loudspeaker comprised in or in
  • the message to the subject comprises a visual message, e.g. displayed by a display comprised in or in communication with the restraint management device 10.
  • the processing unit is arranged to output the signal 12 after a predefined amount of time has passed since the processing unit last output a signal of the same type as the current signal 12 (e.g. in some embodiments an instruction or control signal arranged to cause adjustment of a particular restraint would comprise a first type of signal, and an instruction or control signal arranged to cause adjustment of a different restraint, a control signal arranged to generate an alarm, a warning message to a caregiver, etc. would each be considered to be a signal of a different type to the first signal). This can reduce the stress experienced by the subject, since if restraint parameters are changed too often this can confuse or surprise the subject.
  • the signal 12 comprises a control signal and is sent to a functional module of the restraint management apparatus 10.
  • the control signal may be arranged, for example, to cause one or more of:
  • an alarm module to generate an alarm
  • a subject feedback module to generate a message to the subject
  • a communications module to generate a message to a caregiver
  • a communications module to send a signal to a remote device
  • a communications module to send a signal to the restraint management apparatus
  • a display module to display information
  • a lighting module to activate a light
  • the processing unit is arranged to output a plurality of signals, for example a control signal to the restraint device 14 and a control signal to an alarm module of the restraint management apparatus 10 causing it to generate an alarm.
  • the processing unit is arranged to output a signal 12 continuously, in real-time or near real-time. In some embodiments the processing unit is arranged to output a signal in response to a change in the determined parameter. In some embodiments the processing unit is arranged to output a signal 12 at periodic intervals.
  • the restraint device 14 can be any device arranged to restrain (i.e. to restrict the movement of) a body part of a subject.
  • the restraint device 14 comprises multiple restraint components, each of which is arranged to restrain a particular body part.
  • the restraint components comprise one or more straps (e.g. wrist straps, ankle straps, a waist strap, etc.).
  • the restraint components comprise one or more tethers.
  • the restraint device 14 comprises one or more bed rails.
  • the restraint device 14 comprises a bed sheet or cover which can be adjustably tightened over the subject.
  • the restraint device is attached to or integrated in a hospital bed.
  • the restraint device 14 is arranged such that one or more restraint parameters (e.g. strap tightness/length, tether length, bed rail height, resistance to movement) are adjustable during use of the restraint device 14.
  • the restraint device is arranged to be as user friendly as possible.
  • straps can comprise a soft and/or padded material. Straps and/or tethers may be arranged to vary the resistance to movement in dependence on the
  • the restraint device 14 includes at least one sensor for acquiring sensor data. In some embodiments the restraint device 14 includes at least one sensor associated with each restraint component of the restraint device 14. The at least one sensor can comprise, for example, a strain gauge; an accelerometer; a position sensor. In some embodiments the restraint device 14 includes a sensor arranged to detect a parameter of the subject (e.g. position of a given body part, movement of a given body part, etc.). In some embodiments the restraint device 14 includes a sensor arranged to detect a parameter of the restraint device (e.g.
  • the restraint device 14 is arranged to output sensor data, e.g. to the restraint management apparatus 10 or to another remote device (such as a patient monitoring device or a hospital computer system).
  • the restraint device 14 can be an automated restraint device or a manually operated restraint device.
  • a manually operated restraint device requires a human operator to adjust its restraint parameters.
  • the restraint management device 10 is used in conjunction with a manually operated restraint device, the restraint management device is arranged to alert a caregiver to a change in the determined restraint parameter, so that the caregiver can implement the newly determined parameter.
  • caregiver intervention to adjust the restraint device is still required, in such situations the restraint management system significantly reduces the burden on caregivers by reducing or removing the need for a caregiver to continuously monitor the status of subject.
  • An automated restraint device is able to adjust one or more of its restraint parameters without input from a human operator, e.g. in response to a control signal received from the restraint management apparatus 10.
  • the use of an automated restraint device in conjunction with a restraint management apparatus according to embodiments of the invention can enable the rapid (preferably real-time) adjustment of the nature and degree of restraint applied to a given subject, based on their current status. This can significantly reduce or even eliminate periods of inappropriate restraint.
  • an automated restraint device is arranged such that restraint parameters can also be adjusted manually, e.g. via a user interface of the restraint device.
  • an automated restraint device is arranged such that restraint parameters can also be adjusted manually, e.g. via a user interface of the restraint device.
  • an automated restraint device is arranged such that restraint parameters can also be adjusted manually, e.g. via a user interface of the restraint device.
  • an automated restraint device is
  • automated restraint device is arranged such that a caregiver is required to approve/confirm a restraint parameter generated by the restraint management apparatus before the automated restraint device will apply that parameter.
  • the approval can be done remotely, via a device (such as a nurses' station, or portable electronic device of a caregiver) which is in communication with the restraint management apparatus and the restraint device.
  • Figure 2 shows an example of a method for use in restraining a subject, which can be implemented by the restraint management apparatus 10.
  • sensor data 16 is received, e.g. from one or more sensor devices, from a restraint device, from a user input module of the restraint management apparatus 10, from a remote electronic device, etc.
  • the sensor data is received as a continuous stream.
  • the sensor data is received periodically. In some embodiments the sensor data is received in response to a request previously sent to a sensor device/restraint device. In some embodiments the sensor data has been time-stamped at the time of its generation. In some embodiments the sensor data is time-stamped at the time of its receipt, e.g. by the processing unit of the restraint management apparatus 10. In some embodiments the received sensor data directly relates to a subject, in that it comprises measurements of a physiological characteristic of that subject. In some embodiments the received sensor data indirectly relates to a subject. For example, data which indirectly relates to a subject could comprise an image or audio recording of the subject's hospital room which includes the subject as well as their surroundings. In some embodiments the received sensor data is stored, e.g. in a memory of the restraint management apparatus 10.
  • an optional step 202 of receiving non-sensor data relating to the subject e.g. from a user input module of the restraint management apparatus, a remote electronic device, etc. is performed.
  • the subject-related non-sensor data can comprise, e.g. test results; a medical diagnosis; historical physiological characteristic data; clinical outcome data; treatment data, restraint data, etc.
  • the received subject- related non-sensor data is stored, e.g. in a memory of the restraint management apparatus 10.
  • a status of a subject i.e. the subject to which the received sensor data either directly or indirectly relates
  • is determined based on the received sensor data e.g. by the processing unit of the restraint management apparatus 10.
  • a status of the subject is determined based additionally on the non-sensor data.
  • the received sensor data and the received non-sensor data are combined by a data fusion module of the processing unit of the restraint management apparatus 10.
  • the combining comprises assigning one or more weights to the received sensor data and the received non-sensor data. For example, non-sensor data in the form of data input manually by a caregiver may be assigned a higher weight than sensor data.
  • the non-sensor data can comprise a subject status manually input by a caregiver.
  • the performance of step 203 may comprise ignoring the received sensor data and determining the subject status to be the subject status received as non-sensor data.
  • rules e.g. generated by the machine learning module of the processing unit, are used in the combining.
  • the determined subject status may take any of the forms described above in relation to the processing unit of the restraint management apparatus 10.
  • determining a status of a subject can comprise any of the processing techniques described above in relation to the processing unit of the restraint management apparatus 10.
  • a restraint parameter is determined based on the determined subject status.
  • the determined restraint parameter may take any of the forms described above in relation to the processing unit of the restraint management apparatus 10 or in relation to the restraint device 14.
  • determining a restraint parameter can comprise any of the processing techniques described above in relation to the processing unit of the restraint management apparatus. It is generally expected that a main consideration in the determining of a restraint parameter will be to maintain a balance between the comfort and freedom of the subject, and the clinical need to prevent certain adverse events occurring.
  • the determined restraint parameter may be output in any suitable manner (including any of the manners described above in relation to the output signal 12) such that it can be implemented in restraining (or releasing) the subject.
  • the method includes the optional additional step 205 of applying a restraint device to the subject in accordance with the determined restraint parameter.
  • the restraint device can have any or all of the features described above in relation to the restraint device 14.
  • applying the restraint device comprises a caregiver manually setting a parameter of the restraint device to be equal to the determined parameter.
  • applying the restraint device receiving a control signal based on a determined restraint parameter, and in response to the control signal, setting a parameter of its operation/configuration to be equal to the determined parameter.
  • FIG. 3 shows a specific example of an automated restraint device 30 for use with a restraint management apparatus according to the invention.
  • the restraint device 30 comprises various restraint components arranged on a hospital bed 31 , and a control unit 32 which is in communication (either wired or wireless) with the restraint components so as to be able to send control signals to each of the restraint components.
  • the restraint components comprise a pair of bed rails 33, the height of which can be adjusted in response to a control signal from the control unit 32 (e.g. by means of any suitable actuator).
  • the restraint components also comprise a pair of wrist cuffs 34, each of which is connected to the hospital bed 31 by a tether 35.
  • the restraint components also comprise a pair of ankle cuffs 36, each of which is connected to the hospital bed 31 by a tether 37.
  • the tightness of the wrist and ankle cuffs is manually adjustable (although it will be appreciated that
  • the tightness of the wrist and/or ankle cuffs can be automatically adjusted in response to a control signal from the control unit 32, e.g. by means of any suitable actuator).
  • the length of each wrist and ankle tethers is individually and automatically adjustable in response to a control signal from the control unit 32, e.g. by means of any suitable actuator.
  • the restraint components also comprise a waist strap 38.
  • the waist strap 38 is arranged to be manually applied (although it will be appreciated that embodiments are possible in which the waist strap can be automatically applied) but once applied can be tightened automatically in response to a control signal from the control unit 32.
  • the restraint components can comprise, e.g. a wrist strap, with one end of a connector string or cord attached thereto, and the other end attached to a spool inside a restraint actuator.
  • the restraint actuator controls the freedom of movement of the spool, e.g. using a braking system (known to those skilled in the art of e.g.
  • the applied braking force can be proportional to the effective restraining force in the string or cord.
  • the braking system may include a force sensor for force feedback control.
  • F min a minimal restraining force
  • the restraint device 30 further comprises a plurality of sensors (not shown) which are in communication (wired or wireless) with the control unit 32 and are arranged to send data to the control unit 32.
  • the sensors comprise an accelerometer attached to each wrist cuff 34 and ankle cuff 36, a plurality of pressure sensors distributed over the base of the hospital bed 31, a position sensor attached to each bed rail 33, and a strain gauge attached to each tether 35, 37.
  • the control unit is in communication with a restraint management apparatus (e.g. the restraint management apparatus 10 of Figure 1) and is arranged to send data received from the plurality of sensors to the restraint management apparatus.
  • the restraint management apparatus is provided a separate device to the restraint device, however; alternative embodiments are envisaged in which the restraint management apparatus is provided as a module of the control unit 32.
  • the control unit 32 is arranged to receive a control signal output by a restraint management apparatus, where the control signal is arranged to cause the restraint device to implement one or more restraint parameters determined by the restraint management apparatus.
  • the control unit is arranged to, in response to receiving the control signal, implement a determined restraint parameter by sending a control signal to a restraint component arranged to cause an operational parameter of that restraint component to become equal to the determined restraint parameter.
  • the control unit 32 sends a control signal to each of the wrist tethers 35, the control signal being arranged to cause an actuator associated with each of the wrist tethers to adjust the tether length to be equal to 50cm.
  • the control unit 32 further comprises a user interface, via which restraint parameters can be manually input by a caregiver.
  • the control unit 32 is arranged to, in response to receiving a manually- input restraint parameter, implement the manually- input restraint parameter by sending a control signal to a restraint component arranged to cause an operational parameter of that restraint component to become equal to the manually-input restraint parameter.
  • the control unit 32 is also arranged to send the manually-input restraint parameter to the restraint management apparatus, e.g. for use by the restraint management apparatus in generating or updating rules/relationships relating subject status to restraint parameters, as described above in relation to Figure 1.
  • the control unit 32 is arranged to wait for a predetermined amount of time before implementing a change to an operating parameter of a given restraint component.
  • the predetermined waiting time may be longer if the last change was performed to implement a manually- input restraint parameter than if the last change was performed to implement a restraint parameter determined by the restraint management apparatus.
  • the control unit is arranged such that a manual input overrides signals received from the restraint management apparatus for a predetermined period of time.
  • Embodiments of the invention therefore advantageously enable the amount of time for which a subject is inappropriately restrained to be reduced, whilst also enabling restraints to be applied rapidly in response to potentially harmful movements by the subject, thus reducing their risk of harming them self, other people, or medical equipment.
  • Embodiments of the invention can facilitate automatic restraint implementation and adjustment, and therefore also have the potential to significantly reduce the burden on caregivers, by reducing or eliminating the requirement for caregiver involvement in both monitoring a subject and adjusting restraint devices.
  • a computer program may be stored/distributed on a suitable medium, such as an optical storage medium or a solid-state medium supplied together with or as part of other hardware, but may also be distributed in other forms, such as via the Internet or other wired or wireless telecommunication systems. Any reference signs in the claims should not be construed as limiting the scope.

Landscapes

  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Engineering & Computer Science (AREA)
  • Public Health (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Veterinary Medicine (AREA)
  • Animal Behavior & Ethology (AREA)
  • Biomedical Technology (AREA)
  • Physics & Mathematics (AREA)
  • Biophysics (AREA)
  • Surgery (AREA)
  • Pathology (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Molecular Biology (AREA)
  • Physiology (AREA)
  • Theoretical Computer Science (AREA)
  • Dentistry (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Software Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Nursing (AREA)
  • Mathematical Physics (AREA)
  • Artificial Intelligence (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Radiology & Medical Imaging (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Anesthesiology (AREA)
  • Business, Economics & Management (AREA)
  • General Business, Economics & Management (AREA)
  • Epidemiology (AREA)
  • Primary Health Care (AREA)
  • Computational Linguistics (AREA)
  • Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
  • Invalid Beds And Related Equipment (AREA)
  • Measuring And Recording Apparatus For Diagnosis (AREA)

Abstract

There is provided a restraint management apparatus. The restraint management apparatus comprises a processing unit arranged to: receive one or more types of sensor data; determine a status of a subject based on the received sensor data; determine, based on the determined subject status, a restraint parameter for a restraint device configured to restrain a body part of the subject; and output a signal based on the determined restraint parameter.

Description

RESTRAINT MANAGEMENT
TECHNICAL FIELD OF THE INVENTION
The invention relates to a restraint management apparatus and method for use in restraining a subject, and in particular relates to a restraint management apparatus and method for use in conjunction with a medical restraint system.
BACKGROUND TO THE INVENTION
In clinical environments such as hospitals and care homes, it is sometimes necessary to restrain a subject, e.g. by fixating them to their bed by means of restraining straps or the like. Use of such restraints may be required, for example, when patients
(subjects) become very active due to a certain disease (e.g. one which causes delirium), such that they risk injuring themselves or dislodging or damaging medical equipment such as IV lines, endotracheal tubes or feeding tubes. Some subjects may not be permitted to get out of bed without assistance, for instance because they are connected to medical equipment and/or because they are at a high risk of falling over if they attempt to do so. It is therefore an important patient safety issue to restrain certain subjects when necessary. Currently, restraining straps are applied manually by medical staff. However; with current systems changing restraint parameters such as range/length (i.e. the freedom of movement the restraining straps permit the subject) requires manual intervention by a caregiver, and this means that the restraint parameters are normally kept the same for the whole period during which the restraints are in place.
Decisions about when to apply/release restraints for a given patient are made by medical staff based on their observations. However; it is generally not possible for medical staff to observe every subject in their care, all of the time. This means that there are times when a subject's restraint status is not appropriate to their current condition (e.g. they remain restrained even when lying quietly, they are not restrained whilst moving violently, they are restrained with too great or too little freedom of movement for their current condition).
There is therefore a need for a system which can reduce the amount of time for which a subject is inappropriately restrained. Preferably such a system would be able to achieve a more optimal balance between the clinical need to ensure the safety of the subject and the subject's need to be comfortable and free to move.
SUMMARY OF THE INVENTION
According to a first aspect of the invention, there is provided a restraint management apparatus. The restraint management apparatus comprises a processing unit arranged to: receive one or more types of sensor data; determine a status of a subject based on the received sensor data; determine, based on the determined subject status, a restraint parameter for a restraint device configured to restrain a body part of the subject; and output a signal based on the determined restraint parameter.
Thus, embodiments of the invention advantageously provide an automatic way to predict when a given subject should be restrained, released, or allowed more or less freedom of movement, based on the detected condition of the subject. This creates the possibility of automatically updating restraint parameters in real time, in response to changes in a subject's condition. Thus, an optimum balance between the clinical need to ensure the safety of the subject and the subject's need to be comfortable and free to move can be achieved.
In some embodiments the processing unit is further arranged to receive one or more types of non-sensor data relating to the subject. In some such embodiments the processing unit is arranged to determine a status of a subject based additionally on the received non-sensor data. In some embodiments the one or more types of non-sensor data comprises data originating from a caregiver. In some such embodiments the one or more types of non-sensor data comprises data manually input by a caregiver to the restraint management apparatus or to a device in communication with the restraint management apparatus.
In some embodiments the processing unit comprises a memory containing one or more predefined signatures, each predefined signature relating to a particular type of feature. In some such embodiments the processing unit is arranged to determine a status of a subject by comparing the received data to the one or more predefined signatures to detect one or more features in the received sensor data.
In some embodiments the processing unit is arranged to determine a status of a subject by determining one or more of: a physiological condition of the subject, a position of the subject; a position of a body part of the subject; a movement level of a body part of the subject; an activity level of the subject; whether the subject is asleep or awake; whether the subject's eyes are open; whether the subject is delirious; whether movement of a body part of the subject meets a predefined criterion; whether movement of the subject meets a predefined criterion.
In some embodiments the processing unit is arranged to determine a status of a subject such that the determined status comprises a status value for one or more factors. In some such embodiments status value comprises one or more of: a descriptive indication; a numerical score; a non-numerical level indication.
In some embodiments the processing unit comprises a memory containing rules relating subject status values to restraint parameters. In some such embodiments the processing unit is arranged to determine a restraint parameter by applying one or more of the rules to the determined subject status value. In some such embodiments the processing unit comprises a machine learning module and the rules have been generated by the machine learning module based on historical data relating to the subject. In some embodiments the processing unit is arranged to receive a restraint parameter which has been manually- input to a restraint device, and the machine learning module is arranged to update the generated rules based on the received manually- input restraint parameter.
In some embodiments the processing unit is arranged to determine a restraint parameter such that the determined restraint parameter comprises one of: activation of a restraint device; inactivation of a restraint device; activation of a given restraining component of a restraint device; inactivation of a given restraining component of a restraint device; tightness of a given restraining component of a restraint device; length of a given restraining component of a restraint device; duration of activation of a restraint device; duration of activation of a given restraining component of a restraint device.
In some embodiments the processing unit is arranged to output a signal comprising one or more of: a control signal to an automated restraint device arranged to cause the automated restraint device to apply the determined restraint parameter; a signal to a device associated with a caregiver, the signal comprising an instruction to the caregiver to implement the determined restraint parameter; a control signal to an alarm module of the restraint management apparatus arranged to cause the alarm module to generate an alarm in dependence on the determined restraint parameter; a control signal to a display module of the restraint management apparatus arranged to cause the display module to display information in dependence on the determined restraint parameter; a control signal to a subject feedback module arranged to cause the subject feedback module to generate a message to the subject; a control signal to a lighting module arranged to cause the lighting module to activate a light. In some embodiments the processing unit is arranged to output the signal after a predefined amount of time has passed since the processing unit last output a signal of the same type as the signal.
The invention also provides, according to a second aspect, a restraint system. The restraint system comprises a restraint management apparatus according to the first aspect; and a restraint device configured to restrain a body part of a subject. The restraint device is arranged to: receive a signal output by the restraint management apparatus, and in response to a signal received from the restraint management apparatus, apply a restraint to the subject or alter a parameter of a restraint applied to the subject.
In some embodiments the restraint device is further arranged to: receive a manually- input restraint parameter; in response in response to a signal received from the restraint management apparatus, apply a restraint to the subject or alter a parameter of a restraint applied to the subject; and transmit a signal containing the manually- input restraint parameter to the restraint management apparatus.
The invention also provides, according to a third aspect, a method for use in restraining a subject. The method comprises: receiving one or more types of sensor data;
determining a status of a subject based on the received sensor data; and determining, based on the determined subject status, a restraint parameter for a restraint device configured to restrain a body part of the subject.
In some embodiments the received sensor data comprises one or more of: an image including a subject; an image including a subject and one or more objects in the vicinity of the subject; audio data including sounds generated by a subject; accelerometer data obtained by an accelerometer worn by a subject; measurements of one or more physiological characteristics of a subject; restraint-related data obtained by a sensor associated with a restraint device; pressure data obtained by a bed sensor; bed rail position data obtained by a bed rail sensor.
In some embodiments the received sensor data comprises an image including a subject and one or more objects. In some such embodiments determining a status of a subject comprises: determining a position of a body part of the subject; determining a position of an object in the image; calculating a distance between the body part and the object; and comparing the calculated distance to a predetermined threshold. BRIEF DESCRIPTION OF THE DRAWINGS
For a better understanding of the invention, and to show more clearly how it may be carried into effect, reference will now be made, by way of example only, to the accompanying drawings, in which:
Figure 1 is an illustration of a restraint management apparatus according to a general embodiment of the invention;
Figure 2 is a flow chart illustrating a method for use in restraining a subject according to a general embodiment of the invention; and
Figure 3 is an illustration of an example restraint device for use with a restraint management apparatus according to the invention.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
Figure 1 shows a restraint management apparatus 10 according to an embodiment of the invention. The restraint management apparatus 10 is for use with a restraint device 14 configured to restrain a body part of a subject. In embodiments the restraint device 14 is configured to provide a controllable resistance to movement of a body part. The restraint management apparatus 10 comprises a processing unit. In some embodiments the restraint management apparatus also comprises one or more functional modules, e.g. an alarm module, a display module, a communications module, a user input module, a subject feedback module to present messages to the subject (e.g. audio or text messages), a lighting module to illuminate the subject and/or their surroundings, etc. The processing unit of the restraint management apparatus 10 is arranged to receive one or more types of sensor data 16. In some embodiments the processing unit is arranged to receive sensor data 16 from one or more sensor devices 18. Each sensor device 18 can be, e.g., a camera, microphone, accelerometer, medical monitoring device, or any other type of sensor device arranged to collect sensor data relating to a subject. In some embodiments the processing unit is arranged to receive sensor data from the restraint device 14. In some embodiments the processing unit is arranged to receive sensor data from a user input module of the restraint management apparatus (e.g. where a caregiver has manually input readings obtained by a sensor), and/or a communications module of the restraint management apparatus (e.g. where sensor data is sent to the restraint management apparatus from a remote electronic device, such as a laptop or smartphone of a caregiver). In some embodiments the processing unit is arranged to receive sensor data in real time. In alternative embodiments the processing unit is arranged to receive sensor data at periodic intervals, e.g. in response to a request sent from the processing unit to a sensor device 18 or the restraint device 14.
In some embodiments the processing unit is further arranged to receive non- sensor data, which may or may not be related to the subject. Subject-related non-sensor data can comprise, for example, historical subject information, medical test results, diagnoses, etc. Non-subject related non-sensor data can comprise, for example, software updates, information relating to a restraint device, or any other type of information/data that may be used by the restraint management apparatus. In some embodiments the non-sensor data comprises a manually- input restraint parameter which has been input to a restraint device by a caregiver. It will be appreciated that manually- input data comprises any data received by a device which originates from a caregiver or other human. Manually-input data can, but need not, be input via an input interface of the device receiving the data. Manually-input data can also be input via a remote device in communication with the device receiving the data. Non- sensor data may be received in any of the manners described above in relation to the sensor data.
The processing unit is further arranged to determine a status of a subject based on the received sensor data (and, in some embodiments, received non-sensor data). The determined status should be related to the restraint requirements of the subject, and will generally therefore be indicative of a current or expected amount/type of activity/movement of the subject. In some embodiments the processing unit is arranged to determine a status of a subject by determining one or more of: a physiological condition of the subject, a position of the subject; a position of a body part of the subject; a movement level of a body part of the subject; an activity level of the subject; whether the subject is asleep or awake; whether the subject's eyes are open; whether the subject is delirious; whether movement of a body part of the subject meets a predefined criterion (e.g. defined such that only unusual movement will meet the criterion); whether movement of the subject meets a predefined criterion (e.g.
defined such that only unusual movement will meet the criterion).
Various techniques for extracting physiological, position and/or activity information from sensor data are known in the art. For example, data from an accelerometer mounted on the subject, and/or from a pressure sensor mounted on their hospital bed can be analyzed to determine the position of a subject and/or whether and to what extent the subject is moving. Subject movement can also be determined from image data acquired by a camera, using video analysis techniques. Image data can be used to determine the proximity of a body part of a subject to a medical device or other object (e.g. a wall, bed rail, item of furniture, etc.) Audio data collected by a microphone can detect vocalizations, movement noises and/or impact noises generated by the patient, which can be used to determine whether they are in distress, likely to attempt to get out of bed, and/or are at risk of harming them self.
Physiological data (e.g. heart rate, respiration rate, breathing pattern, ECG data) can be used in the determination of the status of a subject, e.g. by indicating whether the subject is asleep or awake, indicating which sleep stage the subject is in, and/or indicating whether they are experiencing increased circulatory effort (which can result from increased activity).
In some embodiments the processing unit is arranged to determine the status of a subject by detecting patterns and/or features in the received sensor data. For example, in some embodiments signatures and/or patterns associated with features of interest (such as a particular body part in an image of the subject; a particular object in an image of the subject; a vocalization in an audio recording of the subject; an impact sound in an audio recording of the subject; a peak in accelerometer data, etc.) are stored in a memory associated with the processing unit, and the processing unit is arranged to determine the status of a subject by comparing detected patterns/features with the stored signatures/patterns. In some
embodiments the determination is performed based on received sensor data from a particular instant in time (i.e. the determination uses a static analysis). In some embodiments the determination is performed using received sensor data covering a time period (i.e. the determination uses a dynamic analysis). In some embodiments determining the status of a subject comprises tracking a detected feature over a time period. In some embodiments determining the status of a subject comprises detecting movement of a body part of the subject. In some such embodiments, detected movements are classified as being one or more of several different predefined movement types. Any suitable known signal analysis and feature extraction techniques may be used in the performance of step 202.
In a particular embodiment in which the received sensor data 16 comprises image data containing the subject and their immediate surroundings, the processing unit is arranged to determine the status of the subject by detecting, using image analysis techniques, the relative positions of body parts of the subjects and objects. The detected body parts can comprise, for example, one or more of a hand, a foot, the head, etc. The detected objects can comprise, for example, one or more of a patient monitor, an IV stand, a bed rail, a bedside table, etc. In some such embodiments the processing unit is arranged to calculate distances between the detected body parts and the detected objects. In some such embodiments the processing unit is arranged to compare the calculated distances to a predefined threshold. The subject status may then be determined in dependence on the result(s) of the comparing. For example, in some embodiments a status indicating that restraint of a body part is required will be determined only if the distance between that body part and an object is less than the predefined threshold. In some embodiments a status indicating that restraint of multiple or all body parts is required will be determined if the distance between at least one body part and an object is less than the predefined threshold. In further or alternative embodiments, the processing unit is arranged to calculate the velocity of a detected body part towards a detected object (for example from a series of measurements of the relative positions of body parts and objects obtained from a time series of images). In some such embodiments the processing unit is arranged to compare the calculated velocity to a predefined velocity threshold. The subject status may then be determined in dependence on the result(s) of the comparing. For example, in some embodiments a status indicating that restraint of a body part is required will be determined only if the velocity of the body part towards the object is more than the predefined velocity threshold. In some embodiments a status indicating that restraint of multiple or all body parts is required will be determined if the velocity of at least one body part towards an object is more than the predefined velocity threshold. An exemplary velocity threshold could be 2m/s.
In some embodiments (i.e. embodiments in which the processing unit is arranged to receive sensor data in real time) the processing unit is arranged to determine a status of a subject in real time or near real time.
In some embodiments the determined subject status comprises a status value
(which may, for example, take the form of a descriptive indication (e.g. asleep with low activity, asleep with high activity, awake with low activity, awake with high activity, delirious, violent, etc.) a numerical score (e.g. representing activity level), and/or a non- numerical status indication (e.g. a color corresponding to a low, medium or high risk that the patient will cause harm if not restrained). In some embodiments the determined subject status comprises a status value in respect of each of multiple factors, where a factor could be, e.g., activity level, medical condition, sleep/wake state, lucidity, etc. In some embodiments the determined subject status is indicated, e.g. by the restraint management apparatus or a device in communication with the restraint management apparatus. In some such embodiments indicating the determined subject status comprises displaying a message or visual status indication. In some embodiments indicating the determined subject status comprises emitting light of a particular color.
The processing unit is further arranged to determine, based on the determined subject status, a restraint parameter for a restraint device configured to restrain a body part of the subject, e.g. the restraint device 14. The determined restraint parameter can comprise, for example, one or more of: activation/inactivation of the restraint device (i.e. as a whole), activation/inactivation of a given restraining component of the restraint device,
tightness/length of a given restraining component, duration of activation of the restraint device or of a given restraining component, etc. In some embodiments the processing unit is provided with a set of rules/relationships (e.g. stored in a memory of the restraint
management device) relating restraint parameters to subject status and is arranged to use these rules/relationships in determining a restraint parameter. An example of such a rule could be, "if subject status = excessive hand movement, set hand restraining components to maximum tightness' ' .
In embodiments where the restraint device can provide a controllable resistance to movement of a body part, the restraint parameter can indicate a level of resistance to movement to be provided by the restraint device.
In some embodiments the processing unit is arranged to determine a restraint parameter by applying at least one predefined criterion to the determined subject status. In some embodiments the predefined criterion comprises a set of conditions. In some embodiments the set of conditions comprises a time-based condition and a value-based condition. A value-based condition is defined such that whether or not a status value meets the condition depends on the value itself. A time-based condition, on the other hand, specifies a condition (e.g. a minimum) relating to the amount of time for which the status information has consistently met a value-based condition. Providing a time-based condition can be advantageous because some status factors may need to be consistently present for a duration of time before they can be considered to reliably indicate that a change in a restraint parameter (e.g. application of the restraint device to a currently unrestrained subject) is required. In some embodiments the set of conditions comprises a plurality of value-based conditions.
In some embodiments the processing unit is arranged to determine a restraint parameter using a function of the distance from a body part to an object in the environment, and/or of the velocity of a body part towards an object in the environment. An exemplary function based on distance x is set out below:
x Λ <— " BT 1 l Fr = 1 F max
X > RT F = F ■
Otherwise Fr = Fmax + (BT - x)/(RT - BT) (Fmax - Fmin) where BT is a blocking threshold distance at which the movement of the body part is to be blocked (i.e. a maximum restraining force, Fmax, is to be applied), RT is a restraining threshold distance beyond which a minimum restraining force (Fmin) is to be applied. This function leads to the restraining force increasing from the minimum force at distances greater than RT up to the maximum force at distances of BT or less. An exemplary value for Fmax can be 1000 Newton (N), although any force that would prevent movement of the body part can be used. An exemplary value for Fmin can be 0.1 N, although any force that would allow free or reasonably free movement of the body part can be used. An exemplary value for BT can be 10 centimeters, and an exemplary value for RT can be 25 cm. The restraining force provided by the above function can be related to a restraint parameter. It will be appreciated that the above function is merely exemplary, and those skilled in the art will be aware of other types or forms of functions that can be used to relate distance and/or velocity to a restraining force.
In embodiments where the subject status comprises a status value for each of multiple factors, the at least one predefined criterion may be factor-specific. For example, in some embodiments the determined subject status comprises a status value for the factor "sleep/wake state" and a status value for the factor "activity level". In some such
embodiments the processing unit is arranged to determine a restraint parameter based on both whether the sleep/wake state value is equal to a predefined value and whether the activity level value exceeds a predefined threshold. For example, in some such embodiments if the sleep/wake status value is "asleep" and the activity level is "medium" then the processing unit will determine that the restraining device should be used, but if the sleep/wake status value is "awake" and the activity level is "medium" then the processing unit will determine that the restraining device should not be used.
In some embodiments the processing unit includes a machine learning module, configured to use standard machine learning techniques to identify or generate rules, relationships, etc. relating restraint parameters to subject status. In some such embodiments the machine learning module is configured to apply machine learning techniques to historical data, e.g. historical subject status data, historical physiological measurement data, and/or historical restraint information (e.g. information about restraint parameters used and/or their appropriateness). In some embodiments (i.e. embodiments in which the processing unit is arranged to receive manually- input restraint parameters from the restraint device 14), the machine learning module is configured to generate rules/relationships relating restraint parameters to subject status based on manually- input restraint parameters received from the restraint apparatus. In some such embodiments the machine learning module is arranged to update existing rules/relationships based on manually- input restraint parameters received from the restraint apparatus. In some such embodiments the processing unit 12 is arranged to determine a restraint parameter based on rules or relationships generated or identified by the machine learning module.
In preferred embodiments the processing unit is arranged to determine a restraint parameter in real time or near real time. Advantageously this feature can prevent inappropriate restraint of a subject, and minimize the amount of time for which a subject is restrained. Conversely, it can ensure that restraint is applied rapidly when required, reducing the risk of harm to the subject.
The processing unit is further arranged to output a signal 12 based on the determined restraint parameter. In some embodiments the signal 12 comprises a control signal, and is output to the restraint device 14, e.g. by a communications module of the restraint management apparatus 10. In some such embodiments the signal 12 causes the restraint device to apply the determined restraint parameter. For example, in one example embodiment in which the determined restraint parameter comprises a specified length of a given restraint component of the restraint device 14, the output signal 12 comprises a control signal which causes the restraint device 14 to set the length of the given restraint component to be the specified length. In another example embodiment in which the determined restraint parameter indicates a level of resistance to movement to be provided by the restraint device 14, the output signal 12 comprises a control signal which causes the restraint device 14 to set the resistance to movement to the specified level of resistance. In some embodiments in which the signal 12 is output to the restraint device 14, the signal 12 is arranged to cause the restraint device 14 to display the determined restraint parameter, e.g. on a display associated with the restraint device 14. In some embodiments the signal comprises an instruction to a caregiver to implement the determined restraint parameter.
In some embodiments the signal comprises a control signal arranged to cause a message to be presented to the subject. Such a message could provide the subject with instructions and/or reassurance. A suitable message could say, for example, "Please calm down. A nurse will be with you shortly". In some embodiments the message to the subject comprises an audio message, e.g. generated by a loudspeaker comprised in or in
communication with the restraint management device 10. In some embodiments the message to the subject comprises a visual message, e.g. displayed by a display comprised in or in communication with the restraint management device 10. In some embodiments the processing unit is arranged to output the signal 12 after a predefined amount of time has passed since the processing unit last output a signal of the same type as the current signal 12 (e.g. in some embodiments an instruction or control signal arranged to cause adjustment of a particular restraint would comprise a first type of signal, and an instruction or control signal arranged to cause adjustment of a different restraint, a control signal arranged to generate an alarm, a warning message to a caregiver, etc. would each be considered to be a signal of a different type to the first signal). This can reduce the stress experienced by the subject, since if restraint parameters are changed too often this can confuse or surprise the subject.
In some embodiments the signal 12 comprises a control signal and is sent to a functional module of the restraint management apparatus 10. In such embodiments the control signal may be arranged, for example, to cause one or more of:
an alarm module to generate an alarm;
a subject feedback module to generate a message to the subject;
a communications module to generate a message to a caregiver;
a communications module to send a signal to a remote device;
a communications module to send a signal to the restraint management apparatus;
a display module to display information;
a lighting module to activate a light.
In some embodiments the processing unit is arranged to output a plurality of signals, for example a control signal to the restraint device 14 and a control signal to an alarm module of the restraint management apparatus 10 causing it to generate an alarm.
In some embodiments the processing unit is arranged to output a signal 12 continuously, in real-time or near real-time. In some embodiments the processing unit is arranged to output a signal in response to a change in the determined parameter. In some embodiments the processing unit is arranged to output a signal 12 at periodic intervals.
The restraint device 14 can be any device arranged to restrain (i.e. to restrict the movement of) a body part of a subject. In some embodiments the restraint device 14 comprises multiple restraint components, each of which is arranged to restrain a particular body part. In some such embodiments, the restraint components comprise one or more straps (e.g. wrist straps, ankle straps, a waist strap, etc.). In some embodiments the restraint components comprise one or more tethers. In some embodiments the restraint device 14 comprises one or more bed rails. In some embodiments the restraint device 14 comprises a bed sheet or cover which can be adjustably tightened over the subject. In some embodiments the restraint device is attached to or integrated in a hospital bed. Preferably the restraint device 14 is arranged such that one or more restraint parameters (e.g. strap tightness/length, tether length, bed rail height, resistance to movement) are adjustable during use of the restraint device 14. Preferably, the restraint device is arranged to be as user friendly as possible. For example, straps can comprise a soft and/or padded material. Straps and/or tethers may be arranged to vary the resistance to movement in dependence on the
length/tightness for variable fixation ranges. This can advantageously achieve a subtle braking of a moving body part rather than a hard stop, which can reduce the stress felt by the subject as a result of being restrained.
In some embodiments the restraint device 14 includes at least one sensor for acquiring sensor data. In some embodiments the restraint device 14 includes at least one sensor associated with each restraint component of the restraint device 14. The at least one sensor can comprise, for example, a strain gauge; an accelerometer; a position sensor. In some embodiments the restraint device 14 includes a sensor arranged to detect a parameter of the subject (e.g. position of a given body part, movement of a given body part, etc.). In some embodiments the restraint device 14 includes a sensor arranged to detect a parameter of the restraint device (e.g. position of a bed rail; length of a strap; tension in a strap/tether; whether or not a given restraint component is in use; etc.). In some embodiments in which the restraint device 14 includes at least one sensor, the restraint device 14 is arranged to output sensor data, e.g. to the restraint management apparatus 10 or to another remote device (such as a patient monitoring device or a hospital computer system).
The restraint device 14 can be an automated restraint device or a manually operated restraint device. A manually operated restraint device requires a human operator to adjust its restraint parameters. In embodiments where the restraint management device 10 is used in conjunction with a manually operated restraint device, the restraint management device is arranged to alert a caregiver to a change in the determined restraint parameter, so that the caregiver can implement the newly determined parameter. Although caregiver intervention to adjust the restraint device is still required, in such situations the restraint management system significantly reduces the burden on caregivers by reducing or removing the need for a caregiver to continuously monitor the status of subject.
An automated restraint device, on the other hand, is able to adjust one or more of its restraint parameters without input from a human operator, e.g. in response to a control signal received from the restraint management apparatus 10. Advantageously, the use of an automated restraint device in conjunction with a restraint management apparatus according to embodiments of the invention can enable the rapid (preferably real-time) adjustment of the nature and degree of restraint applied to a given subject, based on their current status. This can significantly reduce or even eliminate periods of inappropriate restraint. Preferably an automated restraint device is arranged such that restraint parameters can also be adjusted manually, e.g. via a user interface of the restraint device. In some embodiments, an
automated restraint device is arranged such that a caregiver is required to approve/confirm a restraint parameter generated by the restraint management apparatus before the automated restraint device will apply that parameter. In some such embodiments the approval can be done remotely, via a device (such as a nurses' station, or portable electronic device of a caregiver) which is in communication with the restraint management apparatus and the restraint device.
Figure 2 shows an example of a method for use in restraining a subject, which can be implemented by the restraint management apparatus 10. In a first step 201 sensor data 16 is received, e.g. from one or more sensor devices, from a restraint device, from a user input module of the restraint management apparatus 10, from a remote electronic device, etc. In some embodiments the sensor data is received as a continuous stream. In some
embodiments the sensor data is received periodically. In some embodiments the sensor data is received in response to a request previously sent to a sensor device/restraint device. In some embodiments the sensor data has been time-stamped at the time of its generation. In some embodiments the sensor data is time-stamped at the time of its receipt, e.g. by the processing unit of the restraint management apparatus 10. In some embodiments the received sensor data directly relates to a subject, in that it comprises measurements of a physiological characteristic of that subject. In some embodiments the received sensor data indirectly relates to a subject. For example, data which indirectly relates to a subject could comprise an image or audio recording of the subject's hospital room which includes the subject as well as their surroundings. In some embodiments the received sensor data is stored, e.g. in a memory of the restraint management apparatus 10.
In some embodiments an optional step 202 of receiving non-sensor data relating to the subject, e.g. from a user input module of the restraint management apparatus, a remote electronic device, etc. is performed. The subject-related non-sensor data can comprise, e.g. test results; a medical diagnosis; historical physiological characteristic data; clinical outcome data; treatment data, restraint data, etc. In some embodiments the received subject- related non-sensor data is stored, e.g. in a memory of the restraint management apparatus 10. In step 203 a status of a subject (i.e. the subject to which the received sensor data either directly or indirectly relates) is determined based on the received sensor data, e.g. by the processing unit of the restraint management apparatus 10. In embodiments in which subject-related non-sensor data is received, a status of the subject is determined based additionally on the non-sensor data. In some such embodiments the received sensor data and the received non-sensor data are combined by a data fusion module of the processing unit of the restraint management apparatus 10. In some embodiments the combining comprises assigning one or more weights to the received sensor data and the received non-sensor data. For example, non-sensor data in the form of data input manually by a caregiver may be assigned a higher weight than sensor data. In some embodiments the non-sensor data can comprise a subject status manually input by a caregiver. In some such embodiments the performance of step 203 may comprise ignoring the received sensor data and determining the subject status to be the subject status received as non-sensor data. In some embodiments rules, e.g. generated by the machine learning module of the processing unit, are used in the combining. The determined subject status may take any of the forms described above in relation to the processing unit of the restraint management apparatus 10. Likewise, determining a status of a subject can comprise any of the processing techniques described above in relation to the processing unit of the restraint management apparatus 10.
In step 204 a restraint parameter is determined based on the determined subject status. The determined restraint parameter may take any of the forms described above in relation to the processing unit of the restraint management apparatus 10 or in relation to the restraint device 14. Likewise, determining a restraint parameter can comprise any of the processing techniques described above in relation to the processing unit of the restraint management apparatus. It is generally expected that a main consideration in the determining of a restraint parameter will be to maintain a balance between the comfort and freedom of the subject, and the clinical need to prevent certain adverse events occurring. The determined restraint parameter may be output in any suitable manner (including any of the manners described above in relation to the output signal 12) such that it can be implemented in restraining (or releasing) the subject.
In some embodiments the method includes the optional additional step 205 of applying a restraint device to the subject in accordance with the determined restraint parameter. The restraint device can have any or all of the features described above in relation to the restraint device 14. In some embodiments applying the restraint device comprises a caregiver manually setting a parameter of the restraint device to be equal to the determined parameter. In some embodiments applying the restraint device receiving a control signal based on a determined restraint parameter, and in response to the control signal, setting a parameter of its operation/configuration to be equal to the determined parameter.
Figure 3 shows a specific example of an automated restraint device 30 for use with a restraint management apparatus according to the invention. The restraint device 30 comprises various restraint components arranged on a hospital bed 31 , and a control unit 32 which is in communication (either wired or wireless) with the restraint components so as to be able to send control signals to each of the restraint components. The restraint components comprise a pair of bed rails 33, the height of which can be adjusted in response to a control signal from the control unit 32 (e.g. by means of any suitable actuator). The restraint components also comprise a pair of wrist cuffs 34, each of which is connected to the hospital bed 31 by a tether 35. Similarly, the restraint components also comprise a pair of ankle cuffs 36, each of which is connected to the hospital bed 31 by a tether 37. The tightness of the wrist and ankle cuffs is manually adjustable (although it will be appreciated that
embodiments are possible in which the tightness of the wrist and/or ankle cuffs can be automatically adjusted in response to a control signal from the control unit 32, e.g. by means of any suitable actuator). The length of each wrist and ankle tethers is individually and automatically adjustable in response to a control signal from the control unit 32, e.g. by means of any suitable actuator. The restraint components also comprise a waist strap 38. The waist strap 38 is arranged to be manually applied (although it will be appreciated that embodiments are possible in which the waist strap can be automatically applied) but once applied can be tightened automatically in response to a control signal from the control unit 32.
Where the restraint device is to provide a subtle braking of a moving body part rather than a hard stop as described above, the restraint components can comprise, e.g. a wrist strap, with one end of a connector string or cord attached thereto, and the other end attached to a spool inside a restraint actuator. The restraint actuator controls the freedom of movement of the spool, e.g. using a braking system (known to those skilled in the art of e.g.
mechatronics). Depending on the configuration of the actuator system, the applied braking force can be proportional to the effective restraining force in the string or cord. The braking system may include a force sensor for force feedback control. In some embodiments, in the case of using a braking system as the main actuator, it can be important to maintain a minimal restraining force (e.g. Fmin) in the string or cord at all times, to avoid slacking and tangling. This can be realized using a simple spring connected to the spool. Those skilled in the art will be aware of various types of restraint components that can be used to provide subtle braking of a moving body part.
The restraint device 30 further comprises a plurality of sensors (not shown) which are in communication (wired or wireless) with the control unit 32 and are arranged to send data to the control unit 32. The sensors comprise an accelerometer attached to each wrist cuff 34 and ankle cuff 36, a plurality of pressure sensors distributed over the base of the hospital bed 31, a position sensor attached to each bed rail 33, and a strain gauge attached to each tether 35, 37. The control unit is in communication with a restraint management apparatus (e.g. the restraint management apparatus 10 of Figure 1) and is arranged to send data received from the plurality of sensors to the restraint management apparatus. In this embodiment the restraint management apparatus is provided a separate device to the restraint device, however; alternative embodiments are envisaged in which the restraint management apparatus is provided as a module of the control unit 32.
The control unit 32 is arranged to receive a control signal output by a restraint management apparatus, where the control signal is arranged to cause the restraint device to implement one or more restraint parameters determined by the restraint management apparatus. The control unit is arranged to, in response to receiving the control signal, implement a determined restraint parameter by sending a control signal to a restraint component arranged to cause an operational parameter of that restraint component to become equal to the determined restraint parameter. For example, in a case where the determined restraint parameter is "wrist tethers = 50cm", the control unit 32 sends a control signal to each of the wrist tethers 35, the control signal being arranged to cause an actuator associated with each of the wrist tethers to adjust the tether length to be equal to 50cm.
The control unit 32 further comprises a user interface, via which restraint parameters can be manually input by a caregiver. The control unit 32 is arranged to, in response to receiving a manually- input restraint parameter, implement the manually- input restraint parameter by sending a control signal to a restraint component arranged to cause an operational parameter of that restraint component to become equal to the manually-input restraint parameter. The control unit 32 is also arranged to send the manually-input restraint parameter to the restraint management apparatus, e.g. for use by the restraint management apparatus in generating or updating rules/relationships relating subject status to restraint parameters, as described above in relation to Figure 1. The control unit 32 is arranged to wait for a predetermined amount of time before implementing a change to an operating parameter of a given restraint component. In some embodiments the predetermined waiting time may be longer if the last change was performed to implement a manually- input restraint parameter than if the last change was performed to implement a restraint parameter determined by the restraint management apparatus. In some embodiments the control unit is arranged such that a manual input overrides signals received from the restraint management apparatus for a predetermined period of time.
Embodiments of the invention therefore advantageously enable the amount of time for which a subject is inappropriately restrained to be reduced, whilst also enabling restraints to be applied rapidly in response to potentially harmful movements by the subject, thus reducing their risk of harming them self, other people, or medical equipment.
Embodiments of the invention can facilitate automatic restraint implementation and adjustment, and therefore also have the potential to significantly reduce the burden on caregivers, by reducing or eliminating the requirement for caregiver involvement in both monitoring a subject and adjusting restraint devices.
While the invention has been illustrated and described in detail in the drawings and foregoing description, such illustration and description are to be considered illustrative or exemplary and not restrictive; the invention is not limited to the disclosed embodiments. Variations to the disclosed embodiments can be understood and effected by those skilled in the art in practicing the claimed invention, from a study of the drawings, the disclosure and the appended claims. In the claims, the word "comprising" does not exclude other elements or steps, and the indefinite article "a" or "an" does not exclude a plurality. A single processor or other unit may fulfil the functions of several items recited in the claims. The mere fact that certain measures are recited in mutually different dependent claims does not indicate that a combination of these measures cannot be used to advantage. A computer program may be stored/distributed on a suitable medium, such as an optical storage medium or a solid-state medium supplied together with or as part of other hardware, but may also be distributed in other forms, such as via the Internet or other wired or wireless telecommunication systems. Any reference signs in the claims should not be construed as limiting the scope.

Claims

CLAIMS:
1. A restraint management apparatus comprising:
a processing unit arranged to:
receive one or more types of sensor data;
determine a status of a subject based on the received sensor data, wherein the status comprises a distance between a body part of the subject and a selected object in the environment around the subject and/or velocity of a body part of the subject towards a selected object in the environment around the subject;
determine, based on the determined subject status, a restraint parameter for a restraint device configured to restrain the body part of the subject, wherein the restraint device is configured to provide a resistance to movement of the body part, and wherein the restraint parameter indicates a level of resistance to movement to be provided by the restraint device, and wherein the restraint parameter is determined such that the level of resistance to movement provided by the restraint device increases as the distance between the body part and the selected object decreases and/or as the velocity of the body part towards the selected object increases; and
output a signal based on the determined restraint parameter.
2. The restraint management apparatus of claim 1 , wherein the processing unit is further arranged to receive one or more types of non-sensor data relating to the subject and to determine a status of a subject based additionally on the received non-sensor data.
3. A restraint management apparatus according to any preceding claim, wherein the processing unit comprises a memory containing one or more predefined signatures, each predefined signature relating to a particular type of feature, and wherein the processing unit is arranged to determine a status of a subject by comparing the received data to the one or more predefined signatures to detect one or more features in the received sensor data.
4. A restraint management apparatus according to any preceding claim, wherein the processing unit is arranged to determine a status of a subject such that the determined status comprises a status value for one or more factors, wherein each status value comprises one or more of: a descriptive indication; a numerical score; a non-numerical level indication.
5. A restraint management apparatus according to claim 4, wherein the processing unit comprises a memory containing rules relating subject status values to restraint parameters, and is arranged to determine a restraint parameter by applying one or more of the rules to the determined subject status value.
6. A restraint management apparatus according to claim 5, wherein the processing unit comprises a machine learning module, wherein the rules have been generated by the machine learning module based on historical data relating to the subject, wherein the processing unit is arranged to receive a restraint parameter which has been manually- input to a restraint device, and wherein the machine learning module is arranged to update the generated rules based on the received manually- input restraint parameter.
7. A restraint management apparatus according to any preceding claim, wherein the processing unit is arranged to determine a restraint parameter such that the determined restraint parameter comprises one of:
activation of a restraint device;
inactivation of a restraint device;
activation of a given restraining component of a restraint device; inactivation of a given restraining component of a restraint device;
tightness of a given restraining component of a restraint device; length of a given restraining component of a restraint device; duration of activation of a restraint device;
duration of activation of a given restraining component of a restraint device.
8. A restraint management apparatus according to any preceding claim, wherein the processing unit is arranged to output a signal comprising one or more of:
a control signal to an automated restraint device arranged to cause the automated restraint device to apply the determined restraint parameter;
a signal to a device associated with a caregiver, the signal comprising an instruction to the caregiver to implement the determined restraint parameter;
a control signal to an alarm module of the restraint management apparatus arranged to cause the alarm module to generate an alarm in dependence on the determined restraint parameter;
a control signal to a display module of the restraint management apparatus arranged to cause the display module to display information in dependence on the determined restraint parameter;
a control signal to a subject feedback module arranged to cause the subject feedback module to generate a message to the subject;
a control signal to a lighting module arranged to cause the lighting module to activate a light.
9. A restraint management apparatus according to any preceding claim, wherein the processing unit is arranged to output the signal after a predefined amount of time has passed since the processing unit last output a signal of the same type as the signal.
10. A restraint system comprising :
a restraint management apparatus according to any of claims 1 to 9; and
a restraint device configured to restrain a body part of a subject, wherein the restraint device is arranged to:
receive a signal output by the restraint management apparatus, and in response to a signal received from the restraint management apparatus, apply a restraint to the subject or alter a parameter of a restraint applied to the subject.
11. A restraint system according to claim 10, wherein the restraint device is further arranged to:
receive a manually- input restraint parameter;
in response in response to a signal received from the restraint management apparatus, apply a restraint to the subject or alter a parameter of a restraint applied to the subject; and
transmit a signal containing the manually- input restraint parameter to the restraint management apparatus.
12. A method for use in restraining a subject, the method comprising:
receiving one or more types of sensor data; determining a status of a subject based on the received sensor data, wherein the status comprises a distance between a body part of the subject and a selected object in the environment around the subject and/or velocity of a body part of the subject towards a selected object in the environment around the subject; and
determining, based on the determined subject status, a restraint parameter for a restraint device configured to restrain the body part of the subject, wherein the restraint device is configured to provide a resistance to movement of the body part, and wherein the restraint parameter indicates a level of resistance to movement to be provided by the restraint device, and wherein the restraint parameter is determined such that the level of resistance to movement provided by the restraint device increases as the distance between the body part and the selected object decreases and/or as the velocity of the body part towards the selected object increases.
13. A method according to claim 12, wherein the received sensor data comprises one or more of:
an image including a subject;
an image including a subject and one or more objects in the vicinity of the subject;
audio data including sounds generated by a subject;
accelerometer data obtained by an accelero meter worn by a subject;
measurements of one or more physiological characteristics of a subject;
restraint-related data obtained by a sensor associated with a restraint device; pressure data obtained by a bed sensor;
bed rail position data obtained by a bed rail sensor.
14. A method according to claim 12 or claim 13, wherein the received sensor data comprises an image including a subject and one or more objects, and wherein determining a status of a subject comprises:
determining a position of a body part of the subject;
determining a position of an object in the image;
calculating a distance between the body part and the object; and comparing the calculated distance to a predetermined threshold.
15. A method according to claim 12 or claim 13, wherein the received sensor data comprises a series of images including a subject and one or more objects, and wherein determining a status of a subject comprises:
determining a position of a body part of the subject in each of the images; determining a position of an object in each of the images;
calculating a distance between the body part and the object in each of the images;
calculating a velocity of the body part towards the object from the calculated distances; and
comparing the calculated velocity to a predetermined velocity threshold.
PCT/EP2016/055483 2015-03-23 2016-03-15 Restraint management WO2016150750A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
EP16711556.7A EP3274884B1 (en) 2015-03-23 2016-03-15 Restraint management
JP2017549428A JP6730304B2 (en) 2015-03-23 2016-03-15 Restraint management
CN201680017575.1A CN107408159A (en) 2015-03-23 2016-03-15 Constraints Management
US15/552,906 US10945667B2 (en) 2015-03-23 2016-03-15 Restraint management

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP15160212.5 2015-03-23
EP15160212 2015-03-23

Publications (1)

Publication Number Publication Date
WO2016150750A1 true WO2016150750A1 (en) 2016-09-29

Family

ID=52780839

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2016/055483 WO2016150750A1 (en) 2015-03-23 2016-03-15 Restraint management

Country Status (5)

Country Link
US (1) US10945667B2 (en)
EP (1) EP3274884B1 (en)
JP (1) JP6730304B2 (en)
CN (1) CN107408159A (en)
WO (1) WO2016150750A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11424028B2 (en) * 2018-04-19 2022-08-23 University Of Florida Research Foundation, Incorporated Method and apparatus for pervasive patient monitoring
JP7140264B2 (en) * 2019-02-18 2022-09-21 日本電気株式会社 Abnormality determination device, its operation method, and program

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5649541A (en) * 1996-03-14 1997-07-22 Stuckey; Judith A. Patient restraint
US20070288263A1 (en) * 2005-12-09 2007-12-13 Valence Broadband, Inc. Methods and systems for monitoring quality and performance at a healthcare facility
US20140345060A1 (en) * 2012-05-22 2014-11-27 Hill-Rom Services, Inc. Systems, methods, and devices for the treatment of sleep disorders
US20140379224A1 (en) * 2013-06-20 2014-12-25 Elwha Llc Systems and methods for adjusting the position of a wheelchair occupant

Family Cites Families (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5826216A (en) * 1997-02-07 1998-10-20 Hittite Microwave Corp. Anticipatory collision sensor system
US6160478A (en) * 1998-10-27 2000-12-12 Sarcos Lc Wireless health monitoring system
US7319400B2 (en) * 2003-10-17 2008-01-15 Bed-Check Corporation Method and apparatus for monitoring a restraint device
CN1658116A (en) * 2005-03-19 2005-08-24 李兵 Health-care arm support
GB2424982A (en) * 2005-04-07 2006-10-11 Autoliv Dev Seatbelt pretensioner dependent on steering rate
US20080029104A1 (en) * 2006-08-07 2008-02-07 The Board Of Regents Of The University Of Texas System Arm movement restraining and permission cord for patients receiving cardiac pacemakers and similar implants
US7389552B1 (en) * 2007-12-31 2008-06-24 Monster Medic, Inc. Ambulance cot system
US20090275807A1 (en) * 2008-05-02 2009-11-05 General Electric Company Method for managing alarms in a physiological monitoring system
CN101590838B (en) * 2009-07-06 2010-12-08 中国科学技术大学 Lap-belt retractor
CN102043577B (en) * 2009-10-26 2012-12-19 乔山健康科技股份有限公司 Human-machine interface method of sports apparatus and human-machine interface device
US8417662B2 (en) * 2010-02-18 2013-04-09 The University Of Utah Research Foundation Adjustable alert rules for medical personnel
US8799011B2 (en) 2010-02-19 2014-08-05 Hill-Rom Services, Inc. Patient room and bed management apparatus and system
US10271768B1 (en) * 2010-12-23 2019-04-30 Jogohealth Inc. Method for determining rehab protocol and behavior shaping target for rehabilitation of neuromuscular disorders
CN103517669B (en) * 2011-03-11 2016-04-20 太空实验室健康护理有限公司 The method and system of multiparameter administrative alert grade is determined during patient monitoring
US20130127620A1 (en) 2011-06-20 2013-05-23 Cerner Innovation, Inc. Management of patient fall risk
US8206047B1 (en) * 2011-06-24 2012-06-26 TouchFire, Inc. Keyboard overlay for optimal touch typing on a proximity-based touch screen
JP6053802B2 (en) 2011-10-07 2016-12-27 コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. A monitoring system that monitors patients and detects patient delirium
US8941487B2 (en) * 2012-03-09 2015-01-27 Salutron, Inc. Transferring a mobile tag using a light based communication handshake protocol
US9041530B2 (en) * 2012-04-18 2015-05-26 Qualcomm Incorporated Biometric attribute anomaly detection system with adjusting notifications
US10265022B2 (en) * 2012-10-30 2019-04-23 Nokia Technologies Oy Determining biometrics utilizing a display-embedded distance-measuring sensor
US11229364B2 (en) * 2013-06-14 2022-01-25 Medtronic, Inc. Patient motion analysis for behavior identification based on video frames with user selecting the head and torso from a frame
US10279201B2 (en) * 2014-07-17 2019-05-07 Elwha Llc Monitoring and treating pain with epidermal electronics
US10390755B2 (en) * 2014-07-17 2019-08-27 Elwha Llc Monitoring body movement or condition according to motion regimen with conformal electronics
US10835450B2 (en) * 2016-12-30 2020-11-17 Stryker Corporation CPR chest compression system periodically reminding attendant to check patient

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5649541A (en) * 1996-03-14 1997-07-22 Stuckey; Judith A. Patient restraint
US20070288263A1 (en) * 2005-12-09 2007-12-13 Valence Broadband, Inc. Methods and systems for monitoring quality and performance at a healthcare facility
US20140345060A1 (en) * 2012-05-22 2014-11-27 Hill-Rom Services, Inc. Systems, methods, and devices for the treatment of sleep disorders
US20140379224A1 (en) * 2013-06-20 2014-12-25 Elwha Llc Systems and methods for adjusting the position of a wheelchair occupant

Also Published As

Publication number Publication date
US20180032694A1 (en) 2018-02-01
CN107408159A (en) 2017-11-28
US10945667B2 (en) 2021-03-16
EP3274884A1 (en) 2018-01-31
JP2018515161A (en) 2018-06-14
JP6730304B2 (en) 2020-07-29
EP3274884B1 (en) 2020-05-27

Similar Documents

Publication Publication Date Title
US10004447B2 (en) Systems and methods for collecting and displaying user orientation information on a user-worn sensor device
JP6599580B1 (en) User monitoring system
JP6249943B2 (en) Fever detection device
EP3378386A1 (en) Method and apparatus for determining a health status of an infant
RU2706973C2 (en) Falling detection system and method
CN110366385B (en) Patient monitoring system and method with location-specific contextual alerts
EP3432772B1 (en) Using visual context to timely trigger measuring physiological parameters
JP2017519596A (en) Individual tourniquet system
US10945667B2 (en) Restraint management
CA3019931A1 (en) Pressure ulcer detection methods, devices and techniques
JP6164202B2 (en) Display control device and relaxation support system
US11495351B2 (en) Health monitoring system and method thereof
US10426344B2 (en) Medical service support device, medical service support system and medical service support method
US20220345799A1 (en) Sensor device, sensing method, and program storage medium
KR20170127394A (en) Patient monitoring system built in the algorithm for the management of the proper sedation monitoring and patient monitoring method using the same
JP2022122257A (en) State management method, program, and dialysis system
JP2019513457A (en) Wrinkle detection method, device and technique
KR20170124334A (en) Patient monitoring system built in the algorithm for the management of the proper sedation monitoring and patient monitoring method using the same

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16711556

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2017549428

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

REEP Request for entry into the european phase

Ref document number: 2016711556

Country of ref document: EP