WO2008042219A2 - Interface utilisateur et identification intégrés à un dispositif médical, systèmes et procédés - Google Patents
Interface utilisateur et identification intégrés à un dispositif médical, systèmes et procédés Download PDFInfo
- Publication number
- WO2008042219A2 WO2008042219A2 PCT/US2007/020847 US2007020847W WO2008042219A2 WO 2008042219 A2 WO2008042219 A2 WO 2008042219A2 US 2007020847 W US2007020847 W US 2007020847W WO 2008042219 A2 WO2008042219 A2 WO 2008042219A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- medical device
- main unit
- gesture
- set forth
- medical
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/0304—Detection arrangements using opto-electronic means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/014—Hand-worn input/output arrangements, e.g. data gloves
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0346—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0354—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
- G06F3/03542—Light pens for emitting or receiving light
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0354—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
- G06F3/03545—Pens or stylus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/038—Control and interface arrangements therefor, e.g. drivers or device-embedded control circuitry
- G06F3/0386—Control and interface arrangements therefor, e.g. drivers or device-embedded control circuitry for light pen
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H40/00—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices
- G16H40/60—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices
- G16H40/63—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices for local operation
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B2017/00017—Electrical control of surgical instruments
- A61B2017/00207—Electrical control of surgical instruments with hand gesture control or hand gesture recognition
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2560/00—Constructional details of operational features of apparatus; Accessories for medical measuring apparatus
- A61B2560/02—Operational features
- A61B2560/0266—Operational features for monitoring or limiting apparatus function
- A61B2560/0276—Determining malfunction
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0002—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/145—Measuring characteristics of blood in vivo, e.g. gas concentration, pH value; Measuring characteristics of body fluids or tissues, e.g. interstitial fluid, cerebral tissue
- A61B5/1455—Measuring characteristics of blood in vivo, e.g. gas concentration, pH value; Measuring characteristics of body fluids or tissues, e.g. interstitial fluid, cerebral tissue using optical sensors, e.g. spectral photometrical oximeters
Definitions
- the present invention relates generally to medical devices and, more particularly, to user interfaces and identification systems integrated with medical devices.
- pulse oximetry For example, one technique for monitoring certain physiological characteristics of a patient is commonly referred to as pulse oximetry, and the devices built based upon pulse oximetry techniques are commonly referred to as pulse oximeters.
- Pulse oximetry may be used to measure various blood flow characteristics, such as the blood-oxygen saturation of hemoglobin in arterial blood, the volume of individual blood pulsations supplying the tissue, and/or the rate of blood pulsations corresponding to each heartbeat of a patient.
- Pulse oximeters and other medical devices are typically mounted on stands that are positioned around a patient's bed or around an operating room table.
- a caregiver desires to command the medical device (e.g., program, configure, and so-forth) they manipulate controls or push buttons on the monitoring device itself.
- the monitoring device typically provides results or responses to commands on a Liquid Crystal Diode (“LCD”) screen mounted in an externally visible position within the medical device.
- LCD Liquid Crystal Diode
- this conventional configuration has several disadvantages.
- this conventional configuration relies upon physical contact with the monitoring device to input commands (e.g., pushing a button, turning a knob, and the like).
- Such physical contact raises several concerns. Among these concerns are that in making contact with the medical device, the caregiver may spread illness or disease from room to room. More specifically, a caregiver may accidentally deposit germs (e.g., bacteria, viruses, and so forth) on the medical device while manipulating the device's controls. These germs may then be spread to the patient when a subsequent caregiver touches the medical device and then touches the patient. Moreover, if medical devices are moved from one patient room to another, germs transferred to the medical device via touch may be carried from one patient room to another. Even in operating rooms where medical devices are typically static, germs may be transferred onto a monitoring device during one surgery and subsequently transferred off the medical device during a later performed surgery.
- germs e.g., bacteria, viruses, and so forth
- medical devices that rely on physical contact for command input may create clutter the caregiver's workspace. For example, because the medical device must be within an arm's length of the caregiver, the medical device may crowd the caregiver — potentially even restricting free movement of the caregiver.
- caregivers may have difficulty manipulating controls with gloved hands. For example, it may be difficult to grasp a knob or press a small button due to the added encumbrance of a latex glove.
- HIPPA Health Insurance Portability and Accountability Act
- conventional techniques for outputting medical data also have several potential drawbacks.
- conventional techniques for displaying outputs rely on LCD screens mounted on the medical device itself. Besides constantly consuming power, these LCD screens must be large enough to be visually accessed by a caregiver. As such, the conventional LCD screens employed in typical medical devices also may be a barrier towards miniaturization of the medical device. Further, conventional screen-based output techniques may be impersonal to the patient and may lack configurability by the caregiver.
- Fig. 1 is a diagrammatical representation of a medical device including a gesture interface in accordance with one embodiment of the present invention
- Fig. 2 is a flowchart illustrating an exemplary technique for processing a gesture command in accordance with one embodiment of the present invention
- Fig. 3 is a diagrammatical representation of a medical device including another gesture interface in accordance with one embodiment of the present invention.
- Fig. 4 is a diagrammatical representation of a pulse oximeter configured with a pen-based interface in accordance with one embodiment of the present invention
- Fig. 5 is a diagram of an operating room and a medical device including a laser- based interface in accordance with one embodiment of the present invention
- Fig. 6 is a diagrammatical representation of a remote control for interfacing with a medical device, in accordance with one embodiment of the present invention
- Fig. 7 is a diagrammatical representation of a remote control for interfacing with a medical device incorporated into a badge holder in accordance with one embodiment of the present invention
- Fig. 8 is a bottom view of the badge holder of Fig. 7 in accordance with one embodiment of the present invention.
- Fig. 9 illustrates a patient room and a medical device configured to interface with a personal digital assistant in accordance with one embodiment of the present invention
- Fig. 10 is a diagram of a patient, a caregiver, and a medical device configured to output to a personal caregiver display in accordance with one embodiment of the present invention
- Fig. 1 1 is an enlarged view of the caregiver of Fig. 10 further including a microphone to interface with the medical device of Fig. 10 in accordance with one embodiment of the present invention
- Fig. 12 A is a diagram of an exemplary hospital room configured to identify caregivers or patients in accordance with one embodiment of the present invention
- Fig. 12B is an enlarged view of a doorway in a hospital room configured to identify caregivers or patients in accordance with one embodiment of the present invention
- Fig. 13 is a diagram of a caregiver identifier configured to enable caregiver or patient identification in accordance with one embodiment of the present invention
- Fig. 14 is a flow chart illustrating an exemplary technique for identifying caregivers or patients in accordance with one embodiment of the present invention.
- Fig. 15 is a block diagram of an exemplary system for identifying caregivers or patients in accordance with one embodiment of the present invention.
- the medical device 10 comprises a pulse oximeter.
- the medial device 10 may include a main unit 12 that houses hardware and/or software configured to calculate various physiological parameters or produce various medical outputs.
- the main unit 12 may include a display 14 for displaying the calculated physiological parameters, such as oxygen saturation or pulse rate, to a caregiver or patient.
- the display 14 may be omitted from the main unit 12.
- the medical device 10 may also include a sensor 16 that may be connected to a body part (e.g., finger, forehead, toe, or earlobe) of a patient or a user.
- the sensor 16 may be configured to emit signals or waves into the patient's or user's tissue and detect these signals or waves after dispersion and/or reflection by the tissue.
- the sensor 16 may be configured to emit light from two or more light emitting diodes ("LEDs") into pulsatile tissue ⁇ e.g., finger, forehead, toe, or earlobe) and then detect the transmitted light with a light detector ⁇ e.g., a photodiode or photo-detector) after the light has passed through the pulsatile tissue.
- LEDs light emitting diodes
- the amount of transmitted light that passes through the tissue generally varies in accordance with a changing amount of blood constituent in the tissue and the related light absorption.
- the heart pumps an incremental amount of arterial blood into the pulsatile tissue, which then drains back through the venous system.
- the amount of light that passes through the blood-perfused tissue varies with the cardiac-induced cycling arterial blood volume. For example, when the cardiac cycle causes more light-absorbing blood to be present in the tissue, less light travels through the tissue to strike the sensor's photo-detector.
- These pulsatile signals allow the medical device 10 to measure signal continuation caused by the tissue's arterial blood, because light absorption from other tissues remains generally unchanged in the relevant time span.
- the senor 16 may take other suitable forms beside the form illustrated in Fig. 1.
- the sensor 16 may be configured to be clipped onto a finger or earlobe or may be configured to be secured with tape or another static mounting technique.
- the sensor 16 may be connected to the main unit 12 via a cable 18 and a connector 20.
- the medical device 10 may also include a speaker 22 to broadcast alarms or alerts.
- the pulse oximeter main unit 12 may also include an integral camera 24.
- the integral camera 24 may be configured to receive gesture commands from a caregiver or user that can be processed into commands for the medical device 10.
- Fig. 1 illustrates the integral camera 24 as being located on a top surface of the main unit 12, it will be appreciated that in alternate embodiments, the integral camera 24 may be located at another suitable location on or within the main unit 12, such as the front or side facades.
- an external camera such as a universal serial bus (“USB”) web camera
- USB universal serial bus
- the external camera may also be wirelessly connected to the main unit 12 via radio, infrared, or optical signals.
- WLAN wireless local area networking
- multiple cameras may be used to reduce the effects of parallax and occlusions.
- the cameras may be all external cameras, all integral cameras, or a combination of external and integral cameras.
- Fig. 2 illustrates a flowchart of technique 30 for processing a gesture command in accordance with one embodiment.
- the technique 30 may be executed by the medical device 10.
- other medical devices such as a respirator, cardiac monitor, or multi-parameter monitoring system, may execute the technique 30.
- the technique 30 may begin by receiving a gesture.
- the camera 24 of medical device 10 may be configured to detect hand gestures.
- the medical device 10 first receives one or more images of the gesture (block 32) via the camera 24 and then processes the gesture (block 34).
- the camera 24 may be an analog camera that converts the gesture into an analog signal and feeds it into a digitizer board, or the camera 24 may be a digital camera that records the gesture as a digital signal.
- the camera 24 may be configured to detect other gestures originating from another bodily motion or state, such as arm gestures, finger pointing, hand poses, or facial expressions.
- the gesture processing may be performed by a gesture processing system integrated into the medical device 10. For example, during processing, images captured from the gesture may be normalized, enhanced, or transformed, and then features may be extracted from the images. Next, the processed gesture may be compared to a gesture database, as indicated by block 36.
- the gesture database may be pre-populated or programmed with a plurality of feature combinations that are associated with commands for the medical device 10. For example, feature combinations associated with the gesture command "turn alarm off may be associated with a command for the medical device 10 to silence an alarm. However, in alternate embodiments the feature combinations may be programmed into the gesture database using a gesture training program.
- the gesture processing may be located in an external central processing unit connected to the medical device 10 via a cable or by wireless technology such as radio, infrared, or optical signals.
- a command associated with the gesture may be identified, as indicated by block 38.
- a hand gesture consisting of passing a hand over the camera from right to left with the palm facing the camera may be programmed into the gesture database to correspond with the command "turn alarm off.”
- the command may be executed, as indicated by block 40.
- the command to turn off the alarm may be transmitted to a medical device control system which would turn off the alarm.
- the medical device 50 which may be a pulse oximeter, may include a tracking glove 52.
- the tracking glove 52 may be configured to receive gesture commands from a caregiver 62 or user. As with the gesture command described in regard to Fig. 2, these gesture commands can be processed into commands for the medical device 50.
- the tracking glove 52 may include a battery pack 56 connected to the tracking glove 52 via a cable. Although the battery pack is worn on the forearm in this embodiment, in alternative embodiments the battery pack may be located in other locations such as around the waist of the caregiver 62. Additionally, in other embodiments, the tracking glove 52 may be replaced by another tracking device such as a finger sensor. In yet another embodiment, the tracking glove 52 may have a light emitting diode ("LED") located on the glove and a software programmable switch to permit other functions to be directly programmed into the glove. For example, a button may be included on the glove that can be programmed so that when a caregiver presses the button an alarm on the medical device SO is silenced.
- LED light emitting diode
- the caregiver 62 may make hand gestures while wearing the tracking glove 52.
- the tracking glove 52 may then record the movement (i.e., the gesture) and transmit the gesture to the medical device 50 via a wireless receiver 60 connected to the medical device 50.
- the tracking glove 52 may communicate with a wireless receiver integrated into the main unit 12 or may be connected to the medical device 50 via a cable such as a fiber optic cable or a serial cable.
- the medical device 50 may be configured to interpret the tracking glove 52 movement and execute a command associated with the movement. For example, a hand movement, such as making a fist, may be associated with the command "turn alarm off.” As such, when the caregiver 62 makes a fist while wearing the tracking glove 52, the medical device 50 may interpret the movement and sends a signal to the medical device 50 to silence an alarm.
- a hand movement such as making a fist
- the medical device 50 may interpret the movement and sends a signal to the medical device 50 to silence an alarm.
- the medical device 50 may include calibration software which may allow a caregiver 62 to program movement combinations into the gesture database within the medical device.
- the gesture database may be located in an external central processing unit connected to the medical device 50 via a cable or by wireless technology such as radio, infrared, or optical signals.
- an exemplary medical device configured with a pen- based interface is illustrated and generally designated by the reference numeral 70.
- the medical device 70 may include a stylus 72.
- the caregiver 62 may use the stylus 72 to control the medical device 70.
- the medical device 70 may have a separate display screen 74 connected to the main unit 12 via a cable or wireless means such as radio, infrared, or optic signals.
- the display screen 74 may be a touch screen with selection boxes corresponding to medical device commands. For example, when the caregiver 62 touches the stylus 72 to the selection box corresponding to "turn alarm off," the display screen 74 may transmit a signal to the main unit 12 which silences the alarm. In an alternate embodiment, the caregiver 62 may touch the screen directly without using the stylus. In still other embodiments, the stylus 72 may be used to touch selection boxes directly on the medical device 70, and the separate display 74 may be omitted.
- the stylus 72 may be used to draw symbols or characters representative of medical device commands on the display screen 74.
- the medical device 70 main unit 12 may include symbol recognition software which recognizes the symbols drawn on the display 74 and executes commands corresponding to the recognized symbols. For example, the letter
- the symbol recognition software may interpret the symbol, and the medical device 70 may lower the alarm limit by a predetermined amount.
- the symbol recognition software may be pre-populated or programmed with a plurality of symbols associated with medical device commands. In alternate embodiments, the symbol recognition software may include a calibration program to allow the caregiver 62 to associate symbols with medical device commands.
- the stylus 72 may include an ultrasound transmitter.
- the ultrasound transmitter may be configured to transmit movements of the stylus 72 back to the medical device 70 or another suitable receiver.
- the movements of the stylus 72 may be tracked by one or more sensors positioned around an operating room and coupled to the medical device 70.
- the medical device 80 may include a laser wand 84 and/or the display 64.
- the caregiver 62 may use the laser wand 84 to control the medical device 80.
- the operating room shown in Fig. 5 is merely one possible application of the medical device 80.
- the medical device 80 may be employed in patients' rooms, doctors' offices, or other suitable locations.
- the medical devices described in regard to Figs. 1, 3, and 4, as well as those described below may be employed in each of these locations as well.
- the caregiver 62 may be able to use the laser wand 84 to position a cursor on the display 64.
- the caregiver 62 may focus a laser pointer dot on the display 64.
- a location of a laser pointer dot can be translated to the cursor position on a display 64.
- the laser pointer dot may alternatively be focused on the display 14.
- the display 64 (or the display 14) may employ a camera, such as the camera 24 discussed above, to detect the laser pointer dot.
- the camera may be internal to the display 64 or may be externally connected to it via a cable or wirelessly.
- other suitable laser pointer detection techniques may be employed.
- the display 64 may contain a plurality of selection boxes or regions corresponding to commands for medical device 80.
- the display 64 may contain a selection box for the command "turn alarm off.”
- the software within the medical device 80 may first position the cursor at the selection box location.
- the software within the medical device 80 may then select the box and execute the command associated with the selection box. In this example, the software may then silence the alarm.
- the laser wand 84 may have an integrated selection button. Once the caregiver 62 has positioned the cursor on the selection box, the caregiver 62 may then push the button to select the box and execute the pulse oximeter command associated with the box.
- the integrated selection button may employ standard remote control technology, such as transmitting an infrared signal to an infrared receiver integrated into the medical device 80. In alternate embodiments, an external receiver connected to the medical device 80 via a cable may be used.
- the laser wand 84 may allow the medical device 80 to be controlled from a distance by the caregiver 62. Consequently, the medical device 80 may be placed at a location away from the patient 82 allowing the caregivers 62 more room to maneuver.
- each of the caregivers 62 may have their own laser wand 84 to further reduce the risks of cross-contamination.
- the medical device 80 may be controlled using a remote control style wand 90, as illustrated in Fig. 6.
- the wand 90 may contain a plurality of buttons each programmed to correspond to one or more medical device commands.
- the buttons may be programmed as follows: the button labeled "1" 96 may be programmed to correspond to the command “Raise Alarm Limit;” the button “2" 98 may be programmed to correspond to the command “Lower Alarm Limit;” the button “3” 100 may be programmed to correspond to the command “Reset Alarm Limits;” and the button “4" 102 may be programmed to correspond to the command "Turn Alarm Off.” It will be appreciated that these commands are exemplary. Although the buttons 96-102 are shown in Fig.
- buttons 6 may be customized with different shapes, sizes, and labels. Additionally, the number of buttons present on the wand 90 may vary.
- the wand 90 also may contain a pen-style clip for attaching the wand 90 to the caregiver 62.
- the wand 90 may contain a light emitting diode ("LED") 92 that transmits light pulses or infrared signals corresponding to a medical device command.
- LED light emitting diode
- an integrated circuit within the wand 90 may send a command to the LED 92.
- the LED 92 may then send out a signal corresponding to this command.
- a receiver integrated into the medical device may receive the signal and respond by raising the alarm limit by a predetermined unit.
- the LED transmitter 92 may alternatively be replaced by a radio frequency ("rf ') transmitter.
- the medical device 80 may include an integrated rf receiver.
- the rf transmitter may employ the Bluetooth radio frequency standard or other suitable standard.
- the technology of the wand 90 may also be incorporated in other packages.
- it may be incorporated into a badge holder, as illustrated in Fig. 7.
- the badge holder 110 in addition to holding a caregiver's badge 1 12, may also contain the several command buttons 96- 102.
- the badge holder 110 may also contain the transmitter 92 on the bottom of the badge holder 1 10.
- the transmitter 92 may be located on another facade such as the top, front, or sides of the badge holder 110.
- the control buttons 96-102 may be of various shapes and sizes and be located on other facades of the badge holder 110.
- the medical device 130 may include a PDA 134.
- the caregiver 62 may use the PDA 134 to control the medical device 130.
- the PDA 134 may be configured to present the caregiver 62 with one or more buttons or selectable locations on its screen that correspond to medical device controls or commands. Accordingly, when the caregiver 62 selects one of these controls or commands, the PDA 134 may be configured to transmit this control or command back to the medical device 130, which may subsequent execute the control or command.
- the PDA 134 may be configured to generate a volume control display for the medical device 130. Upon accessing this volume display on the PDA 134, the caregiver may adjust the volume of the medical device 130 up or down.
- the PDA 134 enables the caregiver 62 to control medical device 130 without physically touching or manipulating it.
- the PDA 134 may also supplement a display on the medical device 130.
- the PDA 134 may be configured to mirror or reproduce some or all of the contents displayed on the medical device's 130 internal display. In this way, the medical device 130 could advantageously be located away from the patient bed 132 or the caregiver 62, possibly even out of sight, as the inputs and outputs to the medical device 130 can be supported by the PDA 134.
- the PDA 134 may be configured to interface with a plurality of medical devices 130 in a plurality of patient rooms 131.
- a hospital may issue each of each caregivers 62 their own PDA 134, which they may use to access and/or control medical devices within a plurality of patient rooms.
- the caregiver 62 may use their PDA 134 to access one or more medical devices within a first patient's room and then use the same PDA 134 to access medical devices within a subsequent patient's room. In this way, the caregiver 62 may access and/or control medical devices within a plurality of patient rooms without ever touching the actual medical devices - substantially decreasing the chances of cross-contamination.
- the PDA 134 may supplement or replace the internal screen on the medical device 130.
- Fig. 10 is a diagram of a patient 141, the caregiver 62, and a medical device 140 configured to output to a personal caregiver display 142 in accordance with one embodiment.
- the caregiver personal display 142 may include a pair of glasses or other suitable wearable optics or eyewear ⁇ e.g., a monocular) which may be configured to display outputs from the medical device 140.
- the caregiver personal display 142 may be configured to create a transparent or semi-transparent image that the caregiver 62 may be able to view while still being able to see the patient 141.
- the caregiver personal display may include a pair of glasses with an integral liquid crystal display (“LCD”) that may be configured to display a pleth signal
- the caregiver personal display 142 effectively creates a "heads-up" display for the caregiver 62, allowing them to see the pleth signal 144 or other suitable medical information as if it were floating in front of them.
- caregiver personal display 142 is merely one potential embodiment of a suitable caregiver personal display.
- the caregiver personal display may be a video display mounted on a pair of glasses or other mount, which the caregiver 62 may view by shifting his or her focus towards the display.
- medical information in this embodiment may not appear transparent to the caregiver 62, the caregiver 62 may still able to readily access information from the medical device 140 without having the medical device 140 within visual range of the caregiver 62.
- the caregiver personal display may also include a speaker 146 to enable the caregiver 62 to hear alarms or alerts from the medical device 140.
- the speaker 146 enables the caregiver 62, who is monitoring medical device 140, to hear alerts or alarms without the alarms or alerts bothering other caregiver 62, who may be focused on other activities.
- the caregiver personal display 142 may also include a microphone 148 to enable voice control of the medical device 140, as further described in commonly assigned U.S. Patent Application Serial No. 1 1/540,457 entitled SYSTEM AND METHOD FOR INTEGRATING VOICE WITH A MEDICAL DEVICE and filed on September 29, 2006, which is hereby incorporated by reference.
- Fig. 12A is a diagram of an exemplary patient room 160 configured to identify caregivers or patients in accordance with one embodiment.
- the hospital room .160 may include a patient bed 162, a medical device 164, and a doorway 166.
- the hospital room 160 may also include the patient 141 and the caregiver 62.
- the patient 141 may be located in the bed 162 with the caregiver 62 positioned over the patient 141 and in general proximity with the medical device 164.
- the medical device 164 may include the main unit 12, the display 14, and/or the display 64. Moreover, the medical device 164 may be configured to work in conjunction with an identification ("ID") tag, such as a caregiver ID 168 and/or a patient ID 170 to identify caregivers and/or patients within the hospital room 160. More specifically, in one embodiment, the medical device 164 may be coupled to door sensors 172 A and 172B, which are located in close proximity to the door 166 and configured to detect when the caregiver ID 168 and/or the patient ID 170 pass through the doorway 166.
- ID identification
- the medical device 164 may be coupled to door sensors 172 A and 172B, which are located in close proximity to the door 166 and configured to detect when the caregiver ID 168 and/or the patient ID 170 pass through the doorway 166.
- the door sensors 172a and 172b may be configured to detect when the caregiver ID passes through the doorway 166 ⁇ e.g., when a caregiver enters or exits the patient room 160).
- the door sensors 172a and 172b may be configured to detect when the patient ID 170 passes through the doorway 166 ⁇ e.g., the patient 141 walks or is pushed through the doorway 166).
- the door sensors 172a and 172b may include radio frequency ("rf ') sensors configured to detect an rf transmitter within the caregiver ID 168 and/or the patient ID 170.
- rf ' radio frequency
- the caregiver ID 168 including an rf ID tag 174.
- the rf ID tag 174 may be a passive rf ID configured to receive transmissions from the door sensors 172a or 172b and to broadcast an identifying signal in response.
- the door sensors 172a or 172b may detect this identifying signal, and, thus, identify/detect the entry or exit of the caregiver 62 and/or the patient 141.
- the caregiver ID may be an active ID (e.g., a Bluetooth enabled cell phone).
- the door sensors 172a and 172b may be located elsewhere besides the doors.
- the sensors 172a and 172b may be located in the ceiling of the patient room 160 and configured to detect when the sensors 168 and/or 170 enter are located in the patient room 160
- Fig. 14 is a flow chart illustrating an exemplary technique 210 for identifying caregivers or patients in accordance with one embodiment.
- the technique 210 will be described in conjunction with Fig. 15 which illustrates a block diagram o the medical device 164 in accordance with one embodiment.
- the technique 210 may begin by detecting an ID tag.
- the door sensors 172 may detect the caregiver ID 168 or the patient ID 170.
- the medical device 164 may alternatively or additionally include a bed sensor 250 to detect the ID tag. More specifically, the bed sensor 250 may be mounted on the patient bed 162 and configured to detect the patient ID 170. The bed sensor 250 may be particularly advantageous in hospital rooms including multiple patient beds 162, as the door sensors 172 may not be able to determine which of the plurality of patient beds 162 a particular patient is occupying. Moreover, the bed sensor 250 may also be configured to detect the patient ID 170 and to communicate patient identification information to the medical device 164, as set forth in further detail below.
- the technique 210 may include reading the detected ID tag, as indicated in block 214.
- an ID reader 252 may be configured to read the identity information from the caregiver ID 168 and/or the patient ID 170.
- the technique 210 may include sending the identity information from the ID tag to an ID recognition system, as indicated in block 216.
- the ID reader 252 may transmit the identity information to an ID recognition system 254.
- the technique 210 may include determining an individual type of the detected ID tag, as indicated by block 218.
- the technique may include determining whether the detected ID tag corresponds to the caregiver 62 or the patient 141.
- the ID recognition system 254 may make this determination based on an ID database 256, which includes information regarding a plurality of ID tags and the individual type corresponding to each of the plurality of ID tags.
- the individual type may be encoded on the caregiver ID 168 or the patient ID 170 and communicated to the ID recognition system 254 via the door sensor 172 and/or the bed sensor 250.
- the technique 210 may include multiple branches for various suitable individual types.
- the technique 210 may respond differently to different types of caregivers, such as doctors, nurses, orderlies, and so-forth.
- the technique 210 may next include unlocking access to the medical device at an appropriate permissions level.
- an access control system in the medical device 164 may be configured to unlock the medical device 164 and allow the medical device control system 226 to execute instructions and/or commands commensurate with the permission level of the caregiver 62.
- the technique 210 may then continue to allow the caregiver 62 to execute commands until the same ID tag is again detected by the door sensor 172 or the bed sensor 250, as indicated by block 228 (i.e., the caregiver leaves the patient room 160).
- the door sensors 172a and 172b may be configured to detect when the caregiver ID 168 leaves the proximity of the sensor.
- the door sensors 172a and 172b may be configured to detect when the caregiver ID 168 is located within a certain distance of the door sensors 172a and 172b (i.e., the caregiver is located in the patient room 160).
- the technique 210 may include locking further access to the medical device 164 to prevent the patient 141 or other unauthorized individuals from adjusting the medical device 164 in the absence of the caregiver 162 (block 230).
- the technique 210 enables the medical device 164 or other suitable medical device automatically unlock when the caregiver 62 enters the patient room 160, to accept commands freely from the caregiver 62 while they are in the room, and then to relock automatically when the caregiver 62 leaves the patient room.
- the medical device 164 may be configured to record which caregiver 62 gave which commands to the medical device 164, because each caregiver in a hospital may be assigned a unique caregiver ID 168. In this way, it may be possible for a hospital to reconstruct patient treatment history, if desired.
- the technique 210 may include displaying a patient ID on a display of the medical device 164, as indicated by block 220.
- a patient identification system 258 within the medical device 164 may be configured to display the patient ID on the display 14.
- the medical device 164 may also be configured to annotate any patient medical data subsequently stored by the medical device 164 with the patient information.
- the technique 210 may include clearing the patient information from the medical device 164, as indicated by blocks 222 and 224. Accordingly, the medical device 164 may be able to automatically identify the identity of patients being monitored or treated without the need for caregivers to manually enter this information into the medical device 164. Advantageously, this reduces the chances of cross- contamination and automates one additional caregiver function.
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- General Business, Economics & Management (AREA)
- Public Health (AREA)
- Primary Health Care (AREA)
- Medical Informatics (AREA)
- General Health & Medical Sciences (AREA)
- Epidemiology (AREA)
- Business, Economics & Management (AREA)
- Measuring And Recording Apparatus For Diagnosis (AREA)
- Medical Treatment And Welfare Office Work (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
L'invention concerne des systèmes et des procédés s'appliquant à une interface utilisateur et des systèmes d'identification intégrés à un dispositif médical. Plus spécifiquement, selon un mode de mise en oeuvre, l'invention concerne un procédé permettant de commander un dispositif médical, ce procédé consistant à recevoir une commande gestuelle, traiter la commande gestuelle reçue, la comparer à une base de données de commandes gestuelles comprenant plusieurs commandes gestuelles et plusieurs commandes du dispositif médical correspondant à chaque pluralité de commandes gestuelles, identifier une des commande gestuelles à partir de la base de données de commandes gestuelles correspondant à la commande gestuelle reçue et exécuter la commande.
Applications Claiming Priority (8)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/540,895 US20080097176A1 (en) | 2006-09-29 | 2006-09-29 | User interface and identification in a medical device systems and methods |
US11/540,895 | 2006-09-29 | ||
US11/540,385 US7706896B2 (en) | 2006-09-29 | 2006-09-29 | User interface and identification in a medical device system and method |
US11/540,016 | 2006-09-29 | ||
US11/540,016 US7698002B2 (en) | 2006-09-29 | 2006-09-29 | Systems and methods for user interface and identification in a medical device |
US11/541,069 US20080097177A1 (en) | 2006-09-29 | 2006-09-29 | System and method for user interface and identification in a medical device |
US11/540,385 | 2006-09-29 | ||
US11/541,069 | 2006-09-29 |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2008042219A2 true WO2008042219A2 (fr) | 2008-04-10 |
WO2008042219A3 WO2008042219A3 (fr) | 2008-08-07 |
Family
ID=38961132
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2007/020847 WO2008042219A2 (fr) | 2006-09-29 | 2007-09-26 | Interface utilisateur et identification intégrés à un dispositif médical, systèmes et procédés |
PCT/US2007/020848 WO2008042220A2 (fr) | 2006-09-29 | 2007-09-26 | Interface utilisateur et identification intégrés et un dispositif médical, systèmes et procédés |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2007/020848 WO2008042220A2 (fr) | 2006-09-29 | 2007-09-26 | Interface utilisateur et identification intégrés et un dispositif médical, systèmes et procédés |
Country Status (1)
Country | Link |
---|---|
WO (2) | WO2008042219A2 (fr) |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2283790A1 (fr) * | 2009-08-14 | 2011-02-16 | Karl Storz GmbH & Co. KG | Commande et procédé de fonctionnement d'une lampe d'opération |
WO2012172383A1 (fr) * | 2011-06-15 | 2012-12-20 | Debreceni Egyetem | Système pour déterminer une zone cible chirurgicale et système de positionnement automatique pour des lampes de salles d'opération |
WO2013089549A1 (fr) * | 2011-12-16 | 2013-06-20 | Alarcon Alfaro Roberto | Interface électronique de commande corporelle |
CN103258107A (zh) * | 2012-02-17 | 2013-08-21 | 普天信息技术研究院有限公司 | 一种监护方法和监护辅助系统 |
WO2013168056A1 (fr) * | 2012-05-10 | 2013-11-14 | Koninklijke Philips N.V. | Commande de geste |
EP2805685A1 (fr) * | 2013-05-24 | 2014-11-26 | Surgiris | Système d'éclairage médical, notamment d'éclairage operatoire, et procédé de commande d'un tel système d'éclairage |
CN106361349A (zh) * | 2015-11-19 | 2017-02-01 | 北京智谷睿拓技术服务有限公司 | 身体左右侧确定方法和设备 |
EP2480157B1 (fr) * | 2009-11-13 | 2017-03-22 | Intuitive Surgical Operations, Inc. | Système pour la commande de main d'un instrument chirurgical asservi à effraction minimale actionné à distance |
CN106725860A (zh) * | 2009-11-13 | 2017-05-31 | 直观外科手术操作公司 | 用于微创外科手术系统中的手势控制的设备 |
US9743989B2 (en) | 2010-09-21 | 2017-08-29 | Intuitive Surgical Operations, Inc. | Method and system for hand presence detection in a minimally invasive surgical system |
US9901402B2 (en) | 2010-09-21 | 2018-02-27 | Intuitive Surgical Operations, Inc. | Method and apparatus for hand gesture control in a minimally invasive surgical system |
US10067569B2 (en) | 2015-08-14 | 2018-09-04 | Fresenius Medical Care Holdings, Inc. | Touchless interface for a medical treatment system |
KR101900922B1 (ko) | 2009-11-13 | 2018-09-21 | 인튜어티브 서지컬 오퍼레이션즈 인코포레이티드 | 최소 침습 수술 시스템에서 손 존재 검출을 위한 방법 및 시스템 |
US10288881B2 (en) | 2013-03-14 | 2019-05-14 | Fresenius Medical Care Holdings, Inc. | Wearable interface for remote monitoring and control of a medical device |
EP3499511A1 (fr) * | 2017-12-18 | 2019-06-19 | Löwenstein Medical Technology S.A. | Ventilateur avec un dispositif pour la détection sans contact des opérations de l'utilisateur |
US10623188B2 (en) | 2017-04-26 | 2020-04-14 | Fresenius Medical Care Holdings, Inc. | Securely distributing medical prescriptions |
US11167071B2 (en) | 2015-12-15 | 2021-11-09 | Fresenius Medical Care Deutschland Gmbh | System and method for detecting an operating state or a course of treatment in a blood treatment |
US11355235B2 (en) | 2011-07-15 | 2022-06-07 | Fresenius Medical Care Deutschland Gmbh | Method and device for remote monitoring and control of medical fluid management devices |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP4071586A1 (fr) * | 2021-04-09 | 2022-10-12 | Fresenius Medical Care Deutschland GmbH | Commande sans contact des appareils médicaux |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6359612B1 (en) * | 1998-09-30 | 2002-03-19 | Siemens Aktiengesellschaft | Imaging system for displaying image information that has been acquired by means of a medical diagnostic imaging device |
WO2002037466A1 (fr) * | 2000-11-02 | 2002-05-10 | Essential Reality, Inc | Dispositif d'interface electronique porte par l'utilisateur |
US20030048312A1 (en) * | 1987-03-17 | 2003-03-13 | Zimmerman Thomas G. | Computer data entry and manipulation apparatus and method |
EP1643401A1 (fr) * | 2004-10-01 | 2006-04-05 | General Electric Company | Procédé et appareil pour salle d'opérations chirurgicales, affichage d'informations, détection de gaz et hiérarchisation des utilisateurs pour contrôle |
WO2006087689A2 (fr) * | 2005-02-18 | 2006-08-24 | Koninklijke Philips Electronics N. V. | Commande automatique de dispositif medical |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE19932147A1 (de) * | 1999-07-12 | 2001-01-25 | Insys Ges Fuer Microcontroller | Elektronisches System zur Erfassung und Kontrolle von Patientendaten |
DE10325106A1 (de) * | 2003-06-03 | 2004-12-30 | Disetronic Licensing Ag | Vorrichtung und Verfahren zur Erkennung eines Benutzers eines medizinischen Geräts |
US20050062603A1 (en) * | 2003-08-06 | 2005-03-24 | Oren Fuerst | Secure, networked and wireless access, storage and retrival system and method utilizing tags and modular nodes |
CA2439722A1 (fr) * | 2003-09-04 | 2005-03-04 | Vernon George Houle | Methode de controle des deplacements a l'interieur et a l'exterieur d'une installation |
DE102004013814A1 (de) * | 2004-03-20 | 2005-10-13 | B. Braun Medizintechnologie Gmbh | Verfahren zur Gestattung von Bedienereingaben an einem medizinischen Gerät |
-
2007
- 2007-09-26 WO PCT/US2007/020847 patent/WO2008042219A2/fr active Application Filing
- 2007-09-26 WO PCT/US2007/020848 patent/WO2008042220A2/fr active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030048312A1 (en) * | 1987-03-17 | 2003-03-13 | Zimmerman Thomas G. | Computer data entry and manipulation apparatus and method |
US6359612B1 (en) * | 1998-09-30 | 2002-03-19 | Siemens Aktiengesellschaft | Imaging system for displaying image information that has been acquired by means of a medical diagnostic imaging device |
WO2002037466A1 (fr) * | 2000-11-02 | 2002-05-10 | Essential Reality, Inc | Dispositif d'interface electronique porte par l'utilisateur |
EP1643401A1 (fr) * | 2004-10-01 | 2006-04-05 | General Electric Company | Procédé et appareil pour salle d'opérations chirurgicales, affichage d'informations, détection de gaz et hiérarchisation des utilisateurs pour contrôle |
WO2006087689A2 (fr) * | 2005-02-18 | 2006-08-24 | Koninklijke Philips Electronics N. V. | Commande automatique de dispositif medical |
Cited By (34)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2283790A1 (fr) * | 2009-08-14 | 2011-02-16 | Karl Storz GmbH & Co. KG | Commande et procédé de fonctionnement d'une lampe d'opération |
US8817085B2 (en) | 2009-08-14 | 2014-08-26 | Karl Storz Gmbh & Co. Kg | Control system and method to operate an operating room lamp |
EP2480157B1 (fr) * | 2009-11-13 | 2017-03-22 | Intuitive Surgical Operations, Inc. | Système pour la commande de main d'un instrument chirurgical asservi à effraction minimale actionné à distance |
KR101956900B1 (ko) | 2009-11-13 | 2019-03-12 | 인튜어티브 서지컬 오퍼레이션즈 인코포레이티드 | 최소 침습 수술 시스템에서 손 존재 검출을 위한 방법 및 시스템 |
KR20180105270A (ko) * | 2009-11-13 | 2018-09-27 | 인튜어티브 서지컬 오퍼레이션즈 인코포레이티드 | 최소 침습 수술 시스템에서 손 존재 검출을 위한 방법 및 시스템 |
KR101900922B1 (ko) | 2009-11-13 | 2018-09-21 | 인튜어티브 서지컬 오퍼레이션즈 인코포레이티드 | 최소 침습 수술 시스템에서 손 존재 검출을 위한 방법 및 시스템 |
EP3320875A1 (fr) * | 2009-11-13 | 2018-05-16 | Intuitive Surgical Operations Inc. | Appareil de commande de gestes manuels dans un système chirurgical à invasion minimale |
EP2498711B1 (fr) * | 2009-11-13 | 2018-01-10 | Intuitive Surgical Operations, Inc. | Appareil de commande des gestes de la main dans un système de chirurgie à effraction minimale |
KR101762638B1 (ko) * | 2009-11-13 | 2017-07-28 | 인튜어티브 서지컬 오퍼레이션즈 인코포레이티드 | 최소 침습 수술 시스템에서 손 제스처 제어를 위한 방법 및 장치 |
CN106725860A (zh) * | 2009-11-13 | 2017-05-31 | 直观外科手术操作公司 | 用于微创外科手术系统中的手势控制的设备 |
US10543050B2 (en) | 2010-09-21 | 2020-01-28 | Intuitive Surgical Operations, Inc. | Method and system for hand presence detection in a minimally invasive surgical system |
US9743989B2 (en) | 2010-09-21 | 2017-08-29 | Intuitive Surgical Operations, Inc. | Method and system for hand presence detection in a minimally invasive surgical system |
US9901402B2 (en) | 2010-09-21 | 2018-02-27 | Intuitive Surgical Operations, Inc. | Method and apparatus for hand gesture control in a minimally invasive surgical system |
US11707336B2 (en) | 2010-09-21 | 2023-07-25 | Intuitive Surgical Operations, Inc. | Method and system for hand tracking in a robotic system |
WO2012172383A1 (fr) * | 2011-06-15 | 2012-12-20 | Debreceni Egyetem | Système pour déterminer une zone cible chirurgicale et système de positionnement automatique pour des lampes de salles d'opération |
US11869660B2 (en) | 2011-07-15 | 2024-01-09 | Fresenius Medical Care Deutschland Gmbh | Method and device for remote monitoring and control of medical fluid management devices |
US11355235B2 (en) | 2011-07-15 | 2022-06-07 | Fresenius Medical Care Deutschland Gmbh | Method and device for remote monitoring and control of medical fluid management devices |
WO2013089549A1 (fr) * | 2011-12-16 | 2013-06-20 | Alarcon Alfaro Roberto | Interface électronique de commande corporelle |
CN103258107A (zh) * | 2012-02-17 | 2013-08-21 | 普天信息技术研究院有限公司 | 一种监护方法和监护辅助系统 |
US9483122B2 (en) | 2012-05-10 | 2016-11-01 | Koninklijke Philips N.V. | Optical shape sensing device and gesture control |
WO2013168056A1 (fr) * | 2012-05-10 | 2013-11-14 | Koninklijke Philips N.V. | Commande de geste |
US10288881B2 (en) | 2013-03-14 | 2019-05-14 | Fresenius Medical Care Holdings, Inc. | Wearable interface for remote monitoring and control of a medical device |
EP2805685A1 (fr) * | 2013-05-24 | 2014-11-26 | Surgiris | Système d'éclairage médical, notamment d'éclairage operatoire, et procédé de commande d'un tel système d'éclairage |
FR3006034A1 (fr) * | 2013-05-24 | 2014-11-28 | Surgiris | Systeme d'eclairage medical, notamment d'eclairage operatoire, et procede de commande d'un tel systeme d'eclairage |
US10067569B2 (en) | 2015-08-14 | 2018-09-04 | Fresenius Medical Care Holdings, Inc. | Touchless interface for a medical treatment system |
CN106361349A (zh) * | 2015-11-19 | 2017-02-01 | 北京智谷睿拓技术服务有限公司 | 身体左右侧确定方法和设备 |
US11167071B2 (en) | 2015-12-15 | 2021-11-09 | Fresenius Medical Care Deutschland Gmbh | System and method for detecting an operating state or a course of treatment in a blood treatment |
US10623188B2 (en) | 2017-04-26 | 2020-04-14 | Fresenius Medical Care Holdings, Inc. | Securely distributing medical prescriptions |
US11424934B2 (en) | 2017-04-26 | 2022-08-23 | Fresenius Medical Care Holdings, Inc. | Securely distributing medical prescriptions |
US12101414B2 (en) | 2017-04-26 | 2024-09-24 | Fresenius Medical Care Holdings, Inc. | Securely distributing medical prescriptions |
JP2019107447A (ja) * | 2017-12-18 | 2019-07-04 | レーヴェンシュタイン メディカル テクノロジー エス.アー. | 利用者の操作プロセスを無接触式に検出する装置を有する呼吸器具 |
EP3499511A1 (fr) * | 2017-12-18 | 2019-06-19 | Löwenstein Medical Technology S.A. | Ventilateur avec un dispositif pour la détection sans contact des opérations de l'utilisateur |
JP7204466B2 (ja) | 2017-12-18 | 2023-01-16 | レーヴェンシュタイン メディカル テクノロジー エス.アー. | 利用者の操作プロセスを無接触式に検出する装置を有する呼吸器具 |
US11815866B2 (en) | 2017-12-18 | 2023-11-14 | Loewenstein Medical Technology S.A. | Ventilator comprising a device for contactless detection of operations carried out by a user |
Also Published As
Publication number | Publication date |
---|---|
WO2008042220A3 (fr) | 2009-11-26 |
WO2008042220A2 (fr) | 2008-04-10 |
WO2008042219A3 (fr) | 2008-08-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7698002B2 (en) | Systems and methods for user interface and identification in a medical device | |
US8160726B2 (en) | User interface and identification in a medical device system and method | |
US20080097176A1 (en) | User interface and identification in a medical device systems and methods | |
US20080097177A1 (en) | System and method for user interface and identification in a medical device | |
WO2008042219A2 (fr) | Interface utilisateur et identification intégrés à un dispositif médical, systèmes et procédés | |
US20210145300A1 (en) | Spot check measurement system | |
US9801542B2 (en) | Health monitoring appliance | |
US20190295096A1 (en) | Smart watch and operating method using the same | |
US20080082338A1 (en) | Systems and methods for secure voice identification and medical device interface | |
US10152150B2 (en) | Facilitating user input via arm-mounted peripheral device interfacing with head-mounted display device | |
US20230058803A1 (en) | Method, device, and system for remotely monitoring a medical device | |
KR101793587B1 (ko) | 원격진료환자 본인확인을 위한 사용자 인증가능 산소포화도 측정 기능을 구비한 모바일기기 및 이를 이용한 산소포화도 측정 방법 | |
Kalpana et al. | Smart Glove with Gesture Based Communication and Monitoring of Paralyzed Patient | |
EP4345838A1 (fr) | Visualisation d'une indication d'un emplacement dans une installation médicale | |
FERLINI et al. | Mobile health with head-worn devices: Challenges and opportunities.(2022) |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 07838931 Country of ref document: EP Kind code of ref document: A2 |
|
NENP | Non-entry into the national phase in: |
Ref country code: DE |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC, EPO FORM 1205A DATED 18.08.2009 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 07838931 Country of ref document: EP Kind code of ref document: A2 |