US11515030B2 - System and method for artificial agent based cognitive operating rooms - Google Patents
System and method for artificial agent based cognitive operating rooms Download PDFInfo
- Publication number
- US11515030B2 US11515030B2 US16/310,565 US201716310565A US11515030B2 US 11515030 B2 US11515030 B2 US 11515030B2 US 201716310565 A US201716310565 A US 201716310565A US 11515030 B2 US11515030 B2 US 11515030B2
- Authority
- US
- United States
- Prior art keywords
- data
- current context
- surgical procedure
- patient
- next steps
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
- 238000000034 method Methods 0.000 title claims abstract description 58
- 230000001149 cognitive effect Effects 0.000 title abstract description 12
- 238000001356 surgical procedure Methods 0.000 claims abstract description 56
- 238000011002 quantification Methods 0.000 claims abstract description 38
- 238000005259 measurement Methods 0.000 claims abstract description 33
- 238000002059 diagnostic imaging Methods 0.000 claims abstract description 21
- 238000004422 calculation algorithm Methods 0.000 claims description 15
- 230000011218 segmentation Effects 0.000 claims description 12
- 239000004020 conductor Substances 0.000 claims description 9
- 238000005094 computer simulation Methods 0.000 claims description 7
- 230000002787 reinforcement Effects 0.000 claims description 7
- 238000013179 statistical model Methods 0.000 claims description 5
- 230000003190 augmentative effect Effects 0.000 claims description 4
- 230000000007 visual effect Effects 0.000 claims description 4
- 230000004044 response Effects 0.000 claims 2
- 238000007670 refining Methods 0.000 claims 1
- 239000003795 chemical substances by application Substances 0.000 description 57
- 230000004927 fusion Effects 0.000 description 16
- 230000010354 integration Effects 0.000 description 10
- 238000003384 imaging method Methods 0.000 description 9
- 238000013473 artificial intelligence Methods 0.000 description 8
- 238000002560 therapeutic procedure Methods 0.000 description 8
- 238000004590 computer program Methods 0.000 description 6
- 230000000747 cardiac effect Effects 0.000 description 5
- 238000002591 computed tomography Methods 0.000 description 5
- 238000002360 preparation method Methods 0.000 description 5
- 238000002604 ultrasonography Methods 0.000 description 5
- 230000008901 benefit Effects 0.000 description 4
- 238000012544 monitoring process Methods 0.000 description 4
- 238000013439 planning Methods 0.000 description 4
- 210000003484 anatomy Anatomy 0.000 description 3
- 238000013499 data model Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- ACGUYXCXAPNIKK-UHFFFAOYSA-N hexachlorophene Chemical compound OC1=C(Cl)C=C(Cl)C(Cl)=C1CC1=C(O)C(Cl)=CC(Cl)=C1Cl ACGUYXCXAPNIKK-UHFFFAOYSA-N 0.000 description 3
- 210000000056 organ Anatomy 0.000 description 3
- 238000011477 surgical intervention Methods 0.000 description 3
- 230000003044 adaptive effect Effects 0.000 description 2
- 238000002583 angiography Methods 0.000 description 2
- 230000036772 blood pressure Effects 0.000 description 2
- 238000013135 deep learning Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000002001 electrophysiology Methods 0.000 description 2
- 230000007831 electrophysiology Effects 0.000 description 2
- 238000010191 image analysis Methods 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 238000002595 magnetic resonance imaging Methods 0.000 description 2
- 238000007726 management method Methods 0.000 description 2
- 210000004115 mitral valve Anatomy 0.000 description 2
- 239000000523 sample Substances 0.000 description 2
- 210000001519 tissue Anatomy 0.000 description 2
- 206010002091 Anaesthesia Diseases 0.000 description 1
- 238000002679 ablation Methods 0.000 description 1
- 230000006978 adaptation Effects 0.000 description 1
- 230000002411 adverse Effects 0.000 description 1
- 230000037005 anaesthesia Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000000740 bleeding effect Effects 0.000 description 1
- 210000000988 bone and bone Anatomy 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000003745 diagnosis Methods 0.000 description 1
- 239000003814 drug Substances 0.000 description 1
- 229940079593 drug Drugs 0.000 description 1
- 238000001647 drug administration Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000036512 infertility Effects 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 238000010859 live-cell imaging Methods 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 239000003550 marker Substances 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000035479 physiological effects, processes and functions Effects 0.000 description 1
- 230000002980 postoperative effect Effects 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 238000011155 quantitative monitoring Methods 0.000 description 1
- 238000011084 recovery Methods 0.000 description 1
- 238000009877 rendering Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 238000012384 transportation and delivery Methods 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H30/00—ICT specially adapted for the handling or processing of medical images
- G16H30/20—ICT specially adapted for the handling or processing of medical images for handling medical images, e.g. DICOM, HL7 or PACS
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/25—User interfaces for surgical systems
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7235—Details of waveform analysis
- A61B5/7264—Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems
- A61B5/7267—Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems involving training the classification device
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/74—Details of notification to user or communication with user or patient ; user input means
- A61B5/7405—Details of notification to user or communication with user or patient ; user input means using sound
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/74—Details of notification to user or communication with user or patient ; user input means
- A61B5/742—Details of notification to user or communication with user or patient ; user input means using visual displays
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/74—Details of notification to user or communication with user or patient ; user input means
- A61B5/746—Alarms related to a physiological condition, e.g. details of setting alarm thresholds or avoiding false alarms
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/30—Determination of transform parameters for the alignment of images, i.e. image registration
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B7/00—Signalling systems according to more than one of groups G08B3/00 - G08B6/00; Personal calling systems according to more than one of groups G08B3/00 - G08B6/00
- G08B7/06—Signalling systems according to more than one of groups G08B3/00 - G08B6/00; Personal calling systems according to more than one of groups G08B3/00 - G08B6/00 using electric transmission, e.g. involving audible and visible signalling through the use of sound and light sources
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B19/00—Teaching not covered by other main groups of this subclass
- G09B19/003—Repetitive work cycles; Sequence of movements
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H20/00—ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance
- G16H20/40—ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance relating to mechanical, radiation or invasive therapies, e.g. surgery, laser therapy, dialysis or acupuncture
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H30/00—ICT specially adapted for the handling or processing of medical images
- G16H30/40—ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H40/00—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices
- G16H40/60—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices
- G16H40/63—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices for local operation
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/20—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
- A61B2034/101—Computer-aided simulation of surgical operations
- A61B2034/102—Modelling of surgical devices, implants or prosthesis
- A61B2034/104—Modelling the effect of the tool, e.g. the effect of an implanted prosthesis or for predicting the effect of ablation or burring
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
- A61B2034/101—Computer-aided simulation of surgical operations
- A61B2034/105—Modelling of the patient, e.g. for ligaments or bones
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
- A61B2034/107—Visualisation of planned trajectories or target regions
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/25—User interfaces for surgical systems
- A61B2034/252—User interfaces for surgical systems indicating steps of a surgical procedure
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/25—User interfaces for surgical systems
- A61B2034/258—User interfaces for surgical systems providing specific settings for specific users
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H70/00—ICT specially adapted for the handling or processing of medical references
- G16H70/20—ICT specially adapted for the handling or processing of medical references relating to practices or guidelines
Definitions
- the present invention relates to an artificial agent for providing artificial intelligence based operating room support, and more particularly to a system and method for providing an intelligent artificial agent that supports the clinical workflow around surgical procedures.
- the operating room With the boom of medical technologies and increasingly complex therapies, the operating room (OR) is becoming an extremely complex environment, with several actors and multiple, real-time but asynchronous information streams (e.g., live imaging, physiological data, patient history, medical body of knowledge, etc.) to mentally integrate.
- the main operator e.g., surgeon, interventional cardiologist, etc.
- the main operator may be overloaded with information from multiple sources while at the same time performing complex tasks. This is true for the entire patient care workflow, from treatment preparation (planning, patient preparation, etc.) to the treatment itself, to follow up care.
- the present invention provides an artificial agent based cognitive operating room system and method thereof.
- the present inventors have recognized that there is a need to centralize data acquired from multiple sources in the operating room (OR) in a consistent, time-synchronized fashion, present the data in a unified way to the OR actors, and assist the operators in their tasks with an intelligent system that understands the current situation and can predict what will happen next for context-aware support and medical reasoning.
- OR operating room
- Such a system would not only simplify the OR environment, thus increasing safety and efficacy, but also increase throughput through automation with intelligent imaging and medical systems, and smoothen the learning curve of the operators.
- Embodiments of the present invention provide an intelligent, ubiquitous artificial agent that supports the clinical workflow around surgical procedures.
- the artificial agent supports a next generation cognitive OR in which an intelligent system automatically integrates information from multiple sources and selectively presents the information to the operators based on the surgical procedure, the stage of the intervention, the user, etc.
- the intelligent system is context aware and reasons based on the various data streams and prior knowledge, such as patient medical history, operator experience and preferences, and medical knowledge like guidelines, recommendations, clinical trial results, etc.
- data related to the surgical procedure from multiple data sources is fused based on a current context.
- the data includes medical images of a patient acquired using one or more medical imaging modalities.
- Real-time quantification of patient measurements based on the data from the multiple data sources is performed based on the current context.
- Short-term predictions in the surgical procedure are forecasted based on the current context, the fused data, and the real-time quantification of the patient measurements.
- Suggestions for next steps in the surgical procedure and relevant information in the fused data are determined based on the current context and the short-term predictions.
- the suggestions for the next steps and the relevant information in the fused data are presented to an operator.
- FIG. 1 illustrates a high-level workflow of an intelligent artificial agent based cognitive operating room (OR) system according to an embodiment of the present invention
- FIG. 2 a cognitive OR system for providing automated assistance for surgery or therapy according to an embodiment of the present invention
- FIG. 3 illustrates a method for providing automated assistance for a surgical procedure using the system of FIG. 2 according to an embodiment of the present invention
- FIG. 4 illustrates a computation infrastructure for performing the real-time quantification according to an embodiment of the present invention
- FIG. 5 illustrates a distributed multi-agent system for providing automated AI-based assistance for a surgical procedure according to an embodiment of the present invention
- FIG. 6 is a high-level block diagram of a computer capable of implementing the present invention.
- the present invention relates to an artificial agent based cognitive operating room system and method thereof.
- Embodiments of the present invention provide a system and method in which an intelligent artificial agent provides support for the clinical workflow of procedures based on medical images and other data streams.
- a digital image is often composed of digital representations of one or more objects (or shapes).
- the digital representation of an object is often described herein in terms of identifying and manipulating the objects.
- Such manipulations are virtual manipulations accomplished in the memory or other circuitry/hardware of a computer system. Accordingly, is to be understood that embodiments of the present invention may be performed within a computer system using data stored within the computer system or a remote computer system.
- Embodiments of the present invention provide an intelligent, ubiquitous artificial agent that supports the clinical workflow around surgical procedures, including planning, patient preparation, intervention, monitoring, and/or follow-up.
- the intelligent artificial agent supports a next generation cognitive operating room (OR) in which an intelligent system automatically integrates data from multiple sources and selectively presents the data to operators based on the surgical procedure, the stage of the intervention, the user, etc.
- the intelligent artificial agent is context aware and performs reasoning and/or prediction based on the data streams and prior knowledge, such as patient medical history, operator experience and preferences, and medical body of knowledge (e.g., guidelines, recommendations, clinical trial results, etc.).
- FIG. 1 illustrates a high-level workflow of an intelligent artificial agent based cognitive OR system according to an embodiment of the present invention.
- the artificial agent based cognitive OR system performs data integration (shown in image 102 ), in which OR “signals” or data streams (e.g., medical images from medical imaging scanners, real-time quantification and fusion (shown in image 104 ), context-awareness, and reasoning to automatically generate short term predictions to anticipate upcoming events, suggest next steps, and reduce risks.
- the results of the context-aware reasoning performed by the system are provided in real-time to operators in the OR, as shown in image 106 .
- real-time data streams are integrated into a common data model and representation system (time-synchronized but asynchronous).
- the data streams can include medical imaging devices, cameras, physiological information, etc.
- Information of the integrated data streams is then fused based on the current context, for example using multi-modality image and data fusion algorithms.
- the information of the integrated data streams is quantified based on the current context, for example using segmentation and tracking algorithms and/or live measurements. Analyzing and predicting next phases are performed based on the fused information, the quantification results, computational and statistical models, and background knowledge (e.g., medical body of knowledge, patient history, etc.). Next steps are proposed and alerts are raised based on the current context background information, and short-term predictions.
- the intelligent artificial agent interacts with the operator(s) using ubiquitous human-machine interactions and context-aware data presentation. In a possible embodiment, several artificial agents can work together, collaboratively or competitively, for distributed support.
- FIG. 2 a cognitive OR system 200 for providing automated assistance for surgery or therapy according to an embodiment of the present invention.
- the system 200 includes data integration middleware 205 and a virtual agent 210 .
- the data integration middleware 205 integrates data from various data sources, including real-time data streams, into a common data model and representation system.
- the data sources can include medical imaging scanners 207 and other sensors/actuators 209 .
- the medical imaging scanners 207 can include any medical imaging modalities, such as computed tomography (CT), magnetic resonance imaging (MRI), ultrasound, DynaCT, X-ray, laparoscopic, endoscopic, etc.
- the medical imaging scanners can be used to acquire pre-operative medical images of the patient (e.g., CT, MRI, etc.) and intra-operative medical images of the patient (e.g., ultrasound, X-ray angiography, laparoscopic, endoscopic, etc.)
- the intraoperative medical images may be received as a real-time data stream.
- the sensors/actuators 209 may include physiological sensors that acquire physiological measurements of the patient (e.g., ECG, blood pressure, temperature, heart rate, etc.), cameras (3D or 2D cameras), sound/voice sensors and/or actuators, devices (e.g., medical devices, probes, etc.), and/or robots (e.g., robotic systems with actuators that are controllable by the virtual agent 210 ).
- Data from the sensors/actuators 209 may be received as real-time data streams.
- the data integration middleware 205 is implemented on a computer system and stores the data acquired from the various data sources (medical images, clinical measurements, etc.) in order to centralize the data from the various sources.
- the data integration middleware 105 also time-synchronizes the data in the real-time data streams.
- the data integration middleware 105 can be implemented using an information technology (IT) infrastructure that centralizes and synchronizes the data streams.
- IT information technology
- the virtual agent 210 is a computer-based intelligent artificial agent that uses artificial intelligence to intelligently fuse data, quantify patient measurements from the data, reason, and present information to provide context aware guidance during a surgical intervention.
- the virtual agent 210 can also perform actions such as automatically deciding to fuse or quantify some aspects in an image or autonomously controlling devices in the OR.
- the virtual agent 210 is implemented on one or more computers or processors by executing computer program instructions (code) loaded into memory.
- the virtual agent 210 determines a current context for the surgical intervention based on the data integrated by the data integration middleware 205 (e.g., medical images, physiological measurements, etc.) patient information 204 , and information from a knowledge base of medical procedures 202 , and autonomously performs the data fusion, quantification, reasoning, and presentation based on the current context.
- the virtual agent 210 can be implemented on a local computer device in the OR.
- the virtual agent 210 and the data integration middleware 205 may be run on separate computer devices or on the same computer device.
- the virtual agent 210 can be implemented on one or more remote devices, such as in a cloud-based computing system.
- the virtual agent 210 includes a real-time computing engine 212 and an adaptive presentation layer 214 .
- the real-time computing engine 212 performs real-time data fusion, quantification, and reasoning based on the data integrated by the data integration middleware 205 , patient information 204 , and information the knowledge base of medical procedures 202 . Additional details regarding the real-time fusion, quantification, and reasoning are described in greater detail below.
- the presentation layer 214 provides a seamless interface with the user(s)/operator(s) by showing or telling only the relevant information at any specific time of the surgical procedure.
- the presentation layer 214 can interact with the user(s) and/or OR environment in various ways.
- the presentation layer 214 can provide information, such as guidance for next steps of the surgical procedure, alerts, etc., by displaying information on a display device, using an augmented reality (AR) view, speaking information using a synthetic voice, and/or using digital commands to control other devices in the OR. Additional details regarding the presentation of information by the presentation layer 214 are described below.
- AR augmented reality
- Patient information 204 such as a patient diagnosis, therapy planning information, clinical history, and the patient position in the OR is input to the virtual agent 210 .
- Patient information 204 may be stored in a patient information database.
- patient information 204 may be stored on a medical information system, such as a picture archiving and communication system (PACS).
- PACS picture archiving and communication system
- the virtual agent 210 can also retrieve information, such as past recordings, population analysis, surgical process models, surgical ontologies, medical guidelines and recommendations, and clinical trial results, from a knowledge base of medical procedures 202 .
- FIG. 3 illustrates a method for providing automated assistance for a surgical procedure using the system of FIG. 2 according to an embodiment of the present invention.
- data related to the surgical procedure is received from multiple data sources.
- the data can include medical images (pre-operative and intra-operative) acquired using medical imaging scanners 207 for various medical imaging modalities, physiological measurements, camera images (e.g., RGB-D images), and other sensor/actuator data. Some of the data (e.g., intra-operative images, physiological measurements, camera images) can be received in real-time data streams.
- the data integration middleware 205 integrates the data into a common data model and representation system. This results in the data from the various sources being centralized and time-synchronized.
- the data from the different data sources is fused in a common frame of reference.
- This data fusion is performed by the real-time computing engine 212 of the virtual agent 210 .
- the data fusion is performed at step 306 and the real-time quantification is performed at step 308 , it is also possible that the data fusion and real-time quantification steps be performed in parallel.
- image registration technologies While standard image registration techniques apply, in an advantageous embodiment artificial intelligence is used to achieve the image registration in a robust, semantic, and context-aware way.
- task-specific intelligent artificial agents are trained to perform various registration tasks corresponding to anatomies being registered and the medical imaging modalities of the images being registered.
- the task-specific intelligent artificial agents can utilize trained deep neural networks to iteratively select registration actions to implement a registration policy learned using deep reinforcement learning, or other artificial agent registration techniques described in U.S. patent application Ser. No. 15/587,094, filed May 4, 2017, entitled “Method and System for Image Registration Using an Intelligent Artificial Agent,” which is incorporated herein in its entirety by reference.
- an artificial registration conductor is trained.
- the trained artificial registration conductor is an intelligent artificial agent that automatically selects the artificial registration agent to apply, given the surgical context, the input images and the organs visible in the images. For instance, if the artificial registration conductor is given a pre-op chest CT and a DynaCT, and it knows a cardiac procedure is being performed, it will call the cardiac registration agent to perform the fusion. However, if instead a spine surgery is being done, then the artificial registration conductor agent would call the spine registration agent.
- the artificial registration conductor can be trained in a supervised fashion using deep learning for instance.
- the input to the artificial registration conductor agent is the images to be registered and meta data about the surgical context, and the output is the task-specific registration agent to call.
- An active learning framework can also be used, coupled with reinforcement learning, to have the conductor learn and adapt continuously as more interventions are performed.
- Clinical data of various types can also be “fused” with imaging data. This could be achieved in a statistical fashion, where imaging features are combined with the clinical features in a statistical model trained in supervised way (e.g., risk scores, etc.).
- computational, multi-scale models can be employed to mechanistically integrate the information in a physiologically-plausible fashion.
- an organ model can be built from the medical images, and the clinical data can be used to estimate the underlying physiological parameters, such as cardiac electrical conductivity, tissue elasticity, stiffness, etc.
- an artificial agent can be trained based on reinforcement learning to perform such a task, as described in Neumann et al., “Self-Taught Artificial Agent for Multi-physics Computational Model Personalization”, Medical Image Analysis, Vol. 34, December 2016, pp. 52-64.
- a deep learning method can be used to directly infer electrical properties from ECG signals.
- Sensors intial measurement unit (IMU), ultra-wide band, etc.
- depth cameras 3D cameras
- IMU intial measurement unit
- depth cameras 3D cameras
- the localization information is thus used to associate a spatial coordinate to the object with respect to the room coordinate system.
- transformation composition any element visualized in the image by devices can then be mapped into the room coordinate system (which can also be aligned with the patient), for comprehensive information fusion.
- real-time quantification of patient-measurements is performed.
- the real-time computing engine 212 of the virtual agent 210 performs real-time quantification to extract live measurements for quantitative monitoring and logging.
- the real-time quantification is an advantageous part of the cognitive OR system, as it is important to be able to associate quantifiable measurements/numbers to what is happening during the surgical procedure in order to understand the scene and anticipate the next steps.
- FIG. 4 illustrates a computation infrastructure for performing the real-time quantification according to an embodiment of the present invention.
- the quantification relies on a computation infrastructure 400 that is multi-threaded, lock-free, and asynchronous.
- the core 402 stores algorithms that are used to perform the quantification.
- the algorithms are executed in the analyzer 404 .
- two core algorithm elements are used: robust landmark tracking and anatomical modeling (anatomical object segmentation).
- robust landmark tracking includes detecting and consistently tracking landmarks in medical images and videos, or features in signals.
- an artificial intelligence (AI)-based tracking method may be used to provide a robust, self-adaptive algorithm, which is beneficial for interventional applications.
- Live anatomy modeling i.e., anatomical object segmentation
- Various anatomical modeling/segmentation algorithms can be used.
- [Voigt 2015] describes a detecting and tracking the mitral valve annulus on real-time ultrasound.
- Such a model can be similarly used, for example to monitor MitraClip positioning, device landing, etc.
- anatomy modeling relies on machine learning based segmentation and real-time tracking.
- acquisition 406 collects the real-time data from various real-time data streams at a current time point and inputs the data in a single input channel.
- the real-time data streams can include real-time intra-operative medical images, such as laparoscopic images (LAP), angiography x-ray images (AX), and ultrasound images (US), as well as RGD+D (red, green, blue+depth) images from a 3D camera, videos, sensor data (e.g., ECG, temperature, blood pressure, etc.), and robot data.
- LAP laparoscopic images
- AX angiography x-ray images
- US ultrasound images
- RGD+D red, green, blue+depth images from a 3D camera
- videos e.g., ECG, temperature, blood pressure, etc.
- sensor data e.g., ECG, temperature, blood pressure, etc.
- robot can refer to a robotic system (e.g., robotic arm) that is controllable by the virtual agent 210 or any other smart machine.
- the robot data can include a current state and/position of the robot, or any other data that the robot can measure.
- the analyzer 404 executes robust landmark tracking algorithm elements and/or the anatomical modeling algorithm elements for the input data at the current time point to perform real-time quantitation.
- Flush 408 then outputs the data with the landmark tracking and/or anatomical modeling results for the current time point, so the data from the next time point can be processed by the analyzer 404 .
- Various output data and the quantification results for that output data may be presented (e.g., via the presentation layer 214 ) on various output devices. For example, as shown in FIG. 4 one device is controlled to display the US images with the quantification results and another device is controlled to display the AX images with the quantification results.
- the robot may be controlled based on the quantification results as well by controlling actuators of the robot via the presentation layer 214 .
- the reasoning performed by the real-time computing engine 212 of the virtual agent 210 includes steps 310 and 312 of FIG. 3 .
- short-term predictions for the surgical procedure are forecast based on the current context, as well as the real-time quantification results and the fused data.
- “fast-forward” models are employed to forecast short-term events.
- the fast-forward models can be implemented using computational models or statistical models. For instance, real-time computational models of organ physiology and organ-device interactions can be used to calculate the next stages of the surgical workflow, given the current context.
- real-time cardiac electrophysiology models such as the electrophysiology model described in Zettinig et al., “Data-Driven Estimation of Cardiac Electrical Diffusivity from 12-lead ECG signals”, Medical Image Analysis , Vol. 18, No. 8, 2014, pp. 1361-1376, which is incorporated herein in is entirety by reference
- EP data preop or interventional
- a similar model can be applied for device placement, where the landing zone can be forecasted using the model.
- One advantage of this setup is that it is a closed loop.
- the model can be updated, in real-time, with data for refined predictions, in a similar fashion as weather forecast.
- the system can be applied also at the room level, where models can be employed to predict the trajectories of the persons in the OR for collision avoidance for instance.
- suggestions for next steps in the surgical procedure, alerts, and relevant information in the fused data are determined based on the current context, the short-term predictions, and background information, such as the patient information 204 and the information from the knowledge base of medical procedures 202 .
- the virtual agent 210 performs the data fusion, real-time quantification, and forecasting of the short-term predictions, the virtual agent 210 applies an AI-based reasoning system that infers the next stages and understands the situations to raise alerts, suggests the next steps, and/or brings relevant information in the scene.
- Artificial intelligence and statistical learning techniques can be used to train the virtual agent 210 to effectively and autonomously perform these reasoning tasks.
- an artificial agent is initially trained to approach the problem as a “suggestion problem”.
- the artificial agent is trained to learn distance metrics to recognize the similarity between two events. Based on the fused data, the quantification results, the short-term predictions, and the patient-data 204 , the trained artificial agent can then determine one or more events in the knowledge base of medical procedures 202 that is/are most similar to the current context and then infer the next steps of the surgical procedure to suggest (and/or whether to raise an alert or what information should be identified as relevant) based on the similar event or events in the knowledge base of medical procedures 202 . Reinforcement learning methods can then be applied to adapt the model to a specific user.
- the artificial agent receives a positive reward if the agent performs the task it forecasted/suggested, negative reward otherwise.
- the Q-function can be approximated by a deep reinforcement learning architecture coupled with memory systems (short term, using LSTM architectures, but also longer terms, using derivable memory architectures), where the input are the data streams, and the output the suggested task.
- LSTM architectures long term, using LSTM architectures, but also longer terms, using derivable memory architectures
- the virtual agent 210 interacts with the operators and the OR environment via the presentation layer 214 to present the suggested next steps of the surgical procedure, any alerts determined to be raised, and the relevant information from the data sources.
- the virtual agent 210 uses the adaptive presentation layer 214 to interact with the surgeons/operators and any other actors in the OR (human or robotics).
- the presentation layer 214 controls images and data to be consistently displayed on a screen.
- the presentation layer 214 can utilize hyper-realistic rendering to visualize the medical images and other data, including the results of the data fusion and the real-time quantification.
- the presentation layer 214 can control only the most relevant information in the images and other data to be displayed. Instructions for the suggested next step can also be displayed on the screen, as well as a visual indication of any alerts (e.g., alerts regarding patient safety or an alert regarding an imminent collision in the OR).
- the virtual agent 210 can interact like a virtual assistant, through natural language using a synthetic voice.
- the virtual agent 210 can answer questions (using the real-time reasoning based on the real-time data fusion and quantification), automatically present the relevant information according to the reasoning system, etc.
- the synthetic voice of the virtual agent 210 can be accessible to everyone in the OR, or just to the surgeon (e.g., via an earphone or bone conduction).
- the virtual agent 210 can also be visualized using augmented reality, like a hologram.
- augmented reality could also be used to visualize the suggested next steps of the surgical procedure.
- the virtual agent 210 can perform autonomous actions related to the surgical procedure corresponding to the suggested next steps, for example by directly controlling other devices in the OR.
- the presentation layer 214 of the virtual agent 210 can be connected to actuators (robotics, devices, etc.) to directly perform a certain task in the OR.
- the virtual agent 210 can automatically order devices, decide and adjust drug levels, control anesthesia, etc.
- the virtual agent 210 can connect to actuators of an image scanner, such as a C-arm image acquisition device, to adjust settings, such as the angulation of the C-arm image acquisition device, to automatically control the intra-operative image acquisition.
- a modular, pluggable system can be used, in which devices are interfaced through “drivers” that translate the device data stream into a common internal data representation and management.
- the method of FIG. 3 can be repeated for the duration of the surgical procedure to provide automated real-time assistance from the virtual agent 210 over the whole surgical procedure.
- FIG. 5 illustrates a distributed multi-agent system for providing automated AI-based assistance for a surgical procedure according to an embodiment of the present invention.
- the system 500 includes multiple virtual agents 510 .
- Each of the virtual agents 510 receives data from the sensor and imaging data stream 506 , inputs patient information 504 and information from the knowledge base of medical procedures 502 , and operates similarly to the virtual agent 210 of FIG. 2 described above.
- the various virtual agents 510 can be trained for specific tasks corresponding to specific portions of the intervention workflow, and the virtual agents 510 communicate with each other to share relevant information, suggestions, alerts, etc., which increases the information available to each virtual agent 510 .
- the use of multiple virtual agents provides advantages in that each virtual agent can be trained on specific tasks, and more enhances intelligence can emerge from the multi-agent system.
- the virtual agents can be distributed throughout the entire interventional workflow.
- a virtual agent can be in the hospital room, monitoring and acquiring data before the intervention. This virtual agent could help identify risks and adverse events.
- Another virtual agent can be located in the OR, as previously described above in FIG. 2 .
- Another virtual agent can be located in the ICU monitoring and acquiring data during recovery.
- the above described intelligent artificial agent systems and methods can be applied to various different applications.
- Example applications for the above described intelligent artificial agent system are described herein, but it is to be understood that the present invention is not limited to only these specific applications and may be similarly extended to other applications as well.
- the above described intelligent artificial agent system may be applied for intelligent room and environment control, including providing a centralized alert system and environmental control, live dose and sterility monitoring for patients and operators with an alert system, automatic ordering of tool availability and device preparation based on current context, and/or automatic OR music playlist based on current context.
- the above described intelligent artificial agent system may be applied for smart scanning and workflow automation, including automatic adaptation of imaging scanners and devices to the current context for imaging acquisition (predictive scanning), including automatic selection of imaging modality and parameters,
- the above described intelligent artificial agent system may be applied for intelligent patient management and therapy delivery.
- the intelligent artificial agent system can provide therapy recommendations by using modeling to virtually test different therapies and provide procedural recommendations.
- the intelligent artificial agent system can be used for a body-GPS application, in which given an intervention type, the intelligent artificial agent can automatically perform preparation for the intervention (take pre-op image, segment it, path planning, fusion, etc.), and provide navigational information (visual cues, audio cues, etc.).
- the intelligent artificial agent system can be used for automatic check list and verification (pre-operative, intra-operative, and post-operative), including tool and disposable tracking.
- the intelligent artificial agent system can be used to provide a connection to a pathologist and diagnostics, with a centralized reporting system.
- the intelligent artificial agent system can be used to provide a surgical/interventional error alert system.
- the intelligent artificial agent system can be used to generate predictors for bleeding and other life-threatening events through imaging, cameras, and vitals measurements.
- the intelligent artificial agent system can be used to provide drug administration suggestions based on the current context, physiological measurements, and medical history of the patient.
- the intelligent artificial agent system can be used for end-to-end automatic fusion of images, in which based on context or user-selected pre-op imaging, therapy stage, etc., images are robustly fused to bring the relevant information at the appropriate time.
- the intelligent artificial agent system can be used to provide automatic identification for humans, in which based on the fused images, labeling of what is being seen in the scene is provided.
- the intelligent artificial agent system can be used to control actuators (e.g., robotic systems, probes, etc.) as a “third arm” to assist the surgeon, including robotic scrub nurse.
- Computer 602 contains a processor 604 , which controls the overall operation of the computer 602 by executing computer program instructions which define such operation.
- the computer program instructions may be stored in a storage device 612 (e.g., magnetic disk) and loaded into memory 610 when execution of the computer program instructions is desired.
- a storage device 612 e.g., magnetic disk
- the steps of the method of FIG. 3 may be defined by the computer program instructions stored in the memory 610 and/or storage 612 and controlled by the processor 604 executing the computer program instructions.
- One or more image acquisition device(s) 620 can be connected to the computer 602 to input image data to the computer 602 . It is possible to implement the image acquisition device 620 and the computer 602 as one device. It is also possible that the image acquisition device 620 and the computer 602 communicate wirelessly through a network.
- the computer 602 also includes one or more network interfaces 606 for communicating with other devices via a network.
- the computer 602 also includes other input/output devices 608 that enable user interaction with the computer 602 (e.g., display, keyboard, mouse, speakers, buttons, etc.).
- FIG. 6 is a high level representation of some of the components of such a computer for illustrative purposes.
- the computer that performs all of a subset of the method steps of the above-described method for providing an intelligent artificial agent based support for a surgical procedure may be implemented in network-based cloud computing system.
- a server communicates with one or more client computers via a network.
- a client computer may communicate with the server via a network browser application residing and operating on the client computer, for example.
- a client computer may store data on the server and access the data via the network.
- a client computer may transmit requests for data, or requests for online services, to the server via the network.
- the server may perform requested services and provide data to the client computer(s).
- the server may also transmit data adapted to cause a client computer to perform a specified function, e.g., to perform a calculation, to display specified data on a screen, etc.
- Certain steps of the above described methods may be performed by a server or by other computers/processors in the network-based cloud-computing system. Certain steps of the above described methods may be performed locally by a client computer in a network-based cloud computing system. The steps of the above described methods may be performed by one or more devices in the network-based cloud-computing system or by a local client computer in any combination.
Landscapes
- Health & Medical Sciences (AREA)
- Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Public Health (AREA)
- Medical Informatics (AREA)
- General Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Surgery (AREA)
- Physics & Mathematics (AREA)
- Veterinary Medicine (AREA)
- Heart & Thoracic Surgery (AREA)
- Molecular Biology (AREA)
- Animal Behavior & Ethology (AREA)
- Primary Health Care (AREA)
- Epidemiology (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Pathology (AREA)
- Biophysics (AREA)
- Business, Economics & Management (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- Radiology & Medical Imaging (AREA)
- Theoretical Computer Science (AREA)
- Physiology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Robotics (AREA)
- General Business, Economics & Management (AREA)
- Human Computer Interaction (AREA)
- Signal Processing (AREA)
- Educational Administration (AREA)
- Educational Technology (AREA)
- Entrepreneurship & Innovation (AREA)
- Evolutionary Computation (AREA)
- Fuzzy Systems (AREA)
- Mathematical Physics (AREA)
- Psychiatry (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Urology & Nephrology (AREA)
- Measuring And Recording Apparatus For Diagnosis (AREA)
Abstract
Description
Claims (26)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/310,565 US11515030B2 (en) | 2016-06-23 | 2017-06-23 | System and method for artificial agent based cognitive operating rooms |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201662353907P | 2016-06-23 | 2016-06-23 | |
PCT/EP2017/065560 WO2017220788A1 (en) | 2016-06-23 | 2017-06-23 | System and method for artificial agent based cognitive operating rooms |
US16/310,565 US11515030B2 (en) | 2016-06-23 | 2017-06-23 | System and method for artificial agent based cognitive operating rooms |
Publications (2)
Publication Number | Publication Date |
---|---|
US20190333626A1 US20190333626A1 (en) | 2019-10-31 |
US11515030B2 true US11515030B2 (en) | 2022-11-29 |
Family
ID=59285155
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/310,565 Active 2040-04-05 US11515030B2 (en) | 2016-06-23 | 2017-06-23 | System and method for artificial agent based cognitive operating rooms |
Country Status (2)
Country | Link |
---|---|
US (1) | US11515030B2 (en) |
WO (1) | WO2017220788A1 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20210170579A1 (en) * | 2019-12-09 | 2021-06-10 | Canon Kabushiki Kaisha | Robot controller that controls robot, learned model, method of controlling robot, and storage medium |
US20210326653A1 (en) * | 2020-04-08 | 2021-10-21 | Arizona Board Of Regents On Behalf Of Arizona State University | Systems, methods, and apparatuses for the generation of self-taught models genesis absent manual labeling for the processing of medical imaging |
Families Citing this family (130)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11871901B2 (en) * | 2012-05-20 | 2024-01-16 | Cilag Gmbh International | Method for situational awareness for surgical network or surgical network connected device capable of adjusting function based on a sensed situation or usage |
US20240081611A1 (en) * | 2012-05-20 | 2024-03-14 | Cilag Gmbh International | Method for situational awareness for surgical network or surgical network connected device capable of adjusting function based on a sensed situation or usage |
US11504192B2 (en) | 2014-10-30 | 2022-11-22 | Cilag Gmbh International | Method of hub communication with surgical instrument systems |
US20180233228A1 (en) * | 2017-02-14 | 2018-08-16 | GilAnthony Ungab | Systems and methods for data-driven medical decision making assistance |
WO2018195456A1 (en) * | 2017-04-20 | 2018-10-25 | The Cleveland Clinic Foundation | System and method for holographic image-guided non-vascular percutaneous procedures |
EP3470006B1 (en) * | 2017-10-10 | 2020-06-10 | Holo Surgical Inc. | Automated segmentation of three dimensional bony structure images |
EP3445048A1 (en) | 2017-08-15 | 2019-02-20 | Holo Surgical Inc. | A graphical user interface for a surgical navigation system for providing an augmented reality image during operation |
US11291510B2 (en) | 2017-10-30 | 2022-04-05 | Cilag Gmbh International | Method of hub communication with surgical instrument systems |
US11317919B2 (en) | 2017-10-30 | 2022-05-03 | Cilag Gmbh International | Clip applier comprising a clip crimping system |
US10959744B2 (en) | 2017-10-30 | 2021-03-30 | Ethicon Llc | Surgical dissectors and manufacturing techniques |
US11801098B2 (en) | 2017-10-30 | 2023-10-31 | Cilag Gmbh International | Method of hub communication with surgical instrument systems |
US11229436B2 (en) | 2017-10-30 | 2022-01-25 | Cilag Gmbh International | Surgical system comprising a surgical tool and a surgical hub |
US11311342B2 (en) | 2017-10-30 | 2022-04-26 | Cilag Gmbh International | Method for communicating with surgical instrument systems |
US11026687B2 (en) | 2017-10-30 | 2021-06-08 | Cilag Gmbh International | Clip applier comprising clip advancing systems |
US11564756B2 (en) | 2017-10-30 | 2023-01-31 | Cilag Gmbh International | Method of hub communication with surgical instrument systems |
US11911045B2 (en) | 2017-10-30 | 2024-02-27 | Cllag GmbH International | Method for operating a powered articulating multi-clip applier |
US11510741B2 (en) | 2017-10-30 | 2022-11-29 | Cilag Gmbh International | Method for producing a surgical instrument comprising a smart electrical system |
US11071595B2 (en) * | 2017-12-14 | 2021-07-27 | Verb Surgical Inc. | Multi-panel graphical user interface for a robotic surgical system |
US11304720B2 (en) | 2017-12-28 | 2022-04-19 | Cilag Gmbh International | Activation of energy devices |
US11096693B2 (en) | 2017-12-28 | 2021-08-24 | Cilag Gmbh International | Adjustment of staple height of at least one row of staples based on the sensed tissue thickness or force in closing |
US11559307B2 (en) | 2017-12-28 | 2023-01-24 | Cilag Gmbh International | Method of robotic hub communication, detection, and control |
US11419667B2 (en) | 2017-12-28 | 2022-08-23 | Cilag Gmbh International | Ultrasonic energy device which varies pressure applied by clamp arm to provide threshold control pressure at a cut progression location |
US11844579B2 (en) | 2017-12-28 | 2023-12-19 | Cilag Gmbh International | Adjustments based on airborne particle properties |
US11202570B2 (en) | 2017-12-28 | 2021-12-21 | Cilag Gmbh International | Communication hub and storage device for storing parameters and status of a surgical device to be shared with cloud based analytics systems |
US20190201146A1 (en) | 2017-12-28 | 2019-07-04 | Ethicon Llc | Safety systems for smart powered surgical stapling |
US11446052B2 (en) | 2017-12-28 | 2022-09-20 | Cilag Gmbh International | Variation of radio frequency and ultrasonic power level in cooperation with varying clamp arm pressure to achieve predefined heat flux or power applied to tissue |
US10898622B2 (en) | 2017-12-28 | 2021-01-26 | Ethicon Llc | Surgical evacuation system with a communication circuit for communication between a filter and a smoke evacuation device |
US11179208B2 (en) | 2017-12-28 | 2021-11-23 | Cilag Gmbh International | Cloud-based medical analytics for security and authentication trends and reactive measures |
US11266468B2 (en) | 2017-12-28 | 2022-03-08 | Cilag Gmbh International | Cooperative utilization of data derived from secondary sources by intelligent surgical hubs |
US11304763B2 (en) | 2017-12-28 | 2022-04-19 | Cilag Gmbh International | Image capturing of the areas outside the abdomen to improve placement and control of a surgical device in use |
US11069012B2 (en) | 2017-12-28 | 2021-07-20 | Cilag Gmbh International | Interactive surgical systems with condition handling of devices and data capabilities |
US11376002B2 (en) | 2017-12-28 | 2022-07-05 | Cilag Gmbh International | Surgical instrument cartridge sensor assemblies |
US20190201039A1 (en) | 2017-12-28 | 2019-07-04 | Ethicon Llc | Situational awareness of electrosurgical systems |
US10758310B2 (en) | 2017-12-28 | 2020-09-01 | Ethicon Llc | Wireless pairing of a surgical device with another device within a sterile surgical field based on the usage and situational awareness of devices |
US11166772B2 (en) | 2017-12-28 | 2021-11-09 | Cilag Gmbh International | Surgical hub coordination of control and communication of operating room devices |
US11464559B2 (en) | 2017-12-28 | 2022-10-11 | Cilag Gmbh International | Estimating state of ultrasonic end effector and control system therefor |
US11969216B2 (en) | 2017-12-28 | 2024-04-30 | Cilag Gmbh International | Surgical network recommendations from real time analysis of procedure variables against a baseline highlighting differences from the optimal solution |
US11317937B2 (en) | 2018-03-08 | 2022-05-03 | Cilag Gmbh International | Determining the state of an ultrasonic end effector |
US11540855B2 (en) | 2017-12-28 | 2023-01-03 | Cilag Gmbh International | Controlling activation of an ultrasonic surgical instrument according to the presence of tissue |
US11364075B2 (en) | 2017-12-28 | 2022-06-21 | Cilag Gmbh International | Radio frequency energy device for delivering combined electrical signals |
US11304745B2 (en) | 2017-12-28 | 2022-04-19 | Cilag Gmbh International | Surgical evacuation sensing and display |
US12096916B2 (en) | 2017-12-28 | 2024-09-24 | Cilag Gmbh International | Method of sensing particulate from smoke evacuated from a patient, adjusting the pump speed based on the sensed information, and communicating the functional parameters of the system to the hub |
US11278281B2 (en) | 2017-12-28 | 2022-03-22 | Cilag Gmbh International | Interactive surgical system |
US11253315B2 (en) | 2017-12-28 | 2022-02-22 | Cilag Gmbh International | Increasing radio frequency to create pad-less monopolar loop |
US11424027B2 (en) | 2017-12-28 | 2022-08-23 | Cilag Gmbh International | Method for operating surgical instrument systems |
US11324557B2 (en) | 2017-12-28 | 2022-05-10 | Cilag Gmbh International | Surgical instrument with a sensing array |
US11529187B2 (en) | 2017-12-28 | 2022-12-20 | Cilag Gmbh International | Surgical evacuation sensor arrangements |
US11291495B2 (en) | 2017-12-28 | 2022-04-05 | Cilag Gmbh International | Interruption of energy due to inadvertent capacitive coupling |
US20190201139A1 (en) | 2017-12-28 | 2019-07-04 | Ethicon Llc | Communication arrangements for robot-assisted surgical platforms |
US11786245B2 (en) | 2017-12-28 | 2023-10-17 | Cilag Gmbh International | Surgical systems with prioritized data transmission capabilities |
US11969142B2 (en) | 2017-12-28 | 2024-04-30 | Cilag Gmbh International | Method of compressing tissue within a stapling device and simultaneously displaying the location of the tissue within the jaws |
US11051876B2 (en) | 2017-12-28 | 2021-07-06 | Cilag Gmbh International | Surgical evacuation flow paths |
US11832840B2 (en) | 2017-12-28 | 2023-12-05 | Cilag Gmbh International | Surgical instrument having a flexible circuit |
US11589888B2 (en) | 2017-12-28 | 2023-02-28 | Cilag Gmbh International | Method for controlling smart energy devices |
US11659023B2 (en) * | 2017-12-28 | 2023-05-23 | Cilag Gmbh International | Method of hub communication |
US11666331B2 (en) | 2017-12-28 | 2023-06-06 | Cilag Gmbh International | Systems for detecting proximity of surgical end effector to cancerous tissue |
US11571234B2 (en) | 2017-12-28 | 2023-02-07 | Cilag Gmbh International | Temperature control of ultrasonic end effector and control system therefor |
US20190206569A1 (en) | 2017-12-28 | 2019-07-04 | Ethicon Llc | Method of cloud based data analytics for use with the hub |
US11160605B2 (en) | 2017-12-28 | 2021-11-02 | Cilag Gmbh International | Surgical evacuation sensing and motor control |
US10892995B2 (en) | 2017-12-28 | 2021-01-12 | Ethicon Llc | Surgical network determination of prioritization of communication, interaction, or processing based on system or device needs |
US11432885B2 (en) | 2017-12-28 | 2022-09-06 | Cilag Gmbh International | Sensing arrangements for robot-assisted surgical platforms |
US11389164B2 (en) | 2017-12-28 | 2022-07-19 | Cilag Gmbh International | Method of using reinforced flexible circuits with multiple sensors to optimize performance of radio frequency devices |
US11559308B2 (en) | 2017-12-28 | 2023-01-24 | Cilag Gmbh International | Method for smart energy device infrastructure |
US10966791B2 (en) | 2017-12-28 | 2021-04-06 | Ethicon Llc | Cloud-based medical analytics for medical facility segmented individualization of instrument function |
US11058498B2 (en) | 2017-12-28 | 2021-07-13 | Cilag Gmbh International | Cooperative surgical actions for robot-assisted surgical platforms |
US11132462B2 (en) | 2017-12-28 | 2021-09-28 | Cilag Gmbh International | Data stripping method to interrogate patient records and create anonymized record |
US11100631B2 (en) | 2017-12-28 | 2021-08-24 | Cilag Gmbh International | Use of laser light and red-green-blue coloration to determine properties of back scattered light |
US11896443B2 (en) | 2017-12-28 | 2024-02-13 | Cilag Gmbh International | Control of a surgical system through a surgical barrier |
US11109866B2 (en) | 2017-12-28 | 2021-09-07 | Cilag Gmbh International | Method for circular stapler control algorithm adjustment based on situational awareness |
US11273001B2 (en) | 2017-12-28 | 2022-03-15 | Cilag Gmbh International | Surgical hub and modular device response adjustment based on situational awareness |
US11633237B2 (en) | 2017-12-28 | 2023-04-25 | Cilag Gmbh International | Usage and technique analysis of surgeon / staff performance against a baseline to optimize device utilization and performance for both current and future procedures |
US11864728B2 (en) | 2017-12-28 | 2024-01-09 | Cilag Gmbh International | Characterization of tissue irregularities through the use of mono-chromatic light refractivity |
US11602393B2 (en) | 2017-12-28 | 2023-03-14 | Cilag Gmbh International | Surgical evacuation sensing and generator control |
US11304699B2 (en) | 2017-12-28 | 2022-04-19 | Cilag Gmbh International | Method for adaptive control schemes for surgical network control and interaction |
US11937769B2 (en) | 2017-12-28 | 2024-03-26 | Cilag Gmbh International | Method of hub communication, processing, storage and display |
US11410259B2 (en) | 2017-12-28 | 2022-08-09 | Cilag Gmbh International | Adaptive control program updates for surgical devices |
US11147607B2 (en) | 2017-12-28 | 2021-10-19 | Cilag Gmbh International | Bipolar combination device that automatically adjusts pressure based on energy modality |
US11896322B2 (en) | 2017-12-28 | 2024-02-13 | Cilag Gmbh International | Sensing the patient position and contact utilizing the mono-polar return pad electrode to provide situational awareness to the hub |
US10987178B2 (en) | 2017-12-28 | 2021-04-27 | Ethicon Llc | Surgical hub control arrangements |
US11076921B2 (en) | 2017-12-28 | 2021-08-03 | Cilag Gmbh International | Adaptive control program updates for surgical hubs |
US11672605B2 (en) | 2017-12-28 | 2023-06-13 | Cilag Gmbh International | Sterile field interactive control displays |
US11284936B2 (en) | 2017-12-28 | 2022-03-29 | Cilag Gmbh International | Surgical instrument having a flexible electrode |
US11786251B2 (en) | 2017-12-28 | 2023-10-17 | Cilag Gmbh International | Method for adaptive control schemes for surgical network control and interaction |
US11903601B2 (en) | 2017-12-28 | 2024-02-20 | Cilag Gmbh International | Surgical instrument comprising a plurality of drive systems |
US11744604B2 (en) | 2017-12-28 | 2023-09-05 | Cilag Gmbh International | Surgical instrument with a hardware-only control circuit |
US11998193B2 (en) | 2017-12-28 | 2024-06-04 | Cilag Gmbh International | Method for usage of the shroud as an aspect of sensing or controlling a powered surgical device, and a control algorithm to adjust its default operation |
US12062442B2 (en) | 2017-12-28 | 2024-08-13 | Cilag Gmbh International | Method for operating surgical instrument systems |
US10943454B2 (en) | 2017-12-28 | 2021-03-09 | Ethicon Llc | Detection and escalation of security responses of surgical instruments to increasing severity threats |
US11419630B2 (en) | 2017-12-28 | 2022-08-23 | Cilag Gmbh International | Surgical system distributed processing |
US11464535B2 (en) | 2017-12-28 | 2022-10-11 | Cilag Gmbh International | Detection of end effector emersion in liquid |
US11056244B2 (en) * | 2017-12-28 | 2021-07-06 | Cilag Gmbh International | Automated data scaling, alignment, and organizing based on predefined parameters within surgical networks |
US11857152B2 (en) | 2017-12-28 | 2024-01-02 | Cilag Gmbh International | Surgical hub spatial awareness to determine devices in operating theater |
US11832899B2 (en) | 2017-12-28 | 2023-12-05 | Cilag Gmbh International | Surgical systems with autonomously adjustable control programs |
US11311306B2 (en) | 2017-12-28 | 2022-04-26 | Cilag Gmbh International | Surgical systems for detecting end effector tissue distribution irregularities |
US11308075B2 (en) | 2017-12-28 | 2022-04-19 | Cilag Gmbh International | Surgical network, instrument, and cloud responses based on validation of received dataset and authentication of its source and integrity |
US11423007B2 (en) | 2017-12-28 | 2022-08-23 | Cilag Gmbh International | Adjustment of device control programs based on stratified contextual data in addition to the data |
US11234756B2 (en) | 2017-12-28 | 2022-02-01 | Cilag Gmbh International | Powered surgical tool with predefined adjustable control algorithm for controlling end effector parameter |
US11818052B2 (en) | 2017-12-28 | 2023-11-14 | Cilag Gmbh International | Surgical network determination of prioritization of communication, interaction, or processing based on system or device needs |
US11576677B2 (en) | 2017-12-28 | 2023-02-14 | Cilag Gmbh International | Method of hub communication, processing, display, and cloud analytics |
US11678881B2 (en) | 2017-12-28 | 2023-06-20 | Cilag Gmbh International | Spatial awareness of surgical hubs in operating rooms |
US11257589B2 (en) | 2017-12-28 | 2022-02-22 | Cilag Gmbh International | Real-time analysis of comprehensive cost of all instrumentation used in surgery utilizing data fluidity to track instruments through stocking and in-house processes |
US11612408B2 (en) | 2017-12-28 | 2023-03-28 | Cilag Gmbh International | Determining tissue composition via an ultrasonic system |
US11399858B2 (en) | 2018-03-08 | 2022-08-02 | Cilag Gmbh International | Application of smart blade technology |
US11259830B2 (en) | 2018-03-08 | 2022-03-01 | Cilag Gmbh International | Methods for controlling temperature in ultrasonic device |
US11337746B2 (en) | 2018-03-08 | 2022-05-24 | Cilag Gmbh International | Smart blade and power pulsing |
WO2019181432A1 (en) * | 2018-03-20 | 2019-09-26 | ソニー株式会社 | Operation assistance system, information processing device, and program |
US11090047B2 (en) | 2018-03-28 | 2021-08-17 | Cilag Gmbh International | Surgical instrument comprising an adaptive control system |
US10973520B2 (en) | 2018-03-28 | 2021-04-13 | Ethicon Llc | Surgical staple cartridge with firing member driven camming assembly that has an onboard tissue cutting feature |
US11197668B2 (en) | 2018-03-28 | 2021-12-14 | Cilag Gmbh International | Surgical stapling assembly comprising a lockout and an exterior access orifice to permit artificial unlocking of the lockout |
US11219453B2 (en) | 2018-03-28 | 2022-01-11 | Cilag Gmbh International | Surgical stapling devices with cartridge compatible closure and firing lockout arrangements |
US11207067B2 (en) | 2018-03-28 | 2021-12-28 | Cilag Gmbh International | Surgical stapling device with separate rotary driven closure and firing systems and firing member that engages both jaws while firing |
US11096688B2 (en) | 2018-03-28 | 2021-08-24 | Cilag Gmbh International | Rotary driven firing members with different anvil and channel engagement features |
US11278280B2 (en) | 2018-03-28 | 2022-03-22 | Cilag Gmbh International | Surgical instrument comprising a jaw closure lockout |
US11471156B2 (en) | 2018-03-28 | 2022-10-18 | Cilag Gmbh International | Surgical stapling devices with improved rotary driven closure systems |
US11259806B2 (en) | 2018-03-28 | 2022-03-01 | Cilag Gmbh International | Surgical stapling devices with features for blocking advancement of a camming assembly of an incompatible cartridge installed therein |
EP3608870A1 (en) | 2018-08-10 | 2020-02-12 | Holo Surgical Inc. | Computer assisted identification of appropriate anatomical structure for medical device placement during a surgical procedure |
EP3841590A1 (en) * | 2018-10-12 | 2021-06-30 | Sony Corporation | Surgical support system, data processing apparatus and method |
US11317915B2 (en) | 2019-02-19 | 2022-05-03 | Cilag Gmbh International | Universal cartridge based key feature that unlocks multiple lockout arrangements in different surgical staplers |
US11751872B2 (en) | 2019-02-19 | 2023-09-12 | Cilag Gmbh International | Insertable deactivator element for surgical stapler lockouts |
US11369377B2 (en) | 2019-02-19 | 2022-06-28 | Cilag Gmbh International | Surgical stapling assembly with cartridge based retainer configured to unlock a firing lockout |
US11298129B2 (en) | 2019-02-19 | 2022-04-12 | Cilag Gmbh International | Method for providing an authentication lockout in a surgical stapler with a replaceable cartridge |
US11357503B2 (en) | 2019-02-19 | 2022-06-14 | Cilag Gmbh International | Staple cartridge retainers with frangible retention features and methods of using same |
US11139062B2 (en) | 2019-05-21 | 2021-10-05 | Verb Surgical Inc. | Method and system for combining video, image, and audio data with text data for comprehensive data analytics |
USD950728S1 (en) | 2019-06-25 | 2022-05-03 | Cilag Gmbh International | Surgical staple cartridge |
USD964564S1 (en) | 2019-06-25 | 2022-09-20 | Cilag Gmbh International | Surgical staple cartridge retainer with a closure system authentication key |
USD952144S1 (en) | 2019-06-25 | 2022-05-17 | Cilag Gmbh International | Surgical staple cartridge retainer with firing system authentication key |
FR3111463B1 (en) * | 2020-06-12 | 2023-03-24 | Univ Strasbourg | Processing of video streams relating to surgical operations |
US20230027210A1 (en) | 2021-07-22 | 2023-01-26 | Cilag Gmbh International | Surgical data system and control |
WO2023002379A1 (en) * | 2021-07-22 | 2023-01-26 | Cilag Gmbh International | Surgical data system and management |
WO2023086332A1 (en) * | 2021-11-09 | 2023-05-19 | Genesis Medtech (USA) Inc. | An interactive augmented reality system for laparoscopic and video assisted surgeries |
Citations (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030190954A1 (en) * | 1992-05-22 | 2003-10-09 | Sitrick David H. | System and methodology for mapping and linking based user image integration |
US20080234322A1 (en) * | 2007-03-20 | 2008-09-25 | University Of Utah | Methods and apparatus for drug modeling and displaying drug models |
US7822291B2 (en) | 2004-10-28 | 2010-10-26 | Siemens Medical Solutions Usa, Inc. | Non-rigid multi-modal registration using statistical learning methods |
US20130211230A1 (en) * | 2012-02-08 | 2013-08-15 | Convergent Life Sciences, Inc. | System and method for using medical image fusion |
US20140088990A1 (en) * | 2012-09-17 | 2014-03-27 | Depuy Orthopaedics, Inc. | Systems and methods for surgical and interventional planning, support, post-operative follow-up, and, functional recovery tracking |
US20140108983A1 (en) * | 2012-01-22 | 2014-04-17 | Karen Ferguson | Graphical system for collecting, presenting and using medical data |
US9084531B2 (en) | 2013-02-27 | 2015-07-21 | Siemens Aktiengesellschaft | Providing real-time marker detection for a stent in medical imaging |
US20150364022A1 (en) * | 2014-06-13 | 2015-12-17 | Nant Health, Llc | Alarm fatigue management systems and methods |
US20160067007A1 (en) | 2013-03-15 | 2016-03-10 | Synaptive Medical (Barbados) Inc. | Interamodal synchronization of surgical data |
US9349098B1 (en) | 2015-05-14 | 2016-05-24 | James Albert Ionson | Cognitive medical and industrial inspection system and method |
US20170330075A1 (en) | 2016-05-12 | 2017-11-16 | Siemens Healthcare Gmbh | System and method for deep learning based cardiac electrophysiology model personalization |
US20170337682A1 (en) | 2016-05-18 | 2017-11-23 | Siemens Healthcare Gmbh | Method and System for Image Registration Using an Intelligent Artificial Agent |
US10653368B1 (en) * | 2013-09-09 | 2020-05-19 | Cerner Innovation, Inc. | Determining when to emit an alarm |
-
2017
- 2017-06-23 US US16/310,565 patent/US11515030B2/en active Active
- 2017-06-23 WO PCT/EP2017/065560 patent/WO2017220788A1/en active Application Filing
Patent Citations (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030190954A1 (en) * | 1992-05-22 | 2003-10-09 | Sitrick David H. | System and methodology for mapping and linking based user image integration |
US7822291B2 (en) | 2004-10-28 | 2010-10-26 | Siemens Medical Solutions Usa, Inc. | Non-rigid multi-modal registration using statistical learning methods |
US20080234322A1 (en) * | 2007-03-20 | 2008-09-25 | University Of Utah | Methods and apparatus for drug modeling and displaying drug models |
US20140108983A1 (en) * | 2012-01-22 | 2014-04-17 | Karen Ferguson | Graphical system for collecting, presenting and using medical data |
US20130211230A1 (en) * | 2012-02-08 | 2013-08-15 | Convergent Life Sciences, Inc. | System and method for using medical image fusion |
US20140088990A1 (en) * | 2012-09-17 | 2014-03-27 | Depuy Orthopaedics, Inc. | Systems and methods for surgical and interventional planning, support, post-operative follow-up, and, functional recovery tracking |
US9084531B2 (en) | 2013-02-27 | 2015-07-21 | Siemens Aktiengesellschaft | Providing real-time marker detection for a stent in medical imaging |
US20160067007A1 (en) | 2013-03-15 | 2016-03-10 | Synaptive Medical (Barbados) Inc. | Interamodal synchronization of surgical data |
US10653368B1 (en) * | 2013-09-09 | 2020-05-19 | Cerner Innovation, Inc. | Determining when to emit an alarm |
US20150364022A1 (en) * | 2014-06-13 | 2015-12-17 | Nant Health, Llc | Alarm fatigue management systems and methods |
US9349098B1 (en) | 2015-05-14 | 2016-05-24 | James Albert Ionson | Cognitive medical and industrial inspection system and method |
US20170330075A1 (en) | 2016-05-12 | 2017-11-16 | Siemens Healthcare Gmbh | System and method for deep learning based cardiac electrophysiology model personalization |
US20170337682A1 (en) | 2016-05-18 | 2017-11-23 | Siemens Healthcare Gmbh | Method and System for Image Registration Using an Intelligent Artificial Agent |
Non-Patent Citations (11)
Title |
---|
[Buzz] https://www.brainlab.com/en/surgery-products/overview-platform-products/buzz-digital-or/; retrieved from the internet. |
[Sensis] http://usa.healthcare.siemens.com/angio/workplaces/sensis; retrieved from the Internet. |
El-Zehiry, et al.; "A System and Method for the Live Landmark Tracking in Ultrasound Sequences", invention disclosure; 2015. |
Ghesu, et al.; "Towards Intelligent Image Parsing with Deep Reinforcement Learning", MICCAI; 2016. |
International Search Report dated Sep. 29, 2017 in corresponding International Patent Application No. PCT/EP2017/065560. |
Lalys, et al. "Surgical process modelling: a review." International journal of computer assisted radiology and surgery 9, No. 3 (2014): 495-511. |
Makary, et al.; "Medical error—the third leading cause of death in the US." BMJ 353 (2016): i2139. |
Neumann, et al.; "A self-taught artificial agent for multi-physics computational model personalization." Medical image analysis (2016). |
Santoro, et al.; "One-shot Learning with Memory-Augmented Neural Networks." arXiv preprint arXiv:1605.06065 (2016). |
Voigt, et al.; "Robust Live Tracking of Mitral Valve Annulus for Minimally-Invasive Intervention Guidance." In International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer International Publishing, 2015. |
Zettinig, et al. "Data-driven estimation of cardiac electrical diffusivity from 12-lead ECG signals." Medical image analysis 18.8 (2014): 1361-1376. |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20210170579A1 (en) * | 2019-12-09 | 2021-06-10 | Canon Kabushiki Kaisha | Robot controller that controls robot, learned model, method of controlling robot, and storage medium |
US11679496B2 (en) * | 2019-12-09 | 2023-06-20 | Canon Kabushiki Kaisha | Robot controller that controls robot, learned model, method of controlling robot, and storage medium |
US20210326653A1 (en) * | 2020-04-08 | 2021-10-21 | Arizona Board Of Regents On Behalf Of Arizona State University | Systems, methods, and apparatuses for the generation of self-taught models genesis absent manual labeling for the processing of medical imaging |
US11922628B2 (en) * | 2020-04-08 | 2024-03-05 | Arizona Board Of Regents On Behalf Of Arizona State University | Systems, methods, and apparatuses for the generation of self-taught models genesis absent manual labeling for the processing of medical imaging |
Also Published As
Publication number | Publication date |
---|---|
US20190333626A1 (en) | 2019-10-31 |
WO2017220788A1 (en) | 2017-12-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11515030B2 (en) | System and method for artificial agent based cognitive operating rooms | |
JP2023126480A (en) | Surgical system with training or assist functions | |
US20190008598A1 (en) | Fully autonomic artificial intelligence robotic system | |
CN110301013A (en) | Augmented reality for the predictive workflow in operating room | |
US20230157762A1 (en) | Extended Intelligence Ecosystem for Soft Tissue Luminal Applications | |
KR20180099434A (en) | Method and apparatus for monitoring patient | |
US20200375467A1 (en) | Telemedicine application of video analysis and motion augmentation | |
US20220354585A1 (en) | Robotic surgery | |
US20180247714A1 (en) | Methods and apparatuses for monitoring condition of object in medical imaging | |
US20230157757A1 (en) | Extended Intelligence for Pulmonary Procedures | |
JP2020512069A (en) | Vascular occlusion diagnosis method, device and system | |
US20210386489A1 (en) | Surgical support system, data processing apparatus and method | |
CN109256205B (en) | Method and system for clinical decision support with local and remote analytics | |
Ebnali et al. | Artificial intelligence in cardiothoracic surgery: current applications and future perspectives | |
US20220022973A1 (en) | Method, device and system for providing a virtual medical procedure drill | |
Singh | It is time for us to get artificially intelligent! | |
US20210076942A1 (en) | Infrared thermography for intraoperative functional mapping | |
EP3975201A1 (en) | Sterility in an operation room | |
US20220361954A1 (en) | Extended Intelligence for Cardiac Implantable Electronic Device (CIED) Placement Procedures | |
US20240296604A1 (en) | Systems and methods for keypoint detection and tracking-by-prediction for multiple surgical instruments | |
US20240153618A1 (en) | Systems and methods for automated communications | |
CN118575203A (en) | Detection and differentiation of critical structures in surgery using machine learning | |
EP4436508A1 (en) | Extended intelligence ecosystem for soft tissue luminal applications | |
WO2023239743A1 (en) | Use of cath lab images for procedure and device evaluation | |
Korsholm et al. | The Art of Planning a Left Atrial Appendage Closure: Will Computational Modeling Simplify the Intervention? |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SIEMENS MEDICAL SOLUTIONS USA, INC., PENNSYLVANIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MANSI, TOMMASO;KAPOOR, ANKUR;PHEIFFER, THOMAS;AND OTHERS;SIGNING DATES FROM 20181104 TO 20181120;REEL/FRAME:047794/0857 |
|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
AS | Assignment |
Owner name: SIEMENS HEALTHCARE GMBH, GERMANY Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SIEMENS MEDICAL SOLUTIONS USA, INC.;REEL/FRAME:047938/0135 Effective date: 20190108 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: SIEMENS HEALTHINEERS AG, GERMANY Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SIEMENS HEALTHCARE GMBH;REEL/FRAME:066267/0346 Effective date: 20231219 |