US20210142907A1 - System for Surgical Decisions Using Deep Learning - Google Patents

System for Surgical Decisions Using Deep Learning Download PDF

Info

Publication number
US20210142907A1
US20210142907A1 US17/156,705 US202117156705A US2021142907A1 US 20210142907 A1 US20210142907 A1 US 20210142907A1 US 202117156705 A US202117156705 A US 202117156705A US 2021142907 A1 US2021142907 A1 US 2021142907A1
Authority
US
United States
Prior art keywords
patient
surgical
surgery
physiological data
predicted
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US17/156,705
Inventor
Asheleigh Adeline Mowery
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US17/156,705 priority Critical patent/US20210142907A1/en
Publication of US20210142907A1 publication Critical patent/US20210142907A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/20ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/10Computer-aided planning, simulation or modelling of surgical operations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/042Knowledge-based neural networks; Logical representations of neural networks
    • G06N3/0427
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H20/00ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance
    • G16H20/40ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance relating to mechanical, radiation or invasive therapies, e.g. surgery, laser therapy, dialysis or acupuncture
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/10Computer-aided planning, simulation or modelling of surgical operations
    • A61B2034/101Computer-aided simulation of surgical operations
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • A61B2090/364Correlation of different images or relation of image positions in respect to the body
    • A61B2090/365Correlation of different images or relation of image positions in respect to the body augmented reality, i.e. correlating a live optical image with another image
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/50Supports for surgical instruments, e.g. articulated arms
    • A61B2090/502Headgear, e.g. helmet, spectacles
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/30Surgical robots
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/01Dynamic search techniques; Heuristics; Dynamic trees; Branch-and-bound
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N7/00Computing arrangements based on specific mathematical models
    • G06N7/01Probabilistic graphical models, e.g. probabilistic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Public Health (AREA)
  • Medical Informatics (AREA)
  • Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • Surgery (AREA)
  • Software Systems (AREA)
  • Biophysics (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Computing Systems (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Primary Health Care (AREA)
  • General Engineering & Computer Science (AREA)
  • Epidemiology (AREA)
  • Pathology (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Databases & Information Systems (AREA)
  • Robotics (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Animal Behavior & Ethology (AREA)
  • Veterinary Medicine (AREA)
  • Urology & Nephrology (AREA)
  • Medical Treatment And Welfare Office Work (AREA)

Abstract

A method of monitoring and treating a patient using physiological data, a computer learning system, and a predictive model. The method may include generating predicted physiological data for a patient that is compared to a predictive model. When the predicted physiological data is comparable to the predictive model, preemptive care is administered.

Description

    CROSS REFERENCES TO RELATED PATENT
  • This application is a continuation of prior application Ser. No. 16/199,030, filed on Nov. 23, 2018.
  • TECHNICAL FIELD
  • This invention relates to the improvement of the effectiveness of a surgeon's decision during an operation by predicting the procedural success rates and possible complications that could arise from the operation. Furthermore, a deep learning model can predict in real time the operating costs of certain procedures, prevention of medical fraud, life expectancy of patient, and, anesthesia time left until patient awakens. Finally, the next surgical steps can be predicted by a vision system using deep neural networks which can further enhance the medical care given during a complex medical operation.
  • BACKGROUND ART
  • Since the beginning of surgical history, doctors make their decisions based on their clinical and surgical experiences. Surgeons tend to provide better patient care as they age because of past experiences and by developing a basic standard of risks across every single patient based on statistical data. Although surgeons are well-respected for their level of education, surgeons don't always pick the perfect procedure to implement given a patient's condition. They also don't come across certain experiences that another surgeon may have in his or her career. Surgeons are human beings and prone to make mistakes as any normal human being would. Yet, computers do not make mistakes and can potentially acquire knowledge equivalent too and eventually beyond a surgeon's capabilities by using artificial intelligence.
  • If there was supporting data showing the best possible way to proceed at any point during a surgery, a surgeon would clearly pick this best path. This data could change the course of a case if the surgeon knew the risks during surgery and it could change the outcome given that the surgeon would make a certain decision or not based on the risks presented. However, surgeons don't currently have this kind of data in the operation room and it becomes more complicated as the surgery commences because the patient's condition can quickly change at any moment. Yet, if surgeons could have some technology with this knowledge that can also take into account any sudden changes in the patient given the vitals of that patient or the particular step in the surgery, then this can greatly change the course of patient care and save more lives.
  • There has been a rise of robotic surgeries in the past decade with robots making the surgical cut as opposed to human beings. The Da Vinci machine has been used in order to make more efficient and smaller cuts. In terms of cosmetic look, it's been a popular opinion to chose a more minimally invasive approach such as the Da Vinci to prevent larger permanent scars. It also prevents humanly mistakes such as an accidental cuts or not precise cuts from occurring. While the surgeon is still the brain behind the machine, the machine is making the physical cut. Another application just on the rise is using machine learning to help the surgeon make a surgical decision as opposed to just surgeons making their own decisions. Similar to the use of the Da Vinci machine, surgeons are using technology but they are still making the decisions. There probably won't be technology completely taking over surgeon's jobs for years to come. While machine learning takes data and can make predictive algorithms, it cannot recognize its own issues within its own algorithm without human assistance. However, with the introduction of deep learning, technology is acting more and more like the human brain continuing to learn and gather information becoming smarter and smarter over time. Deep learning is better than machine learning because it can recognize potential issues within its own system as more information is input and correct these issues without humans. It is actually a better fit in terms of healthcare since healthcare is widely founded on statistics and data. Every procedure done had to be successfully conducted thousands of times in order to be trusted and continuously be used. Surgeons and doctors use this historical data before beginning any procedure. Likewise, deep learning works in the sense that it interprets data representation and mass data to build highly accurate models.
  • Surgeons often use an FDA approved risk calculator to decide whether or not the surgery is worth the risk of someone's life. This calculator also includes other complications that could possibly arise during or after surgery. Millions and millions of entries are entered everyday into this calculator which ultimately builds its accuracy. However, surgeons usually do this before consulting a particular patient on their condition and before actually operating on him or her. An anesthesiologist or surgery could pull up this calculator during surgery. However, it is impractical to enter in all the patient's vital statistics into this calculator continuously during the surgery and it's more of a generalization as opposed to recognizing what can occur each step within a case. Surgeons are therefore lacking live statistics that can note the potential risks or complications that could arise given a specific time point in surgery. Moreover, the risk calculator is based on historical patient history, but does not take in account current patient changing risks as the surgery is performed. It also doesn't take into account any changes in the patients state between the time is entered into the calculator and the present time. There is an assumption that the patient stays the same which may not necessarily be true. In addition, the risk calculator is more of a historical average versus taking into account the extremes within the population which happens at both ends of the population. These extremes cause unpredicted risks in a procedure that the current risk calculator cannot usually predict.
  • A major issue that medicine faces would be the extremes that exist on the mass population curve, also known as the p-curve. Most people exist in the middle of the p-curve and this is where surgeries don't necessarily have “random” complications. If someone dies from the middle of the p-curve, there is usually a clear reason as to why this happened. However, the people that exist on the ends of the curve are the inexplicable cases or the certain procedures that went awry and it wasn't until the death that there was a clearer explanation. For example, there has been situations where a patient dies because of an allergy to penicillin that the patient didn't know about prior to surgery. With technological advancements, the p-curve will widen and less people will be at the ends of the curve. These “random” complications will soon not be denoted as random because of this predicative model proposed. It might be surprising to see that the extreme cases might have similarities to each other in which they can potentially be compared to each other. Eventually, the extreme can also be analyzed alongside someone in the median population who could have characteristics like someone labeled as an extreme. Therefore, there is a potential for the p-curve to widen and transfer those on the extreme end to the middle of the curve.
  • Some science fiction movies have introduced an interesting concept where currency is built around life expectancy. For example, someone's life is projected onto their wrists and either time can speed up or slow down depending on how time was spent. In a similar manner, imagine if someone's life can be counted down. Well human beings aren't omnipotent, but what if life expectancy can be projected when someone is in critical condition such as on an operation table. Impossible some might say to know when someone is going to die. Surgeons have spent centuries trying to estimate how much time someone has left to live based on statistics and data. Yet, they have been focusing on this data prior to or after surgery as opposed to on the table. However, the current patent described herein can predict someone's life expectancy to the last second in real time while the patient is in surgery. Furthermore, the patent described herein can provide real time risk percentages based on the surgeon's selected procedure and the possible complications that could arise given a chosen path.
  • This presented system will work similarly to 1980's personal computer video games such as the labyrinth 3d video game. Originally, these video games were developed to be three dimensional but actually they were programmed to be two dimensional with set three dimensional views. This invention will work similarly since there are limited amount of variable and choices to be made in a procedure especially because most procedures are done the same and are repeated. In a labyrinth, it started out where there was only two directions that the player could proceed in. However over time, more choices could be made by the player as the game was developed such as maybe going diagonally, backwards, and grabbing objects. In a similar matter, the surgical decision system can work where it starts at a simplified level and grows to include more variables and add more layers to the matrix.
  • Each time a surgery begins, a labyrinth is basically drawn out which represents the steps taken to get to the final end goal of having both a healthy and alive patient. Each labyrinth is drawn specifically relative to that patient, their vitals, what specific point they are in surgery given the video system and usage of object recognition, and the matching logic to other patient's similar to him or her. The labyrinth can change at any point but the system learns what is the best path to take to the end. This logic can be applied to anesthesia time, life expectancy, anesthesia medication dosages which are all variables added on to the labyrinth. This can also apply to the later suggested topic of using scans to find the best pathway to location of surgery in procedures such as the minimally invasive TAVR.
  • Today's operating room consist of data that is separated into different servers and devices. This makes comparing data very difficult. The operating room is connected to the cloud, but the security concerns of moving patient data is prevalent throughout our society. Moreover, there is not a direct and dynamic displaying of the patient's expense per procedure during an operation. This patent attempts to solve these issues so that a physician has a heads up display of the surgical risks, operating expenses, and schedule of events such as time left for patient to awake.
  • To date, there is no prior art that allows a real time predictive system that optimizes surgical decisions based on historic patient data, real time operating data, cost, prior surgical decisions, and provides an autonomous surgical prediction based on real time and historical data like previous operating procedures captured with video. The present invention learns the operating room procedures using video, sound, and surgical instrument's geotags without direct input from the operating room staff. This allows a real-time ongoing predictive model. Furthermore, the invention described herein builds a plurality of predictive models instead of just one for a faster and more accurate result. The predictive model is chosen at a certain step in the surgery and used for the most accurate prediction. Some attempted solutions have tried building encoded neural networks using Deep Learning, but none of them have a methodology to pick a different model during a medical procedure based on the current medical decision.
  • DISCLOSURE OF INVENTION
  • These and other problems are generally solved or circumvented, and technical advantages are generally achieved, by preferred embodiments of the present invention that provides for an advanced predictive method for surgery, specifically to a mechanism for predicting surgical decisions, procedural success rates, surgical complications, pharmaceutical decisions, and cost analysis using deep learning techniques.
  • Research has shown that the success of an operation depends on the patient's personal characteristics and the ability of the physician. The risks of a complex operation is based on the patient's heredity factors, health risks, ongoing health concerns, and unknown health issues. This patent described herein explains a method of how to lower risks and provide the best path forward during an operation. It will also shorten the length of surgeries especially when picking the next operational procedures as the surgery progresses. Instead of taking time making a firm decision, this program will help the surgeon make the best decisions based on a deep learning model.
  • The presently preferred embodiment of the present invention utilizes a neural network on a processing device that has been trained on the historical patient's physical data including hereditary factors, past medical history, and ongoing health concerns utilizing deep learning methods. This processing device will also be trained with surgical decisions using previous surgical videos with sound and deep learning algorithms similar to autonomous driving training with graphical processing units such that the processing device can generate a plurality of predictive models.
  • The second part of the presently preferred embodiment of the present invention is a method to choose a model from the previously generated plurality of predictive pre-generated models in correspondence to the current surgical decision. The method consist of a surgeon providing surgical decisions during the operation. The surgical decisions will be used to search for corresponding single predictive model from the plurality of predictive pre-generated models that references the current surgical decision.
  • The third part of the presently preferred embodiment of the present invention is a method to securely evaluate the current patient's vital statistics and physician's procedural decision versus the chosen trained model. This evaluation using the chosen trained model which is dynamically calculated provides a predictive cost, life expectancy, patient risk, anesthesiology timeline, and surgical procedure timeline which can be viewed securely by the physician or operating staff.
  • The fourth part of the presently preferred embodiment of the present invention is a detailed patient progress report which includes specific information regarding the patient's surgery including ongoing cost, procedural risks, predictive allergies or health concerns, life expectancy, anesthesiology timeline, and next suggested procedure. The fifth part of this presently preferred embodiment is the prediction of the next procedure that the surgeon will take based on ongoing operational awareness using either video, sound, mouse, keyboard or geotags on surgical tools or a combination of all these inputs. As soon as one particular path is chosen by the surgeon, the life expectancy clock will be updated and a suggested next procedure or pharmaceutical medicine to be given will be chosen by the deep learning algorithm.
  • Accordingly, besides the objects and advantages of an surgical decision making using deep learning described herein, and additional objective and advantage of the preferred embodiment of the present invention is to provide a secure method of interchanging data from different servers and medical devices using a secure blockchain communication method.
  • Another additional objective and advantage of the preferred embodiment of the present invention is to provide a secure method of interchanging data from different servers and medical devices using a secure point to point communication method, encrypted communication method, or blockchain method.
  • Another objective and advantage to this invention is to decrease actual surgery time because a procedural guideline with associated risks can be displayed to the surgeon as opposed to searching for a small detail about the patient that would not be immediately known.
  • Another objective and advantage to this invention is to predict the amount of time in each section of the surgery and the cost associated with each section. It also could predict the time it would take to complete other suggested procedures with corresponding cost to complete. This would allow the surgeon to provide the lowest risk care with the associated cost. Today surgeon's chose the best procedure for their patient, but in the case where two or more procedures were essentially equivalent, this would provide the surgeon with the lowest cost procedure with the fastest completion time.
  • Another objective and advantage to this patent is to assist surgeons similar to autonomous driving. Today's cars assist drivers so that there is limited intervention. This patent will allow surgeons to focus on the surgery while getting live predictive data that dynamically changes with their procedural steps in the operation. In the future, just like with autonomous driving, surgeons will use this patent with robotic assistance to completely have autonomous surgery. This will allow the surgeon to intervene only when absolutely necessary allowing the processing device to complete the surgery independent of all human assistance. Essentially, this patent is a stepping stone to autonomous surgery.
  • Another objective and advantage to this patent is to identify possible risks that are not known about the patient by the surgeon. These risks could be allergies, hereditary traits, hereditary risks, or other complications that came up with a patient's family members during their surgeries. Most of the time surgeries proceed as normal, but sometimes risks occur that need expedited resolution, however, vital information may not be known by the surgeon since the patient is unconscious.
  • Another objective and advantage to this patent is to identify the potential for a patient to awaken from under a general anesthesia. It has been shown that patients who are asleep are sometimes awake during an operation. This can be a fearful event for both the patient and anesthesiologist. However, a predictive model using patient data and operating data such as brain waves could help identify when the patient is not in a deep sleep.
  • Another objective and advantage to this patent is to identify when a patient is not within the normal population of previous patients. This will help the surgeon understand that there is additional risks with the current operation. This patent will help match data to drive a predictive model of relative characteristics of a patient that could drive major complications and timelines such as surgical procedures, life expectancy, time to awake, anesthesia medication dosages throughout the surgery, and operating cost.
  • Another objective and advantage to this patent is to use a combination of computer vision and geotags that will recognize and take into account each surgical tool or medication that is used in the surgery. This will lessen the amount of time to prepare for surgery and inventory will no longer need to be done since it will already be taken into account for. This inventory will also take part in the running cost of the surgery. This can also apply to medications and taking into account what medications are present. The geotags and computer vision can also insure that only an authorized user can use the instruments.
  • Another objective and advantage to patent is that wireless connectivity such as bluetooth can portably move anesthesia time on a small wireless display as the patient leaves the operation room. This will allow both the nurses and anesthesiologists to have a general knowledge as to when the patient will wake so they can plan accordingly.
  • Another objective and advantage to the patent is that the system will also be able to predict dosage sizes for certain medications at certain times throughout the surgery given by the matching historical data of other patient's. It will also be correlated to the other variables that has been discussed such as vital signs and the surgery vision system for further support. This will give the anesthesiologist more information in order to support their general knowledge surrounding dosage sizes. Also, the dosages will take part in the running costs. The system will know how much dosage is used based on flow control which will be taken account in cost and the amount of dosages that was originally predicted.
  • Another objective and advantage to the patent is that the system will not only have video cameras interpreting the surgery but also video cameras over where the anesthesiologists are. This will work in a similar manner to the other camera in which it will recognize certain medications based on packaging or label names and take this information into account in the system. It is also acts as a prevention of medical fraud where an anesthesiologist cannot accurately record what medications were used and when they were used during surgery. This will take part in the running cost calculations.
  • Another objective and advantage to patent is that the system is customizable for each surgeon. The surgeon is recognized either thru voice or video recognition. The system is that calibrated towards the surgeons operating room settings. Also, the system could recognize the patient in a similar fashion and have the operating room be setup for their specific characteristics such as temperature of the room. Surgeons also usually like to play music while operating so the system can upload music that the particular patient likes to work to. The system will also be able to learn not only what type of music the surgeon likes, but also the environment the surgeon likes to work best in. Also, the system will soon be able to be more customizable for different specialties. For example, the anesthesiologist can have a version or tab with information relevant to them such as predicted and anesthesia time or predicted dosages. Cardiology will be able to have adaptations such as for TAVR where a routing system might be needed. This can also apply to neurology which is generally a very complex and tedious specialty. Therefore, it's important to be very careful in brain surgery since one slight movement in the wrong place can make major damage.
  • An additional embodiment of this present invention is to provide a program will include certain possible procedures to choose from given past data and success rates and failure rates. The program will rank the procedures based on relevancy and the best possible routes to proceed in. As soon as one particular path is picked, a life expectancy clock will be projected. There will be a separate tab to input the particular patient's statistics prior to surgery such as gender, age, health conditions, past surgeries, present health concerns. This will be useful also to quickly reference and act as the patient's chart. It will also be important so that past patient data can be matched up with the current patient based on similar characteristics. This will allow there to be a list the potential complications given past mass data. Vocal recognition can also be applied to this tab if the surgeon needs to know a basic statistic about the patient. Again, it will also waste less time as opposed to searching for a small detail about the patient that wouldn't be known immediately. The invention will also use its own interpretation of scans and a predicted life expectancy associated with a given a certain procedure. These scans will actually be read more thoroughly and accurate as opposed to a human brain would despite experience. In the future, a camera can be linked up to the program to receive live data images of both medical scans, such as CAT scans or an echocardiogram, and live video footage which will increase the accuracy of the both the life expectancy, procedures to give, a possibility of the patient coming out of anesthesia, and major complications that could arise. Vocal recognition will be used for easy access such as asking a question of how to proceed in a given situation or specific knowledge that needs to be known about the patient anytime throughout the case. As soon as a certain step in the procedure is chosen, the program will make its adjustments and take this into account especially in terms of the predicted life expectancy clock and other variables like anesthesia time, dosage amount, risk percentages, major complications that could arise, and surgical decisions to make given the situation.
  • Another objective and advantage of the additional embodiment of the patent is a projected path for surgeries that are guided with scans using the surgeries live video feed or cat scans. This can apply to certain specialties like cardiology or in neurology. In cardiology, the TAVR procedure, valve replacements are done through the femoral artery, are becoming more common but there isn't necessarily a clear path on the scan of how this is done. Therefore, this system will provide guidance with a highlighted projected path with also vocal recognition that will provide directions to the intended destination. This will be done based on the matching logic to similar patients and use similar logic to labyrinth metaphor addressed earlier. Also, the program will use object recognition in order to recognize a particular route it should proceed given that the program will be fed many scans into the system of healthy, blocked, or narrow arteries. This will allow for the combination of surgery rooms video system and live feed inside the patient for greater accuracy of the machine learning system.
  • Another objective and advantage of the additional embodiment of the patent is to provide complications that could arise for going the wrong path inside a patient. These complications could be shown on the display and warn the surgeon of mistakes during surgery. The system could provide new directions to a final destination similar to navigation systems in a car. The system could also provide rerouting options when a obstacle becomes apparent. This works similar to the traffic logic in the navigation system where the best route is given. However, it will be up to the doctor to proceed in a certain direction and the system will adjust accordingly. This also is a good teaching tool for younger doctors without as much experience as well. However, it can be used at any experience level of a health professional. The system will take this route into account in terms of life expectancy and any complications that could arise especially if the wrong route is taken.
  • Another objective and advantage of the additional embodiment of this patent is that the video interpretation system will record what is seen in the surgery in text form which will eventually eliminate the need for doctor's notes. Doctors can add additional information at the end of surgery but it won't be necessary since this system saw what everyone in the room did. This reduce the amount of workload for health professionals. Given that surgeons already spend so many hours in the operating room, this will also decrease their long work hours throughout the week. This will give the surgeons more rest time and possibly more opportunity to focus in the operating room as opposed to spending time writing these notes. Also, medical scribes in the operation room will no longer be needed. A health care technical word bank can be input into the system and the system can be taught how to write like a professional. This documentation will also be more accurate than a human being trying to recall what happened based off of memory. Furthermore, this information will take part in matching other patients to each other. The system will then become more efficient with time and experience. It will soon be able to interpret its own data input as opposed to human information put into the system. This will also prevent doctors from inaccurately reporting what happened in surgery thus preventing medical fraud. Furthermore, there will be less room for grammatical error that can lead to misinterpretation of data.
  • Another objective and advantage of the additional embodiment of this patent is to use DNA or hereditary searches for the patient so that a more exact model can be used. The system could use patient data or DNA data to pull up records and data to assist the accuracy of predictions.
  • Another objective and advantage of the additional embodiment of this patent is to use a combination of surgical data, patient data, and patient hereditary data to predict blood clots and to provide surgical decisions to overcome this obstacle.
  • Still further objects and advantages will become apparent from a consideration of the ensuing description and drawings. Similar constructions that do not depart from the spirit and scope of this invention set forth in the claims or embodiment should be considered the equivalent.
  • BRIEF DESCRIPTION OF DRAWINGS
  • For a more complete understanding of the present invention, and the advantages thereof, reference is now made to the following descriptions taken in conjunction with the accompanying drawing, in which:
  • FIG. 1 is a diagrammatic illustration of a system for surgical decisions using deep learning; and
  • FIG. 2 is a diagrammatic illustration of a semi autonomous surgical system with limited surgeon intervention that is used in combination with a deep learning algorithm; and
  • FIG. 3 is a diagrammatic illustration of the lowest risk surgical path which depends on a predictive deep learning model.
  • MODE(S) FOR CARRYING OUT THE INVENTION Detailed Description of Best Mode for Carrying Out the Invention
  • The making and using of the presently preferred embodiments are discussed in detail below. It should be appreciated, however, that the present invention provides many applicable inventive concepts that can be embodied in a wide variety of specific contexts. The specific embodiments discussed are merely illustrative of specific ways to make and use the invention, and do not limit the scope of the invention.
  • The present invention will be described with respect to preferred embodiments in a specific context, namely an advanced predictive method of surgery, specifically to a predictive mechanism for procedural risks, operating costs of certain procedures, life expectancy, and anesthesia time left until patient awakes.
  • Operation of Best Mode for Carrying Out the Invention
  • Referring now to the drawings in detail, and initially to FIG. 1, a system for surgical decisions using deep learning. A patient (101) is having an operation on a specific medical issue. The patient (101) is scanned (102) for vital statistics (103) for the ongoing operation. Historical data (104) consisting of previous surgical vital statistics, hereditary data, prescription drug history, and ongoing medical issues may be found in local computer or cloud records and submitted (106) to the deep learning processor (111). The deep learning processor (111) will generate a plurality of predictive models (112) based on the submitted historical data (104). A surgeon (107) provides (108) surgical decisions (107) during the operation submitted (110) to a deep learning processor (111). The surgical decisions (107) will be used to search for corresponding single predictive model that references the current surgical decision. The next predicted surgical decisions can be preloaded in the deep learning processors to increase speed of calculation. The vital statistics (103) is used as input (105) to a deep learning processor (111) that is encoded with the single predictive model (112). The predictive model (112) provides an output corresponding to the lowest risk surgical decision, time expectancy, time for patient to wake, waiting room updates, and surgical risks displayed on an output like a display, augmented reality glasses, or mobile device.
  • Referring now to vital statistics (103) in detail in FIG. 1. Vital statistics could be sensors reading information like oxygen level, brain patterns, electrocardiogram, blood content, heart scans, body scans, organ scans, muscle scans, and bone scans.
  • Referring now to deep learning processor (111) in detail in FIG. 1. A deep learning processor utilizes a neural network to recognize patterns that works well in a systematic setting. The operating room is filled with systematic procedures in which a certain procedure is repeated over and over again. In the operating room, there are displays that project the operations on the screens to assist everyone in the room with having a better view of the operation. It becomes more difficult to see when more and more people surround the patient. Therefore, there is video cameras that basically show what the surgeon is seeing. There will also be other video camera interpreting and recognizing the medications that the anesthesiologists are using. From this footage, the program could live footage as to how the operation is preceding. It'll be able to interpret this in a similar manner as to how autonomous cars read signs on the street. This is known as computer vision basically through the use of neural networks to interpret data. Billions of different versions of the same image are input into a program and then the computer is able to interpret what it is. Videos are basically a sequence of images but together so this program will be able to interpret these images in the chronological order it is put into the system. In this manner, the program will be able to further predict the next step in the surgery. As the program becomes more and more intelligent, it'll be able to interpret this footage and apply it to its statistics and calculations. As of now, the program will use past data and data input before the surgery to assist the surgeon. However, in time the program can evolve to taking both past and live data into account. In a similar way to a child learning, the program will soon learn to become as knowledgeable or even more knowledgeable as a surgeon. As more surgeries are input, the smarter the program will become and apply this knowledge to past historical data from patient files. If the videos taken in surgery are somehow saved and specified under each patient, it'll make the program's predictions more accurate. Eventually this program will be able to predict possible complications from live data before it happens with indications or warnings of something possibly going wrong before even the surgeon notices. This can also be applied in terms of anesthesia where there can be warnings as to the patient possibly coming out of anesthesia.
  • Referring now to historical data (104) in detail from FIG. 1. Historical data can be related to other patient's surgery information that is available for processing. This invention could also use technology in order to read input scans which contains files of the patients as well as the scans. Epic software is an example of a form of data storage for all the patient information in a hospital, local regional area, or nationally connected medical centers. This software is password protected for only use of a healthcare professional as well as only being used in a hospital setting. Epic software's setup varies with each specialty as well as at different hospitals. Away to possibly gain information to be used in the deep learning model is through removal of the patient's names because that would follow HIPPAA protocol. The program will only be used in a hospital setting and not taken out of that setting for any other use. Therefore, professionals will be using this information and they already have an agreement to keep personal information confidential for each patient. Eventually, the goal is to have all the statistics in every operating room in the United States and maybe even one day around the world. As stated elsewhere, the more information the program has, the more accurate the diagnosis will be. Even though human beings are all different, they are very much the same in many ways.
  • Referring now to surgical decisions (107) in detail from FIG. 1. Surgical decisions may be entered through a mouse, keyboard, video interpretation or audio commands. Surgeons often dictate the certain steps they are going to do while operating to direct everyone in sort of a team fashion. It is sort of a follow the leader dynamic. In which, the leader is the surgeon and he or she is dictating how to go about the particular surgery. Their hands are preoccupied with doing the surgery. So, the best way to use the program would be through vocal recognition similar to that used with the common voice recognition software on mobile phones or standalone voice devices. This program will act as another tool to help the surgeon without interrupting his flow of work. It will also be used as a form of multitasking with the priority still being the patient.
  • Referring now to deep learning processor (111) and plurality of predictive models (112). The making and using of a deep learning processor or deep learning model is essential to this patent. However, one skilled in the art of deep learning can develop this model using advanced GPU processors. It should be appreciated, however, that the present invention provides many applicable deep learning concepts that can be embodied in a wide variety of specific contexts. The specific embodiments discussed are merely illustrative of specific ways to make and use the invention, and do not limit the scope of the invention.
  • Referring now to an output (113) in detail in FIG. 1. After a certain procedure is chosen, a life expectancy clock will be provided to the operating room. The patient's predicted life expectancy will be displayed based on the branching logic and ultimately the deep learning analysis. The life expectancy clock can change as the deep learning algorithm takes into account new data. The output (113) could be displayed on a graphical interface or augmented reality glasses that only certain hospital employees can view during an operation. Also, the output (113) could be a mobile application device running on a mobile device. Another output of this device could be waiting time and important health information to family members waiting in the waiting room. Family members could see this information and be informed of predicted time to completion of the surgery. Moreover, the mobile application could also show the predicted and actual cost of the ongoing surgery.
  • Another embodiment of the output (113) could display the predicted time life for a patient to awake which would be provided to the anesthesiologist. According to Current Procedural Terminology (CPT) guidelines, anesthesia begins when the anesthesiologist prepares the patient and then end as soon as the anesthesiologist is said to leave the room, which is also when the surgery is over. Intubation is placing breathing tubes through vocal cords to help a patient breath especially during a surgery. Intubation time is measured from when a tube is placed down a patient's trachea to assist with ventilation to when it is removed. However, extubation refers to taking the breathing tubes outside of a patient. Anesthesiologists make sure a patient can breathe on their own and usually determine this by seeing if a person is awake or following commands. This extubation time would vary depending on the complexity of a case. For example, it would occur in the intensive care unit for a cardiovascular case as opposed to in the operation room causing a lengthier extubation time. Intubation time as well as extubation time are documented in patient records. Specifically, extubation time is noted for ICU cases such as those for cardiovascular surgery. Extubation is usually done when a patient is awake and following commands. This would indicate that a person is able to breathe on his or her own. Length of a case usually can correlate to how long a person is awake or following commands. However, this is not always the case. It usually refers to when a patent is in the operation room. Intubation time, extubation time, and length of a case are documented in patients' records. All of which are variables that help predict when general anesthesia wears off and the patient wakes up. There is a general time frame as to when a patient should wake up but there is yet a way to determine an exact time until he or she wakes up.
  • Anesthesiologists are trained to know the dosage sizes based on the medical literature and they formed their reasoning on dosages based in experiments in which mass data of the effect of the particular dosages were recorded. Similar to the risk calculators, anesthesiologists are using statistics to make their decisions. There is also a pump which gives a recommended dosage of each drug to the patient. The anesthesiologist acts and is the final decider on whether the prediction is accurate. As opposed to the life expectancy clock, which will record a patient's predicted life expectancy in minutes. An anesthesia clock could predict a time in which a patient will awake based on various variables and a deep learning model. Some of theses variables can include length of a case, intubation time, extubation time, medication dosages, family medical history, historical records of other patients, and an individual's statistics all of which are documented. This clock will give the anesthesiologist extra reassurance that their dosage calculations are correct. However, the clock may not be able to account for everything like genetics could play a part. This is a lacking knowledge that anesthesiologists are trying to figure out the influences of dosages on some certain type of genetics. Since past data will be recorded in minutes and protocol for billing is all based on minutes, the clock will project out a time. Anesthesiologists use either online calculators or hand calculations of dosages to give the particular patient prior to surgery. The anesthesia time is recorded in patient records as a main form of billing, but it can also be used to take for a more efficient time accuracy. Alongside the calculations made by the anesthesiologist, this program will have past data of anesthesia times of other patients. As described earlier, the program will match the particular patient to another patient and from that, display the time. Ultimately, the anesthesia time will be more accurate as opposed to just having the calculation from the anesthesiologist. This will provide more efficiency and a less probability of a patient coming out of anesthesia in the middle of a surgery. This will also lessen the risk of the patient dying as well as the probability for a lawsuit especially since anesthesiologists are said to be sued frequently. However, this is just a prediction and nothing is for sure. It just gives a more accurate time duration given past data.
  • Description of Additional Mode for Carrying Out Invention
  • The making and using of the additional preferred embodiments are discussed in detail below. It should be appreciated, however, that the present invention provides many applicable inventive concepts that can be embodied in a wide variety of specific contexts. The specific embodiments discussed are merely illustrative of specific ways to make and use the invention, and do not limit the scope of the invention.
  • The present invention will be described with respect to an additional embodiments in a specific context, namely surgical decisions using deep learning to predict life expectancy, surgical decisions, and risks as shown in FIG. 1 using branch prediction or probability tree. A probability tree could be used in conjunction with the deep learning algorithm so as to filter out unrelated patient or historical data. The system will use a probability tree to search through the input patient records and match the current patient to its best match. Similar to the logic of looking for an organ door or blood donor, the program will look for the closest match. Therefore, the more patient records input into the program, the more accurate the match will be. There are billions of people in the world and although everyone is different, they are very much the same in many ways. It is also similar to the logic of a dating website. A dating website usually works in which a person inserts their basic information such as their hobbies, interests, or occupation. In the same way, patient's stats will be entered. The program will use branching logic to match the patient to a match. It'll first start at a characteristic such as gender, then branch off to age, branch off to past medical history including hereditary and DNA, then branch off to smaller details until it finds the closest match based on past records. Eventually, it will reach a branching point where it'll try to match a certain procedure to the original procedure input into the program. Eventually, the program can evolve to take into account a quick change of procedure to apply to the deep learning portion of the invention. After the branching logic commences, the program will get to a point where procedures will be chosen. Based on the pick, a clock will start running after it—takes into a count the branching logic. This is where a majority of deep learning commences. Deep learning is a form of data representation in time analysis of deep learning. From this logic, the life expectancy clock and the anesthesia clock will run.
  • Operation of an Alternate Mode for Carrying Out the Invention
  • Referring now to the drawings in detail, and specifically to FIG. 2, which is a semi-autonomous surgical system with limited surgeon intervention that is used in combination with a deep learning algorithm. A patient (201) is having an operation on a specific medical issue. The patient (201) is viewed (202) by an operation video camera (203) and scanned (202) for vital statistics (203) for the ongoing operation. Historical data (204) consisting of previous surgical vital statistics, hereditary data, prescription drug history, and ongoing medical issues may be found in computer or cloud records and submitted (206) to the deep learning processor(s) (211). The deep learning processor(s) (211) will generate a plurality of predictive models (212) based on the historical data (204). Moreover, a surgeon (207) completing the operation inputs (208) his surgical decisions using an input device (209). The input device (209) could be keyboard, mouse, voice, or actual video of the operation. The input device (209) also is feed into the neural network (211). The surgical decisions (208) will be used to search for corresponding single predictive model that references the current surgical decision. The next surgical decisions can be preloaded in the deep learning processors to increase speed of calculation. The video and vital statistics (203) is used as input (206) to a deep learning processor(s) (211) that is encoded with the single predictive model (213). The predictive model (213) provides an output corresponding to the lowest risk surgical decision, time expectancy, time for patient to wake, waiting room updates, and surgical risks displayed on an output like a display or mobile device.
  • Referring now to neural network (211) and predictive model (212). The making and using of a neural network or deep learning model is essential to this patent. However, one skilled in the art of neural network or deep learning can develop this model with the assistance of advanced GPU processors. It should be appreciated, however, that the present invention provides many applicable deep learning methods that can be embodied in a wide variety of specific contexts. The specific embodiments discussed are merely illustrative of specific ways to make and use the invention, and do not limit the scope of the invention.
  • Referring now to FIG. 3, which relates to the methodology of deciding which encoded model is loaded into the neural network, the initial procedure 1 encoded model (301) is loaded into the neural network. A patient begins the operation and the patient's vital statistics and operation inputs are passed into the procedure 1 encoded model (301). Procedure 2 encoded model (304) is predicted (302) by procedure 1 encoded model (301). Also, procedure 3 encoded model (305) is predicted (303) by procedure 1 encoded model (301). Both procedure 2 encoded model (304) and procedure 3 encoded model (305) can be preloaded into a neural network for the next surgical decision. Based on the doctor's direction though video or input device, procedure 2 encoded model (304) or procedure 3 encoded model (305) will be chosen. If procedure 3 encoded model (305) is chosen, it is the last procedure in the patient's operation. Based on the number of procedures in an operation, a prediction on length of time can be generated. If procedure 2 encoded model (304) is chosen, the patient's vital statistics and operation inputs are then passed into the new model. Procedure 4 encoded model (307) is predicted (306) by procedure 2 encoded model (304). It is the only procedure model predicted. The patient's vital statistics and operation inputs are then passed into the new chosen procedure 4 encoded model (307). Procedure 5 encoded model (308) is predicted (310) by procedure 4 encoded model (301). Also, procedure 6 encoded model (311) is predicted (309) by procedure 4 encoded model (301). Both procedure 5 encoded model (308) and procedure 6 encoded model (311) can be preloaded into a neural network for the next surgical decision. Based on the doctor's direction though video or input device, procedure 5 encoded model (308) or procedure 6 encoded model (311) will be chosen. If procedure 5 encoded model (308) or procedure 6 encode model (311) are chosen, it is the last procedure in the patient's operation.
  • INDUSTRIAL APPLICABILITY
  • Accordingly, the industrial applicability of this patent is for a sellable biomedical device allowing for reliable surgical decision using deep learning, specifically to an autonomous surgical prediction using video and audio deep learning of the next surgical decision.
  • SEQUENCE LISTING
  • Not applicable

Claims (14)

I claim:
1. A method of monitoring and treating a patient using physiological data, a computer learning system, and a predictive model, the method comprising:
inputting into the computer learning system reference physiological data for previous patients that experienced a surgical complication;
generating, in the computer learning system, a predictive model corresponding to the reference physiological data for patients that experienced the surgical complication;
inputting into the computer learning system a patient's current physiological data that is contemporaneously obtained from a physiological monitor during a surgery;
generating predicted future physiological data for the patient based on the patient's current physiological data and the predictive model;
comparing the patient's future predicted physiological data to the predictive model; and
if the patient's predicted future physiological data is comparable to the predictive model, then administering preemptive care for the surgical complication.
2. The method of claim 1, wherein the computer learning system comprises a deep learning processor.
3. The method of claim 1, wherein administering preemptive care for the surgical complication is selected from the group consisting of omitting surgical actions, administering medication, conducting diagnostic testing, and performing surgical actions.
4. The method of claim 1, further comprising predicting a date and time when the surgical complication will occur based on the comparison of the patient's predicted physiological data with the predictive model.
5. The method of claim 4, further comprising, if the surgical complication is predicted to occur during a surgery, then administering preemptive care during the surgery.
6. The method of claim 5, wherein administering preemptive care for the surgical complication during surgery is selected from the group consisting of administering medication, avoiding surgical actions, conducting diagnostic testing, and performing surgical actions.
7. The method of claim 4, further comprising providing the surgical complication and the predicted date and time on a display for viewing by a medical provider.
8. The method of claim 7, wherein the predicted date and time are provided on the display in the form of a countdown clock.
9. The method of claim 1, wherein the patient's physiological data is selected from the group consisting of pulse, oxygen level, brain wave patterns, electrocardiogram data, blood test results, and organ scans.
10. The method of claim 5, further comprising providing to a medical surgical participant, on a display, identification of the preemptive care to be administered during surgery.
11. The method of claim 5, further comprising generating and providing to a medical surgical participant, on a display and based on at least the surgical complication, identification of contraindicated actions that should not be taken during surgery.
12. The method of claim 1, wherein administering preemptive care comprises the implementation of non-surgical actions and the avoidance of further surgical actions.
13. The method of claim 1, wherein the surgical complication is organ failure.
14. The method of claim 1, wherein the surgical complication is stroke.
US17/156,705 2018-11-23 2021-01-25 System for Surgical Decisions Using Deep Learning Pending US20210142907A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US17/156,705 US20210142907A1 (en) 2018-11-23 2021-01-25 System for Surgical Decisions Using Deep Learning

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US16/199,030 US10930395B2 (en) 2018-11-23 2018-11-23 System for surgical decisions using deep learning
US17/156,705 US20210142907A1 (en) 2018-11-23 2021-01-25 System for Surgical Decisions Using Deep Learning

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US16/199,030 Continuation US10930395B2 (en) 2018-11-23 2018-11-23 System for surgical decisions using deep learning

Publications (1)

Publication Number Publication Date
US20210142907A1 true US20210142907A1 (en) 2021-05-13

Family

ID=70771720

Family Applications (2)

Application Number Title Priority Date Filing Date
US16/199,030 Active 2039-03-20 US10930395B2 (en) 2018-11-23 2018-11-23 System for surgical decisions using deep learning
US17/156,705 Pending US20210142907A1 (en) 2018-11-23 2021-01-25 System for Surgical Decisions Using Deep Learning

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US16/199,030 Active 2039-03-20 US10930395B2 (en) 2018-11-23 2018-11-23 System for surgical decisions using deep learning

Country Status (1)

Country Link
US (2) US10930395B2 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11416247B1 (en) * 2021-02-09 2022-08-16 International Business Machines Corporation Healthcare application data management based on waste priority

Families Citing this family (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110678902A (en) * 2017-05-31 2020-01-10 Eizo株式会社 Surgical instrument detection system and computer program
WO2019106882A1 (en) * 2017-11-30 2019-06-06 テルモ株式会社 Support system, support method, and support program
WO2019245849A1 (en) * 2018-06-19 2019-12-26 Tornier, Inc. Mixed-reality enhanced surgical planning system
US11484384B2 (en) 2019-02-21 2022-11-01 Theator inc. Compilation video of differing events in surgeries on different patients
AU2020224128B2 (en) 2019-02-21 2021-09-30 Theator inc. Systems and methods for analysis of surgical videos
JP2022526937A (en) * 2019-03-27 2022-05-27 サノフイ Residual semi-recurrent neural network
US20210312949A1 (en) 2020-04-05 2021-10-07 Theator inc. Systems and methods for intraoperative video review
EP4147111A4 (en) * 2020-05-08 2024-05-01 Intime Biotech Llc Real-time method of bio big data automatic collection for personalized lifespan prediction
US11763947B2 (en) * 2020-10-14 2023-09-19 Etiometry Inc. System and method for providing clinical decision support
US11682487B2 (en) 2021-01-22 2023-06-20 Cilag Gmbh International Active recognition and pairing sensing systems
US11694533B2 (en) 2021-01-22 2023-07-04 Cilag Gmbh International Predictive based system adjustments based on biomarker trending
US20220233241A1 (en) * 2021-01-22 2022-07-28 Ethicon Llc Surgical procedure monitoring
CN113140318A (en) * 2021-05-10 2021-07-20 中国人民解放军总医院第三医学中心 Lung infection risk prediction method after liver transplantation based on deep learning
WO2023023509A1 (en) * 2021-08-17 2023-02-23 Theator inc. Automated analysis of video data during surgical procedures using artificial intelligence
CN115844545B (en) * 2023-02-27 2023-05-23 潍坊医学院附属医院 Vascular intervention intelligent surgical robot and control method

Family Cites Families (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6678669B2 (en) * 1996-02-09 2004-01-13 Adeza Biomedical Corporation Method for selecting medical and biochemical diagnostic tests using neural network-related applications
WO2001097909A2 (en) 2000-06-14 2001-12-27 Medtronic, Inc. Deep computing applications in medical device systems
US8392342B2 (en) * 2009-11-18 2013-03-05 Empire Technology Development Llc Method and apparatus for predicting movement of a tool in each of four dimensions and generating feedback during surgical events using a 4D virtual real-time space
US8660857B2 (en) 2010-10-27 2014-02-25 International Business Machines Corporation Method and system for outcome based referral using healthcare data of patient and physician populations
EP2637594A4 (en) * 2010-11-11 2015-05-06 Univ Johns Hopkins Human-machine collaborative robotic systems
US10582847B2 (en) * 2010-12-30 2020-03-10 Amo Wavefront Sciences, Llc Method and system for eye measurements and cataract surgery planning using vector function derived from prior surgeries
US10583039B2 (en) * 2010-12-30 2020-03-10 Amo Wavefront Sciences, Llc Method and system for eye measurements and cataract surgery planning using vector function derived from prior surgeries
US10332639B2 (en) * 2017-05-02 2019-06-25 James Paul Smurro Cognitive collaboration with neurosynaptic imaging networks, augmented medical intelligence and cybernetic workflow streams
US9536053B2 (en) 2013-06-26 2017-01-03 WellDoc, Inc. Systems and methods for managing medication adherence
US10201320B2 (en) * 2015-12-18 2019-02-12 OrthoGrid Systems, Inc Deformed grid based intra-operative system and method of use
EP3223183B1 (en) 2016-03-25 2020-11-18 Siemens Healthcare GmbH Case-based reasoning in the cloud using deep learning
US10885463B2 (en) 2016-07-08 2021-01-05 Microsoft Technology Licensing, Llc Metadata-driven machine learning for systems
US10453200B2 (en) * 2016-11-02 2019-10-22 General Electric Company Automated segmentation using deep learned priors
WO2018188466A1 (en) * 2017-04-12 2018-10-18 Bio-Medical Engineering (HK) Limited Automated steering systems and methods for a robotic endoscope
AU2018255892A1 (en) * 2017-04-21 2019-11-07 Medicrea International A system for providing intraoperative tracking to assist spinal surgery

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11416247B1 (en) * 2021-02-09 2022-08-16 International Business Machines Corporation Healthcare application data management based on waste priority

Also Published As

Publication number Publication date
US10930395B2 (en) 2021-02-23
US20200168334A1 (en) 2020-05-28

Similar Documents

Publication Publication Date Title
US20210142907A1 (en) System for Surgical Decisions Using Deep Learning
JP6949128B2 (en) system
US20220383998A1 (en) Artificial-intelligence-based facilitation of healthcare delivery
Pieterse et al. Shared decision making and the importance of time
Maddox et al. Questions for artificial intelligence in health care
Tang et al. Timing of laparoscopic surgery in gallstone pancreatitis
US20220384052A1 (en) Performing mapping operations to perform an intervention
US20210057106A1 (en) System and Method for Digital Therapeutics Implementing a Digital Deep Layer Patient Profile
CN115917492A (en) Method and system for video collaboration
Disler et al. Empowerment in people with COPD
McCoy et al. Improving early childhood development on a global scale: best practices for intervention
Pauker et al. How (should) physicians think?: a journey from behavioral economics to the bedside
US11287876B2 (en) Managing user movement via machine learning
Abbasi The science of persuasion offers lessons for COVID-19 prevention
Morrison et al. Emerging Trends in Digital Health and Care: A Refresh Post-COVID
US20230029678A1 (en) Generating clustered event episode bundles for presentation and action
US20230033160A1 (en) Generating a registry of people using a criteria and performing an action for the registry of people
Cochran et al. Framework for faculty development in resident autonomy and entrustment in the operating room
US20220367054A1 (en) Health related data management of a population
Davenport et al. Advanced Introduction to Artificial Intelligence in Healthcare
Cunha et al. Using Mixed Reality and Machine Learning to Assist Caregivers in Nursing Home and Promote Well-being
US20230162871A1 (en) Care lifecycle tele-health system and methods
Cafiero et al. Improving Healthcare by Designing AI-Human-Centered Systems
Freeman et al. The Power of Data
Price Does prayer really set one apart?

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: APPLICATION DISPATCHED FROM PREEXAM, NOT YET DOCKETED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED