WO2020073114A1 - Image processing of streptococcal infection in pharyngitis subjects - Google Patents

Image processing of streptococcal infection in pharyngitis subjects Download PDF

Info

Publication number
WO2020073114A1
WO2020073114A1 PCT/CA2019/051401 CA2019051401W WO2020073114A1 WO 2020073114 A1 WO2020073114 A1 WO 2020073114A1 CA 2019051401 W CA2019051401 W CA 2019051401W WO 2020073114 A1 WO2020073114 A1 WO 2020073114A1
Authority
WO
WIPO (PCT)
Prior art keywords
subject
images
training
pathogen
image
Prior art date
Application number
PCT/CA2019/051401
Other languages
French (fr)
Inventor
Sarbjit Sarkaria
Steven Rebiffe
Udit GUPTA
Mahendran Maliapen
Peter Whitehead
Original Assignee
Light AI Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US16/589,077 external-priority patent/US11369318B2/en
Application filed by Light AI Inc. filed Critical Light AI Inc.
Priority to AU2019357949A priority Critical patent/AU2019357949B2/en
Priority to EP19871128.5A priority patent/EP3864669A4/en
Priority to CA3115738A priority patent/CA3115738A1/en
Publication of WO2020073114A1 publication Critical patent/WO2020073114A1/en
Priority to IL282169A priority patent/IL282169A/en

Links

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/0059Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
    • A61B5/0082Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence adapted for particular medical purposes
    • A61B5/0084Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence adapted for particular medical purposes for introduction into the body, e.g. by catheters
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/0002Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network
    • A61B5/0015Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network characterised by features of the telemetry system
    • A61B5/0022Monitoring a patient using a global network, e.g. telephone networks, internet
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/0033Features or image-related aspects of imaging apparatus classified in A61B5/00, e.g. for MRI, optical tomography or impedance tomography apparatus; arrangements of imaging apparatus in a room
    • A61B5/0035Features or image-related aspects of imaging apparatus classified in A61B5/00, e.g. for MRI, optical tomography or impedance tomography apparatus; arrangements of imaging apparatus in a room adapted for acquisition of images from more than one imaging mode, e.g. combining MRI and optical tomography
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/0033Features or image-related aspects of imaging apparatus classified in A61B5/00, e.g. for MRI, optical tomography or impedance tomography apparatus; arrangements of imaging apparatus in a room
    • A61B5/004Features or image-related aspects of imaging apparatus classified in A61B5/00, e.g. for MRI, optical tomography or impedance tomography apparatus; arrangements of imaging apparatus in a room adapted for image acquisition of a particular organ or body part
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/0059Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
    • A61B5/0071Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence by measuring fluorescence emission
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/01Measuring temperature of body parts ; Diagnostic temperature sensing, e.g. for malignant or inflamed tissue
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/68Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
    • A61B5/6801Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
    • A61B5/6813Specially adapted to be attached to a specific body part
    • A61B5/6814Head
    • A61B5/682Mouth, e.g., oral cavity; tongue; Lips; Teeth
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/72Signal processing specially adapted for physiological signals or for diagnostic purposes
    • A61B5/7235Details of waveform analysis
    • A61B5/7264Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems
    • A61B5/7267Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems involving training the classification device
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/72Signal processing specially adapted for physiological signals or for diagnostic purposes
    • A61B5/7271Specific aspects of physiological measurement analysis
    • A61B5/7275Determining trends in physiological measurement data; Predicting development of a medical condition based on physiological measurements, e.g. determining a risk factor
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/74Details of notification to user or communication with user or patient ; user input means
    • A61B5/742Details of notification to user or communication with user or patient ; user input means using visual displays
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61KPREPARATIONS FOR MEDICAL, DENTAL OR TOILETRY PURPOSES
    • A61K9/00Medicinal preparations characterised by special physical form
    • A61K9/0012Galenical forms characterised by the site of application
    • A61K9/0053Mouth and digestive tract, i.e. intraoral and peroral administration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • G06V10/443Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by matching or filtering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/60Extraction of image or video features relating to illumination properties, e.g. using a reflectance or lighting model
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/69Microscopic objects, e.g. biological cells or cellular parts
    • G06V20/698Matching; Classification
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H30/00ICT specially adapted for the handling or processing of medical images
    • G16H30/40ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H40/00ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices
    • G16H40/60ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices
    • G16H40/67ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices for remote operation
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/20ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/50ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for simulation or modelling of medical disorders
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B2560/00Constructional details of operational features of apparatus; Accessories for medical measuring apparatus
    • A61B2560/04Constructional details of apparatus
    • A61B2560/0431Portable apparatus, e.g. comprising a handle or case
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B2576/00Medical imaging apparatus involving image processing or analysis
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/08Detecting, measuring or recording devices for evaluating the respiratory organs
    • A61B5/0823Detecting or evaluating cough events
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/41Detecting, measuring or recording for evaluating the immune or lymphatic systems
    • A61B5/414Evaluating particular organs or parts of the immune or lymphatic systems
    • A61B5/418Evaluating particular organs or parts of the immune or lymphatic systems lymph vessels, ducts or nodes
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/44Detecting, measuring or recording for evaluating the integumentary system, e.g. skin, hair or nails
    • A61B5/441Skin evaluation, e.g. for skin disorder diagnosis
    • A61B5/445Evaluating skin irritation or skin trauma, e.g. rash, eczema, wound, bed sore
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61KPREPARATIONS FOR MEDICAL, DENTAL OR TOILETRY PURPOSES
    • A61K39/00Medicinal preparations containing antigens or antibodies
    • A61K39/02Bacterial antigens
    • A61K39/09Lactobacillales, e.g. aerococcus, enterococcus, lactobacillus, lactococcus, streptococcus
    • A61K39/092Streptococcus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10068Endoscopic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/70ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for mining of medical data, e.g. analysing previous cases of other patients
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A90/00Technologies having an indirect contribution to adaptation to climate change
    • Y02A90/10Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation

Definitions

  • the disclosure relates to image processing and particularly to using images captured of a subject’s throat to evaluate for the presence of disease without culturing.
  • a detection system determines disease state predictions, relating to a potential disease and/or medical condition of a subject (also referred to herein as a“patient”), using a chained model.
  • the chained model includes an image model and a classifier, according to some embodiments.
  • the chained model accesses a set of subject images, the subject images capturing a part of a subject’s body, and a set of clinical factors from the subject.
  • the clinical factors are collected by a device or a medical practitioner substantially contemporaneously with the capture of the subject images.
  • the clinical factors may include relevant information for diagnosing the subject with the respective potential disease and/or medical condition.
  • the subject images are inputted into an image model to generate disease metrics for disease prediction for the subject.
  • the image model may be trained using a set of training images and a set of training labels associated with a first set of training subjects, according to some embodiments.
  • the training labels may include a label indicating a presence of a pathogen in the subject associated with the respective training image.
  • the disease metrics generated by the image model and the clinical factors for the subject together are inputted into a classifier to determine the disease state prediction, and the disease state prediction is returned by the chained model.
  • the classifier may be trained using a set of training labels, a set of training disease metrics, and a set of training clinical factors associated with a second set of training subjects, according to some embodiments.
  • the second set of training subjects may be different than the first set of training subjects.
  • the set of training disease metrics may be generated by inputting a set of training subject images associated with the second set of training subjects to the trained image model.
  • the detection system may provide for dry in-situ clinical prediction related to the potential disease and/or medical condition (e.g., the presence/absence of bacterial and viral pathogen infections) without the need for any pathological or laboratory tests, according to some embodiments.
  • the potential disease and/or medical condition e.g., the presence/absence of bacterial and viral pathogen infections
  • FIG. 1 shows a detection system for analyzing a combination of images of a subject’s throat and subject-associated clinical factors to determine a disease state prediction for the subject, according to one embodiment.
  • FIG. 2 is a high-level block diagram illustrating an example of a computing device used either as a client device, application server, and/or database server, according to one embodiment.
  • FIGs. 3A-3I depicts various views of an image capture device, according to one embodiment.
  • FIG. 3J is an interaction diagram for the image capture of a subject’s throat, according to one embodiment.
  • FIG. 4 illustrates a process for training of an image model within a chained model, according to one embodiment.
  • FIG. 5 illustrates a process for training of a classifier within a chained model, according to one embodiment.
  • FIG. 6 illustrates a process for generating disease state predictions using a chained model, according to one embodiment.
  • FIG. 7 illustrates example input and output vectors relevant to the chained model, according to one embodiment.
  • FIG. 8 is a flowchart of returning a disease state prediction for subject determined by a chained model, according to one embodiment
  • FIG. 1 shows a detection system for analyzing a combination of images of the inside of a subject’s throat and subject-associated clinical factors to determine a disease state prediction for the subject (also referred to herein as the“patient”), according to one embodiment.
  • a disease state prediction is related to a disease or medical condition the subject may potentially have.
  • the disease state prediction may indicate a presence or probability of the subject having a streptococcal infection.
  • the detection system analyzes images provided by an image capture device 120 and clinical factors provided by a medical professional 112 or device to determine a disease state related to a type of infection present in the subject.
  • the detection system 100 is used for detecting streptococcal infections in subjects experiencing pharyngitis. In other embodiments the detection system 100 is used for detecting other diseases and conditions.
  • the detection system 100 includes client computing devices 110, 111, an image capture device 120, an application server 130, also referred to herein as server 130, database server 140, and a chained model 150.
  • FIG. 1 illustrates only a single instance of most of the components of the detection system 100, in practice more than one of each component may be present, and additional or fewer components may be used.
  • the chained model 150 is a part of the client device 110, and functions of the chained model 150 are performed locally on the client device 110.
  • the client devices 110, 111 interact with the detection system 100 via a network 160.
  • the network 160, system 100, client devices 110, 111, and/or server 130 are a secure network handling sensitive or confidential information, for example they may be designed to provide for restricted data access, encryption of data, and otherwise may be compliant with medical information protection regulations such as HIPAA.
  • HIPAA medical information protection regulations
  • One type of user is a subject who potentially has pharyngitis or another throat related disease and makes use of the system 100 at least in part to obtain a disease state prediction provided by the server 130.
  • a set of subject throat images of the subject’s throat collected by an image capture device 120 are provided to a client device 110, which in turn reports to the application server 130, which in turn can initiate a process to determine a disease state prediction which is provided to the user through the client device 110.
  • Another type of user is a medical professional 112 who provides clinical factors collected by a device or a medical practitioner substantially contemporaneously with the capture of the set of subject throat images to a client device 111 (which may also be the same as client device 110), which in turn reports to the application server 130, which in turn can be combined with the subject throat images to initiate a process to determine a disease state prediction which is provided to the user through the client device 110.
  • the medical professional 112 may operate the image capture device 120 and client device 110.
  • the subject may instead operate the image capture device 120 and the client device 110.
  • the client device 110, 111 is a computer system.
  • the client device 110, 111 is configured to communicate (e.g., wirelessly or via a wired link) with the detection system 100 via network 160. With network 160 access, the client device 110 transmits to the detection system 100 the set of subject throat images captured by the image capture device 120, and the client device 111 transmits to the detection system 100 the clinical factors provided by the medical professional 112.
  • client devices 110, 111 connected to the detection system 100 may also exchange information with other connected client devices 110, 111.
  • the client device 110 may also perform some data and image processing on the set of subject throat images locally using the resources of client device 110 before sending the processed data through the network 160.
  • the client device 111 may also perform some data processing on the clinical factors locally using the resources of client device 111 before sending the processed data through the network 160. Images and clinical factors sent through the network 160 are received by the application server 130 where they are analyzed and processed for storage and retrieval in conjunction with database server 140.
  • the application server 130 may direct retrieval and storage request to the database system 130 as required by the client devices 110, 111.
  • the client devices 110 may communicate with the image capture device 120 using a network adapter and either a wired or wireless communication protocol, an example of which is the Bluetooth Low Energy (BTLE) protocol.
  • BTLE is a short-ranged, low-powered, protocol standard that transmits data wirelessly over radio links in short range wireless networks.
  • other types of wireless connections are used (e.g., infrared, cellular, 4G, 5G, 802.11).
  • client devices 110 and image capture devices 120 are described above as being separate physical devices (such as a computing device and an image sensor, respectively), in an embodiment, the image capture device 120 may include aspects of the client device 110.
  • an image capture device may include an audiovisual interface including a display or other lighting elements as well as speakers for presenting audible information.
  • the image capture device 120 itself may present the contents of information obtained from server 130, such as the disease state prediction determined by the detection system 100, provided by the server 130 directly, in place of or in addition to presenting them through the client devices 110.
  • the client device 110 may be a smartphone, and part of the image capture device 120 may be a smartphone attachment.
  • a built-in camera of the smart phone combined with optical elements of the smartphone attachment provide the functionality of the image capture device 120.
  • one client device may act as both the client device 110 and the client device 111.
  • the application server 130 is a computer or network of computers. Although a simplified example is illustrated in FIG. 2, typically the application server will be a server class system that uses powerful processors, large memory, and faster network components compared to a typical computing system used, for example, as a client device 110.
  • the server typically has large secondary storage, for example, using a RAID (redundant array of independent disks) array and/or by establishing a relationship with an independent content delivery network (CDN) contracted to store, exchange and transmit data.
  • the computing system includes an operating system, for example, a UNIX operating system, LINUX operating system, or a WINDOWS operating system.
  • the operating system manages the hardware and software resources of the application server 130 and also provides various services, for example, process management, input/output of data, management of peripheral devices, and so on.
  • the operating system provides various functions for managing files stored on a device, for example, creating a new file, moving or copying files, transferring files to a remote system, and so on.
  • the application server 130 includes a software architecture for supporting access to and use of detection system 100 by many different client devices 110, 111 through network 160, and thus at a high level can be generally characterized as a cloud-based system.
  • the application server 130 generally provides a platform for subjects and medical professionals 112 to report data recorded by the client devices 110, 111 associated with the subject’s pharyngitis, collaborate on treatment plans, browse and obtain information relating to their condition, and make use of a variety of other functions.
  • the application server 130 is designed to handle a wide variety of data.
  • the application server 130 includes logical routines that perform a variety of functions including checking the validity of the incoming data, parsing and formatting the data if necessary, passing the processed data to a database server 140 for storage, and confirming that the database server 140 has been updated.
  • the application server 130 stores and manages data at least in part on a subject by subject basis. Towards this end, the application server 130 creates a subject profde for each user.
  • the subject profde is a set of data that characterizes a subject 113 of the detection system 100.
  • the subject profde may include identify information about the subject such as age, gender, a subject’s relevant medical history, and a list of non-subject users authorized to access the subject profde.
  • the profde may further specify a device identifier, such as a unique media access control (MAC) address identifying the one or more client devices 110,
  • MAC media access control
  • image capture devices 120 authorized to submit data (such as a set of subject throat images) for the subject.
  • the application server 130 also creates profiles for health care providers 112.
  • a health care provider profde may include identifying information about the health care provider 112, such as the office location, qualifications and certifications, and so on.
  • the health care provider profde also includes information about their subject population.
  • the provider profde may include access to all of the profiles of that provider’s subjects, as well as derived data from those profiles such as aggregate demographic information. This data may be further subdivided according to any type of data stored in the subject profiles, such as by geographic area (e.g., neighborhood, city) over by time period (e.g., weekly, monthly, yearly).
  • the application server 130 receives client factors and subject throat images from the client devices 110, 111 triggering a variety of routines on the application server 130.
  • the chained model 150 executes routines to access subject throat images as well as clinical factors, analyze the images and data, and output the results of its analysis to subjects or medical professionals 112.
  • the database server 140 stores subject and healthcare provider related data such as profiles, medication events, subject medical history (e.g., electronic medical records).
  • Subject and provider data is encrypted for security and is at least password protected and otherwise secured to meet all Health Insurance Portability and Accountability Act (HIPAA) requirements. Any analyses that incorporate data from multiple subjects and are provided to users is de-identified so that personally identifying information is removed to protect subject privacy.
  • HIPAA Health Insurance Portability and Accountability Act
  • the database server 140 is illustrated in FIG. 1 as being an entity separate from the application server 130 the database server 140 may alternatively be a hardware component that is part of another server such as server 130, such that the database server 140 is implemented as one or more persistent storage devices, with the software application layer for interfacing with the stored data in the database is a part of that other server 130.
  • the database server 140 stores data according to defined database schemas. Typically, data storage schemas across different data sources vary significantly even when storing the same type of data including cloud application event logs and log metrics, due to
  • the database server 140 may also store different types of data such as structured data, unstructured data, or semi- structured data. Data in the database server 140 may be associated with users, groups of users, and/or entities.
  • the database server 140 provides support for database queries in a query language (e.g., SQL for relational databases, JSON NoSQL databases, etc.) for specifying instructions to manage database objects represented by the database server 140, read information from the database server 140, or write to the database server 140.
  • a query language e.g., SQL for relational databases, JSON NoSQL databases, etc.
  • the network 160 represents the various wired and wireless communication pathways between the client devices 110, 111, the image capture device 120, the application server 130, and the database server 140.
  • Network 160 uses standard Internet communications technologies and/or protocols.
  • the network 160 can include links using technologies such as Ethernet, IEEE 802.11, integrated services digital network (ISDN), asynchronous transfer mode (ATM), etc.
  • the networking protocols used on the network 160 can include the transmission control protocol/Intemet protocol (TCP/IP), the hypertext transport protocol (HTTP), the simple mail transfer protocol (SMTP), the file transfer protocol (FTP), etc.
  • TCP/IP transmission control protocol/Intemet protocol
  • HTTP hypertext transport protocol
  • SMTP simple mail transfer protocol
  • FTP file transfer protocol
  • the data exchanged over the network 160 can be represented using technologies and/or formats including the hypertext markup language (HTML), the extensible markup language (XML), etc.
  • all or some links can be encrypted using conventional encryption technologies such as the secure sockets layer (SSL), Secure HTTP (HTTPS) and/or virtual private networks (VPNs).
  • SSL secure sockets layer
  • HTTPS Secure HTTP
  • VPNs virtual private networks
  • the entities can use custom and/or dedicated data communications technologies instead of, or in addition to, the ones described above.
  • FIG. 2 is a high-level block diagram illustrating physical components of an example computer 200 that may be used as part of a client device 110, 111, application server 130, and/or database server 140 from FIG. 1, according to one embodiment. Illustrated is a chipset 210 coupled to at least one processor 205. Coupled to the chipset 210 is volatile memory 215, a network adapter 220, an input/output (I/O) device(s) 225, a storage device 230 representing a non-volatile memory, and a display 235. In one embodiment, the functionality of the chipset 210 is provided by a memory controller 211 and an I/O controller 212. In another embodiment, the memory 215 is coupled directly to the processor 205 instead of the chipset 210. In some embodiments, memory 215 includes high-speed random access memory (RAM), such as DRAM, SRAM, DDR RAM or other random access solid state memory devices.
  • RAM high-speed random access memory
  • the storage device 230 is any non-transitory computer-readable storage medium, such as a hard drive, compact disk read-only memory (CD-ROM), DVD, or a solid-state memory device.
  • the memory 215 holds instructions and data used by the processor 205.
  • the I/O device 225 may be a touch input surface (capacitive or otherwise), a mouse, track ball, or other type of pointing device, a keyboard, or another form of input device.
  • the display 235 displays images and other information from for the computer 200.
  • the network adapter 220 couples the computer 200 to the network 160.
  • a computer 200 can have different and/or other components than those shown in FIG. 2.
  • the computer 200 can lack certain illustrated components.
  • a computer 200 acting as server 140 may lack a dedicated I/O device 225, and/or display 218.
  • the storage device 230 can be local and/or remote from the computer 200 (such as embodied within a storage area network (SAN)), and, in one embodiment, the storage device 230 is not a CD-ROM device or a DVD device.
  • SAN storage area network
  • client devices 110, 111 will vary in size, power requirements, and performance from those used in the application server 130 and the database server 140.
  • client devices 110, 111 which will often be home computers, tablet computers, laptop computers, or smart phones, will include relatively small storage capacities and processing power, but will include input devices and displays. These components are suitable for user input of data and receipt, display, and interaction with notifications provided by the application server 130.
  • the application server 130 may include many physically separate, locally networked computers each having a significant amount of processing power for carrying out the analyses introduced above.
  • the processing power of the application server 130 provided by a service such as Amazon Web ServicesTM or Microsoft AzureTM.
  • the database server 140 may include many, physically separate computers each having a significant amount of persistent storage capacity for storing the data associated with the application server.
  • the computer 200 is adapted to execute computer program modules for providing functionality described herein.
  • a module can be implemented in hardware, firmware, and/or software.
  • program modules are stored on the storage device 230, loaded into the memory 215, and executed by the processor 205.
  • FIGs. 3A-3C depict three views of an exemplary image capture device 120, according to one embodiment.
  • the embodiment depicted is configured for use in a human oral cavity (mouth and if desired upper throat).
  • the image capture device is configured to capture images of other parts of the body or other objects.
  • the image capture device 120 is configured to capture images of a subject’s skin.
  • the scanning and detection device can be any desired shape suitable for a given target site, for example a catheter or endoscope or other configuration (e.g., colposcope, laparascope, etc.) shaped to be inserted into or otherwise introduced into or aimed toward the body of a subject.
  • the image capture device 120 comprises a proximal end 4 and a distal end 6, with the distal end 6 configured to introduce into or aim towards an in vivo biological target site suspected of having an infection.
  • Image capture device 120 comprises housing 8 having an excitation light emitter 10 at the distal end 6, the excitation light emitter 10 configured to emit excitation light selected to elicit fluorescent light from the suspected infection at the target site; if desired, multiple excitation light emitters can be provided, each for a different wavelength/wavelength band of excitation light.
  • the image capture device 120 may further comprise a light sensor as well as a heat sensor 14 (refer, e.g., to FIG. 3D and 3F).
  • the light sensor is configured to detect at least fluorescent light emanating from the target site
  • heat sensor 14 is configured to at least detect and identify heat levels above ambient body temperature emanating from the infection at the target site.
  • the detection system further comprises operably connected computer-implemented programming configured to accept fluorescent light data associated with the fluorescent light and thermal data associated with the heat levels above ambient body temperature and interpret the data to determine a probability whether the target site contains an infection.
  • computer-implemented programming can be contained within housing 8 or can be located externally.
  • Image capture device 120 also contains three buttons for user interaction.
  • the first control button 30 controls the illumination LED (white light emitter).
  • the second button 32 initiates an image/scan acquisition procedure such as a fluorescent image/sensing procedure.
  • the third control button 34 initiates a temperature acquisition procedure.
  • Other or fewer buttons can also be provided as desired.
  • image capture device 120 can comprise an illumination light emitter 16 and an imaging system 26 comprising a camera 18.
  • One or more filters configured to transmit only desirable wavelengths/indicators of light or heat can also be provided, such as first emanating light filter 20, emanating heat filter 22, and second emanating light filter 24.
  • Image capture device 120 further contains a display screen 36, which can display spectrographic results, images of the target site, diagnostic results, false-color representations of the data received from the target site, and the like.
  • the display can also convey other information if desired, such as date, time, subject name, etc.
  • an easily removable separable distal element 38 sized and configured to removably attach to the distal end of the housing.
  • the separable distal element 38 can comprise light-blocking sides 40 and if desired a forward-facing window 42, as shown in FIG. 3E, configured to transmit at least the excitation light, the fluorescent light and the heat levels without substantial alteration.
  • the separable distal element 38 can also comprise recesses 48, 50 to accommodate expected physical structures at a target site, to avoid a side wall from impacting an image/increase scanning/imaging field of view, etc.
  • the distal end 6 of the housing 8 and the separable distal element 38 can be cooperatively configured such that the separable distal element 38 can be snapped on and off the distal end 6 of the housing 8.
  • the distal end 6 of the housing 8 and the separable distal element 38 can comprise cooperative projections 52 and detents 54 configured such that the separable distal element 38 can be snapped on and off the distal end 6 of the housing 8 by cooperatively engaging and releasing such elements.
  • Image capture device 120 can further comprise a plug-port 44 and a battery bay 46.
  • the housing 8 is configured to be held in a single hand of a user, and is configured to fit within a human oral cavity and to scan at least a rear surface of such oral cavity and/or a throat behind such oral cavity.
  • FIG. 3G and 3H show further information about the light emitters, light sensors and heat sensors.
  • all are located at the distal end 6 of the housing 8 (not shown) and are all forward-facing and aimed to substantially cover a same area of the target site, as demonstrated by the overlapping fields of view in the figures.
  • excitation light emitters include red LED 56, green LED 58, and blue LED 60.
  • FIG. 31 shows a further embodiment concerning light emitters, light sensors and heat sensors.
  • the array includes two white light emitting LEDs 62, and two blue LEDs 60, as well as a camera 18 and a radiant heat sensor 14.
  • FIG. 3J shows an interaction diagram of the image capture process for providing the set of subject throat images to the detection system 100, according to one embodiment.
  • the illumination emitter 16 provides light input to the subject’s throat, and the camera 18 simultaneously records a white image of the throat.
  • the white image of the throat may be formed by collecting light reflected from the throat of the subject.
  • excitation images also referred to herein as an“blue images”
  • the excitation emitter 10 provides light input to the subject’s throat at a specific excitation wavelength, and the camera 18 simultaneously records a blue image of the throat, according to one embodiment.
  • the blue image may be formed by collecting light emitted from the throat of the subject as a result of auto-fluorescence, in addition to light reflected by the throat of the subject.
  • the light from auto-fluorescence is a different wavelength than the excitation wavelength.
  • the excitation emitter 10 provides light input to another part of the subject’s body, and the camera 18 simultaneously records an excitation image of the part of the subject’s body.
  • the excitation emitter 10 provides blue light input, but in other embodiments, the excitation emitter may provide light input at wavelengths corresponding to other colors.
  • the subject throat images include images other than the white images and the blue images.
  • the captured subject throat images may include images captured in multiple wavelengths and multiple lighting conditions.
  • the captured subject images may include images other than the white images and the blue images.
  • fluorescent hosts in the bacteria for example a porphyrin, cause the bacteria to auto-fluoresce in response to the light input from the excitation emitter 10.
  • the camera 18 will capture this auto-fluorescence as part of the blue image.
  • viruses or other factors are targeted instead of bacteria.
  • the excitation emitter 10 causes the targeted virus or other factor to fluoresce in response to the light input from the excitation emitter 10.
  • the white image and blue image are included in the set of subject throat images provided to the chained model 150 for use in determining a disease state prediction of the subject.
  • more than one blue image or white image may be included in the set of subject throat images.
  • images other than the blue image or white image may be included in the set of subject throat images, for example images with illumination conditions from the image capture device 120.
  • the subject throat images provided to the chained model 150 may include images captured in other colors or other wavelengths of light, according to some embodiments.
  • the detection system 100 may be used to detect diseases and conditions related to skin legions present on a subject.
  • the image capture device 120 captures white images and excitation images of the skin legions.
  • the image capture device 120 only captures white images.
  • the image capture device 120 only captures excitation images of the skin legions.
  • the captured images are provided to the chained model 150 to determine a disease state prediction related to the skin legion.
  • Clinical factors for the subject are collected substantially contemporaneously with the capture of subject throat images by the image capture device 120.
  • the clinical factors for a subject are collected by the medical professional 112 and submitted to the chained model 150 using the client device 111.
  • the clinical factors are provided by a subject without the aid of or without interacting with the medical professional 112.
  • the subject may report the clinical factors through an application on a client device 110, such as a smartphone.
  • one or more of the clinical factors are not collected contemporaneously to the capture of images by the image capture device 120. For example, if age is a clinical factor for predicting a presence of a disease, the age of the subject may be recorded at a different time than the image capture.
  • the colored light spectra emitted by porphyrins resulting from Streptococci bacteria metabolism in the oral cavity is fdtered by specific wavelength. It is then captured by the image capture device’s camera 18 as white light and blue light digital images. These images are then curated, centered and cropped by an image pre-processing algorithm that assess the quality and suitability of these images for use in the image model.
  • Pre- Processing techniques that may be performed on the set of subject throat images may include: uniform aspect ratio, rescaling, normalization, segmentation, cropping, object detection, dimensionality deduction/increment, brightness adjustment, data augmentation techniques to increase the data size like: Image Shifting, flipping, zoom in/out, rotation etc., determining quality of the image to exclude bad images from being a part of training dataset, image pixel correction, and performing a FV image florescence brightness algorithm.
  • the chained model 150 includes an image model 400 and a classifier 500.
  • the training of the image model 400 and classifier 500 will be discussed below.
  • FIG. 4 illustrates a process for training of an image model 400 within a chained model 150, according to one embodiment.
  • the image model 400 is trained on a first set of training throat images associated with a first set of training subjects and a corresponding first set of training labels.
  • the training images are of sore throats captured under fluorescent light, white light, and ambient light.
  • the fluorescent light may contain blue light at a wavelength for fluorescing porphyrins associated with streptococcal bacteria.
  • Each training subject has one of several pre-determined labels.
  • the pre-determined labels distinguishes the subject as having A) a bacterial pathogen, B) a viral pathogen, or C) an absence of a pathogen.
  • the label may be a categorical label (e.g., A, B, C), or it may be a numerical label (e.g., -1, 0, 1).
  • the first set of training throat images and the associated labels are provided by a training database 415.
  • the first set of training throat images may be captured by the image capture device 120.
  • the labels for the first set of training subjects is provided on the basis that disease states of the first set of training subjects are previously known, for example as determined by traditional cell culturing and evaluation by one or more medical professionals evaluating the training set of subjects.
  • the image model 400 is trained by determining image parameter coefficients 430, each associated with a corresponding image parameter, (not shown). Collectively, the image parameter coefficients 430 are determined so as to best represent the relationship between the first set of training subject throat images input into a function of the image model 400 and their associated labels.
  • the image model 400 is a supervised machine learning technique.
  • the image model 400 is a convolutional neural network model.
  • the convolutional neural network is trained using transfer learning with fine tuning.
  • the image model 400 is specifically a VGG neural network, a ResNet neural network, or an Inception V4 neural network.
  • machine learning models and training methods may be used, examples of which include but are not limited to: stochastic gradient descent, transfer learning algorithms, learning rate annealing, cyclic learning rates, differential learning rates, regularization techniques such as batch normalization, ensembling neural networks, etc.
  • the image model 400 may be used for prediction, as discussed in FIG. 5 and FIG. 6 by accessing the image parameter coefficients 430 and the function specified by the model, and inputting input values for the image parameters to generate a prediction of pathogen presence.
  • the prediction generated for a subject by the image model 400 may include one or more of: a probability of a presence of a bacterial pathogen, a probability of a presence of a viral pathogen, and a probability of an absence of a pathogen.
  • the prediction may be output in the form of a vector including one or more of the above numerical values.
  • the prediction may also output a separate numerical confidence in the prediction.
  • the prediction may include one or more of: a probability of a presence of exudate, a probability of a presence of petechiae, a probability of a presence of swollen tonsils, and a probability of a presence of a swollen uvula.
  • the image model 400 is training with training images and corresponding training labels indicating the presence or absence of these conditions.
  • the prediction may be output in the form of a vector including one or more of the above numerical values, and the prediction may also output a separate numerical confidence in the prediction.
  • the prediction may include one or more of: a presence of plaque, a presence of oral mucosa, a presence of cancer, gastroesophageal reflux disease (GERD) detection, and a presence of bacterial pathogens (e.g, e.coli, salmonella, and other pathogens).
  • a presence of plaque e.g, a plaque, a presence of oral mucosa, a presence of cancer, gastroesophageal reflux disease (GERD) detection, and a presence of bacterial pathogens (e.g, e.coli, salmonella, and other pathogens).
  • GSD gastroesophageal reflux disease
  • the image model 400 is any machine learning model that directly or indirectly generates a prediction of a presence of a disease factor such as a pathogen, a presence of or property of a tumor, or a degree of swelling of a body part.
  • the image model 400 is a machine learning model that performs feature detection on images (e.g., white images, blue images, or images in other wavelengths or lighting conditions) of a subject’s throat, as well as color classification.
  • the feature detection and color classification may be used to determine targeted feature metrics including, but not limited to: presence/size/shape/location of the oral cavity, oral cavity symmetry, presence/size/shape/location tonsils, tonsil redness, tonsil swelling, a soft or hard palate, presence of red spots on the palate, streaks of pus, white patches, and dry mouth.
  • Each of the feature metrics may correspond to an identified feature in an image.
  • a feature metric may indicate a presence of an identified feature or a property of an identified feature.
  • feature detection on the white images may complement the feature detection performed on the blue images.
  • the feature detection and the color classification determines targeted infection metrics including, but not limited to:
  • one or more of the targeted infection metrics generated by the image model for the blue images indicate characteristics of auto fluorescence in one or more regions of a subject’s throat captured in the blue image, in response to illumination from an excitation light source (e.g., blue light from the image capture device).
  • Each of the infection metrics may correspond to an infection in the subject.
  • an infection metric may indicate a presence of a certain infection (e.g., a viral infection or a bacterial infection) in the subject or a property of an infection.
  • the determined feature metrics and infection metrics may then be provided independently of or alongside the prediction of a presence of a pathogen according to the methods described above to the classifier as inputs for generating a patient’s disease state prediction.
  • the determined feature metrics and infection metrics may be provided without the prediction of a presence of a pathogen to the classifier as inputs for generating a patient’s disease state.
  • feature detection and color classification is performed using k-means clustering, however other unsupervised machine learning techniques may also be used.
  • FIG. 5 illustrates a process for training of a classifier within a chained model 150, according to one embodiment.
  • the classifier 500 is trained using a set of training predictions of pathogen presence generated by the pre-trained image model 400 based on a second set of training throat images, training clinical factors associated with a second set of training subjects, and a corresponding second training set of labels.
  • the classifier 500 is trained using feature metrics and infection metrics generated by the pre-trained image model 400 based on the second set of training throat images, in addition to or independently of the training data described above.
  • each subject from the second set of training subjects has a corresponding pre-determined label distinguishing the subject as having a bacterial pathogen, a viral pathogen, or an absence of a pathogen.
  • these labels may be determined by traditional cell culturing and evaluation by one or more medical professionals evaluating the training set of subjects.
  • the labels may alternatively be determined by other methods.
  • the second set of training subject throat images and the associated labels are provided by the training database 415.
  • the training clinical factors are provided by a training clinical database 515.
  • the training clinical database 515 contains clinical factors for each of the second set of training subjects collected by a medical professional or device 120. These images are generally collected substantially simultaneously with the capture of the corresponding training subject throat images for that subject.
  • the classifier 500 is trained by determining classifier parameter coefficients 530, each associated with each classifier parameter (not shown).
  • the coefficients are trained so as to collectively best represent the relationship between the input values (predictions of pathogen presence and clinical factors) of the second set of training subjects and a function of the classifier to the second set of training labels.
  • the classifier 500 is trained using a supervised machine learning technique.
  • the classifier 500 is a neural network model, trained using trained using stochastic gradient descent.
  • other types of classifiers and training methods may be used, examples of which include but are not limited to linear, logistic, and other forms of regression (e.g., elastic net, multinomial regression), decision trees (e.g., random forest, gradient boosting), support vector machines, classifiers (e.g. Naive Bayes classifier), fuzzy matching.
  • the classifier may perform classical statistical analysis methods that include, but are not limited to: correlations, hypothesis tests, and analysis of variance (ANOVA).
  • the classifier model 500 may be used for prediction, as discussed in FIG. 6 by accessing the classifier parameter coefficients 530 and the function specified by the classifier, and inputting input values for the parameters to generate a prediction of disease state.
  • the disease state prediction of the subject generated by the classifier 500 may include one or more of: a probability of bacterial infection, a probability of viral infection, and a probability of no infection. Additionally or alternatively, the disease state prediction may include probabilities indicating the presence of anatomical morphologies or symptoms.
  • the probabilities indicating the presence of anatomical morphologies or systems include one or more of: a probability of a presence of exudate, a probability of a presence of petechiae, a probability of a presence of swollen tonsils, and a probability of a presence of a swollen uvula.
  • the disease state predictions may indicate probabilities of other morphologies or symptoms.
  • the set of clinical factors of the subject used by the classifier 500 in the chained model 150 may include, but are not limited to: an age, a presence or absence of swollen lymph nodes, a subject temperature, a presence or absence of a fever, a presence or absence of coughing symptoms, a presence or absence of a runny nose, a presence or absence of a headache, a presence or absence of body aches, a presence or absence of vomiting, a presence or absence of diarrhea, a presence or absence of fatigue, a presence or absence of chills, a duration of pharyngitis, and a set of symptoms correlated with the Centor procedure.
  • an age a presence or absence of swollen lymph nodes, a subject temperature, a presence or absence of a fever, a presence or absence of coughing symptoms, a presence or absence of a runny nose, a presence or absence of a headache, a presence or absence of body aches, a presence or absence of vomiting, a presence or absence of diarrhea
  • FIG. 6 illustrates a process for generating disease state predictions using a chained model 150, according to one embodiment.
  • the chained model 150 receives as input a set of subject throat images from a subject and a set of clinical factors collected by a medical professional 112 substantially contemporaneously to the capture of the set of subject throat images.
  • the images of the subject are of sore throats captured under fluorescent light, white light, and ambient light.
  • the chained model 150 generates disease state prediction for the subject.
  • the input set of subject throat images may include only white images captured with white lighting conditions or ambient lighting conditions, or only blue images captured using illumination from an excitation light source for fluorescence.
  • the input set of subject throat images may include subject throat images captured under other lighting conditions.
  • the input set of subject throat images may include multiple images capturing multiple wavelengths of light.
  • the generation of the disease state prediction for the subject is a two-step process.
  • a first step includes inputting the set of subject throat images to the image model 400.
  • the image model 400 accesses the image parameter coefficients 430 and generates a pathogen presence prediction for the subject.
  • the pathogen presence prediction is provided together with the set of clinical factors as inputs to the classifier 500.
  • the classifier 500 accesses the classifier parameter coefficients 530 and together with clinical factors and pathogen presence prediction generates a disease state prediction for the subject.
  • the disease state prediction may then be provided to the client device 110 and displayed to a medical professional or the subject.
  • the chained model 150 can provide a disease state prediction solely using the pathogen presence prediction without accessing the clinical factors for the subject.
  • the set of subject throat images is sufficient for determining the disease state prediction, and only the output of the image model 400 is used.
  • the image model 400 is trained using blue images, white images, images captured in a different wavelength of light or different lighting conditions, or some combination thereof, but when generating disease state predictions, the chained model 150 may have input subject throat images that are captured in a different wavelength of light or different lighting conditions than the training images.
  • the image model 400 may be trained using a combination of white images and blue images, but only white images may be used as inputs for the chained model 150 when generating disease state predictions for a subject.
  • FIG. 7 illustrates example input and output vectors relevant to the chained model 150, according to one embodiment.
  • the input vectors include the set of subject throat images and the clinical factors.
  • the resulting output vector of the chained model is a disease state prediction, which includes probabilities for various types of infections in the subject.
  • the clinical factors include age, a presence or absence of swollen lymph nodes, a body temperature, and a presence or absence of a cough.
  • the set of subject throat images includes white images and blue images of the subject’s throat captured with the image capture device 120.
  • the disease state prediction includes a probability of a bacterial infection, a probability of a viral infection, and a probability of no infection, as determined by the chained model 150, based on the input vectors.
  • the input vectors and resulting output vectors of the chained model 150 may be different than what is shown in FIG. 7. For example, if a disease or condition other than pharyngitis is targeted, the input vectors and resulting output vectors may be relevant to the targeted disease.
  • FIG. 8 is a flowchart 800 of returning a disease state prediction for subject determined by a chained model 150, according to one embodiment.
  • the disease state prediction indicates a probability of a subject having a disease or medical condition, according to some embodiments.
  • the chained model 150 accesses 810 a set of subject images associated with the subject.
  • the subject images depict a part of the subject’s body.
  • the subject image may be an image of the subject’s throat.
  • the chained model 150 accesses 820 a set of clinical factors for the subject. The clinical factors are recorded substantially
  • the subject images are inputted 830 into the image model 400 to generate disease metrics.
  • the generated disease metrics and the clinical factors are then inputted into the classifier 500 to determine the disease state prediction for the subject, and the determined disease state prediction is returned 850.
  • the detection system described herein provides for dry in-situ clinical prediction of the presence/absence of bacterial and viral pathogen infections without the need for any pathological or laboratory tests.
  • the detection system may provide subjects with a home diagnostic tool for strep throat. This may effectively reduce the financial burden of treating pharyngitis for both healthcare providers and subjects, as well as reduce the time necessary to determine an accurate diagnosis. Additionally, the detection system may provide accurate predictions for diseases and conditions other than strep throat.
  • any reference to“one embodiment” or“an embodiment” means that a particular element, feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment.
  • the appearances of the phrase“in one embodiment” in various places in the specification are not necessarily all referring to the same embodiment.
  • the terms“comprises,”“comprising,”“includes,”“including,”“has,” “having” or any other variation thereof are intended to cover a non-exclusive inclusion.
  • a process, method, article, or apparatus that comprises a list of elements is not necessarily limited to only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus.
  • “or” refers to an inclusive or and not to an exclusive or. For example, a condition A or B is satisfied by any one of the following: A is true (or present) and B is false (or not present), A is false (or not present) and B is true (or present), and both A and B are true (or present).

Landscapes

  • Health & Medical Sciences (AREA)
  • Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Medical Informatics (AREA)
  • General Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • Public Health (AREA)
  • Biomedical Technology (AREA)
  • Pathology (AREA)
  • Veterinary Medicine (AREA)
  • Animal Behavior & Ethology (AREA)
  • Molecular Biology (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Biophysics (AREA)
  • Surgery (AREA)
  • Artificial Intelligence (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • General Physics & Mathematics (AREA)
  • Epidemiology (AREA)
  • Evolutionary Computation (AREA)
  • Radiology & Medical Imaging (AREA)
  • Data Mining & Analysis (AREA)
  • Primary Health Care (AREA)
  • Physiology (AREA)
  • Multimedia (AREA)
  • Psychiatry (AREA)
  • Databases & Information Systems (AREA)
  • Signal Processing (AREA)
  • Mathematical Physics (AREA)
  • Fuzzy Systems (AREA)
  • Software Systems (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Nutrition Science (AREA)
  • Chemical & Material Sciences (AREA)
  • Medicinal Chemistry (AREA)

Abstract

A method for determining a disease state prediction, relating to a potential disease or medical condition of a subject, includes accessing a set of subject images, the subject images capturing a part of a subject's body, anlaawad accessing a set of clinical factors from the subject. The clinical factors are collected by a device or a medical practitioner substantially contemporaneously with the capture of the subject images. The subject images are inputted into an image model to generate disease metrics for disease prediction for the subject. The disease metrics generated by the image model and the clinical factors are inputted into a classifier to determine the disease state prediction, and the disease state prediction is returned.

Description

IMAGE PROCESSING OF STREPTOCOCCAL INFECTION IN
PHARYNGITIS SUBJECTS
CROSS-REFERENCE TO RELATED PATENT APPLICATION
[0001] This patent application claims priority to U.S. Provisional Patent Application No. 62/743,245, filed on October 9, 2018, and to U.S. Provisional Patent Application No.
62/855,875, filed on May 31, 2019, which are incorporated herein in their entirety.
BACKGROUND
FIELD OF ART
[0002] The disclosure relates to image processing and particularly to using images captured of a subject’s throat to evaluate for the presence of disease without culturing.
DESCRIPTION OF THE RELATED ART
[0003] Sore throats account for around 1% of all primary care visits in North America. All too often the cause of the complaint is viral not bacterial. However, a rapid and accurate diagnosis is challenging for medical practitioners and it is all too easy to adopt a bias towards prescribing antibiotics, thereby contributing to their over-use. This is problematic because it often results in unnecessary spending on medication, as well as contributing to a general increase in antibiotic resistance. Clinical tests, which traditionally involve culturing in vitro for presence of bacteria can be slow, taking up to 72 hours to offer results, are difficult to administer properly, especially in children, and as such can be inaccurate. There is a clear need for a more reliable, fast and automatic detection process.
SUMMARY
[0004] A detection system determines disease state predictions, relating to a potential disease and/or medical condition of a subject (also referred to herein as a“patient”), using a chained model. The chained model includes an image model and a classifier, according to some embodiments. The chained model accesses a set of subject images, the subject images capturing a part of a subject’s body, and a set of clinical factors from the subject. The clinical factors are collected by a device or a medical practitioner substantially contemporaneously with the capture of the subject images. According to some embodiments, the clinical factors may include relevant information for diagnosing the subject with the respective potential disease and/or medical condition. The subject images are inputted into an image model to generate disease metrics for disease prediction for the subject. The image model may be trained using a set of training images and a set of training labels associated with a first set of training subjects, according to some embodiments. For example, the training labels may include a label indicating a presence of a pathogen in the subject associated with the respective training image.
[0005] The disease metrics generated by the image model and the clinical factors for the subject together are inputted into a classifier to determine the disease state prediction, and the disease state prediction is returned by the chained model. The classifier may be trained using a set of training labels, a set of training disease metrics, and a set of training clinical factors associated with a second set of training subjects, according to some embodiments. In some embodiments, the second set of training subjects, may be different than the first set of training subjects. The set of training disease metrics may be generated by inputting a set of training subject images associated with the second set of training subjects to the trained image model. Using the chained model, the detection system may provide for dry in-situ clinical prediction related to the potential disease and/or medical condition (e.g., the presence/absence of bacterial and viral pathogen infections) without the need for any pathological or laboratory tests, according to some embodiments.
BRIEF DESCRIPTION OF DRAWINGS
[0006] FIG. 1 shows a detection system for analyzing a combination of images of a subject’s throat and subject-associated clinical factors to determine a disease state prediction for the subject, according to one embodiment.
[0007] FIG. 2 is a high-level block diagram illustrating an example of a computing device used either as a client device, application server, and/or database server, according to one embodiment.
[0008] FIGs. 3A-3I depicts various views of an image capture device, according to one embodiment.
[0009] FIG. 3J is an interaction diagram for the image capture of a subject’s throat, according to one embodiment.
[0010] FIG. 4 illustrates a process for training of an image model within a chained model, according to one embodiment.
[0011] FIG. 5 illustrates a process for training of a classifier within a chained model, according to one embodiment.
[0012] FIG. 6 illustrates a process for generating disease state predictions using a chained model, according to one embodiment.
[0013] FIG. 7 illustrates example input and output vectors relevant to the chained model, according to one embodiment.
[0014] FIG. 8 is a flowchart of returning a disease state prediction for subject determined by a chained model, according to one embodiment
[0015] The figures depict various embodiments of the presented invention for purposes of illustration only. One skilled in the art will readily recognize from the following discussion that alternative embodiments of the structures and methods illustrated herein may be employed without departing from the principles described herein.
DETAILED DESCRIPTION
I. System Architecture
[0016] FIG. 1 shows a detection system for analyzing a combination of images of the inside of a subject’s throat and subject-associated clinical factors to determine a disease state prediction for the subject (also referred to herein as the“patient”), according to one embodiment. A disease state prediction is related to a disease or medical condition the subject may potentially have. For example, the disease state prediction may indicate a presence or probability of the subject having a streptococcal infection. The detection system analyzes images provided by an image capture device 120 and clinical factors provided by a medical professional 112 or device to determine a disease state related to a type of infection present in the subject. In some embodiments, the detection system 100 is used for detecting streptococcal infections in subjects experiencing pharyngitis. In other embodiments the detection system 100 is used for detecting other diseases and conditions.
[0017] The detection system 100 includes client computing devices 110, 111, an image capture device 120, an application server 130, also referred to herein as server 130, database server 140, and a chained model 150. Although FIG. 1 illustrates only a single instance of most of the components of the detection system 100, in practice more than one of each component may be present, and additional or fewer components may be used. In one embodiment, the chained model 150 is a part of the client device 110, and functions of the chained model 150 are performed locally on the client device 110.
I. A. CLIENT DEVICE AND APPLICATION
[0018] The client devices 110, 111 interact with the detection system 100 via a network 160. In one embodiment, the network 160, system 100, client devices 110, 111, and/or server 130 are a secure network handling sensitive or confidential information, for example they may be designed to provide for restricted data access, encryption of data, and otherwise may be compliant with medical information protection regulations such as HIPAA. For purposes of explanation and clarity it is useful to identify at least two different types of users. One type of user is a subject who potentially has pharyngitis or another throat related disease and makes use of the system 100 at least in part to obtain a disease state prediction provided by the server 130. As will be explained below, a set of subject throat images of the subject’s throat collected by an image capture device 120 are provided to a client device 110, which in turn reports to the application server 130, which in turn can initiate a process to determine a disease state prediction which is provided to the user through the client device 110.
[0019] Another type of user is a medical professional 112 who provides clinical factors collected by a device or a medical practitioner substantially contemporaneously with the capture of the set of subject throat images to a client device 111 (which may also be the same as client device 110), which in turn reports to the application server 130, which in turn can be combined with the subject throat images to initiate a process to determine a disease state prediction which is provided to the user through the client device 110. The medical professional 112 may operate the image capture device 120 and client device 110.
Alternatively, the subject may instead operate the image capture device 120 and the client device 110.
[0020] The client device 110, 111 is a computer system. An example physical
implementation is described more completely below with respect to FIG. 2. The client device 110, 111 is configured to communicate (e.g., wirelessly or via a wired link) with the detection system 100 via network 160. With network 160 access, the client device 110 transmits to the detection system 100 the set of subject throat images captured by the image capture device 120, and the client device 111 transmits to the detection system 100 the clinical factors provided by the medical professional 112.
[0021] In addition to communicating with the application server 130, client devices 110, 111 connected to the detection system 100 may also exchange information with other connected client devices 110, 111.
[0022] The client device 110 may also perform some data and image processing on the set of subject throat images locally using the resources of client device 110 before sending the processed data through the network 160. The client device 111 may also perform some data processing on the clinical factors locally using the resources of client device 111 before sending the processed data through the network 160. Images and clinical factors sent through the network 160 are received by the application server 130 where they are analyzed and processed for storage and retrieval in conjunction with database server 140. The application server 130 may direct retrieval and storage request to the database system 130 as required by the client devices 110, 111.
[0023] The client devices 110 may communicate with the image capture device 120 using a network adapter and either a wired or wireless communication protocol, an example of which is the Bluetooth Low Energy (BTLE) protocol. BTLE is a short-ranged, low-powered, protocol standard that transmits data wirelessly over radio links in short range wireless networks. In other implementations, other types of wireless connections are used (e.g., infrared, cellular, 4G, 5G, 802.11).
[0024] Although client devices 110 and image capture devices 120 are described above as being separate physical devices (such as a computing device and an image sensor, respectively), in an embodiment, the image capture device 120 may include aspects of the client device 110. For example, an image capture device may include an audiovisual interface including a display or other lighting elements as well as speakers for presenting audible information. In such an implementation the image capture device 120 itself may present the contents of information obtained from server 130, such as the disease state prediction determined by the detection system 100, provided by the server 130 directly, in place of or in addition to presenting them through the client devices 110.
[0025] In one embodiment, the client device 110 may be a smartphone, and part of the image capture device 120 may be a smartphone attachment. In such an implementation, a built-in camera of the smart phone combined with optical elements of the smartphone attachment provide the functionality of the image capture device 120.
[0026] In one embodiment, one client device may act as both the client device 110 and the client device 111.
I.B. APPLICATION SERVER
[0027] The application server 130 is a computer or network of computers. Although a simplified example is illustrated in FIG. 2, typically the application server will be a server class system that uses powerful processors, large memory, and faster network components compared to a typical computing system used, for example, as a client device 110. The server typically has large secondary storage, for example, using a RAID (redundant array of independent disks) array and/or by establishing a relationship with an independent content delivery network (CDN) contracted to store, exchange and transmit data. Additionally, the computing system includes an operating system, for example, a UNIX operating system, LINUX operating system, or a WINDOWS operating system. The operating system manages the hardware and software resources of the application server 130 and also provides various services, for example, process management, input/output of data, management of peripheral devices, and so on. The operating system provides various functions for managing files stored on a device, for example, creating a new file, moving or copying files, transferring files to a remote system, and so on.
[0028] The application server 130 includes a software architecture for supporting access to and use of detection system 100 by many different client devices 110, 111 through network 160, and thus at a high level can be generally characterized as a cloud-based system. The application server 130 generally provides a platform for subjects and medical professionals 112 to report data recorded by the client devices 110, 111 associated with the subject’s pharyngitis, collaborate on treatment plans, browse and obtain information relating to their condition, and make use of a variety of other functions.
[0029] Generally, the application server 130 is designed to handle a wide variety of data. The application server 130 includes logical routines that perform a variety of functions including checking the validity of the incoming data, parsing and formatting the data if necessary, passing the processed data to a database server 140 for storage, and confirming that the database server 140 has been updated.
[0030] The application server 130 stores and manages data at least in part on a subject by subject basis. Towards this end, the application server 130 creates a subject profde for each user. The subject profde is a set of data that characterizes a subject 113 of the detection system 100. The subject profde may include identify information about the subject such as age, gender, a subject’s relevant medical history, and a list of non-subject users authorized to access the subject profde. The profde may further specify a device identifier, such as a unique media access control (MAC) address identifying the one or more client devices 110,
111 or image capture devices 120 authorized to submit data (such as a set of subject throat images) for the subject.
[0031] The application server 130 also creates profiles for health care providers 112. A health care provider profde may include identifying information about the health care provider 112, such as the office location, qualifications and certifications, and so on. The health care provider profde also includes information about their subject population. The provider profde may include access to all of the profiles of that provider’s subjects, as well as derived data from those profiles such as aggregate demographic information. This data may be further subdivided according to any type of data stored in the subject profiles, such as by geographic area (e.g., neighborhood, city) over by time period (e.g., weekly, monthly, yearly).
[0032] The application server 130 receives client factors and subject throat images from the client devices 110, 111 triggering a variety of routines on the application server 130. In the example implementations described below, the chained model 150 executes routines to access subject throat images as well as clinical factors, analyze the images and data, and output the results of its analysis to subjects or medical professionals 112.
I.C. DATABASE SERVER
[0033] The database server 140 stores subject and healthcare provider related data such as profiles, medication events, subject medical history (e.g., electronic medical records). Subject and provider data is encrypted for security and is at least password protected and otherwise secured to meet all Health Insurance Portability and Accountability Act (HIPAA) requirements. Any analyses that incorporate data from multiple subjects and are provided to users is de-identified so that personally identifying information is removed to protect subject privacy.
[0034] Although the database server 140 is illustrated in FIG. 1 as being an entity separate from the application server 130 the database server 140 may alternatively be a hardware component that is part of another server such as server 130, such that the database server 140 is implemented as one or more persistent storage devices, with the software application layer for interfacing with the stored data in the database is a part of that other server 130.
[0035] The database server 140 stores data according to defined database schemas. Typically, data storage schemas across different data sources vary significantly even when storing the same type of data including cloud application event logs and log metrics, due to
implementation differences in the underlying database structure. The database server 140 may also store different types of data such as structured data, unstructured data, or semi- structured data. Data in the database server 140 may be associated with users, groups of users, and/or entities. The database server 140 provides support for database queries in a query language (e.g., SQL for relational databases, JSON NoSQL databases, etc.) for specifying instructions to manage database objects represented by the database server 140, read information from the database server 140, or write to the database server 140.
[0036] With respect to the descriptions of Figs. 4-6, the contents of the databases described with respect to those figures may be stored in databases physically proximate to the application server 130 and separate from database server 140 as illustrated.
I D. NETWORK
[0037] The network 160 represents the various wired and wireless communication pathways between the client devices 110, 111, the image capture device 120, the application server 130, and the database server 140. Network 160 uses standard Internet communications technologies and/or protocols. Thus, the network 160 can include links using technologies such as Ethernet, IEEE 802.11, integrated services digital network (ISDN), asynchronous transfer mode (ATM), etc. Similarly, the networking protocols used on the network 160 can include the transmission control protocol/Intemet protocol (TCP/IP), the hypertext transport protocol (HTTP), the simple mail transfer protocol (SMTP), the file transfer protocol (FTP), etc. The data exchanged over the network 160 can be represented using technologies and/or formats including the hypertext markup language (HTML), the extensible markup language (XML), etc. In addition, all or some links can be encrypted using conventional encryption technologies such as the secure sockets layer (SSL), Secure HTTP (HTTPS) and/or virtual private networks (VPNs). In another embodiment, the entities can use custom and/or dedicated data communications technologies instead of, or in addition to, the ones described above.
II. EXAMPLE COMPUTING DEVICES
[0038] FIG. 2 is a high-level block diagram illustrating physical components of an example computer 200 that may be used as part of a client device 110, 111, application server 130, and/or database server 140 from FIG. 1, according to one embodiment. Illustrated is a chipset 210 coupled to at least one processor 205. Coupled to the chipset 210 is volatile memory 215, a network adapter 220, an input/output (I/O) device(s) 225, a storage device 230 representing a non-volatile memory, and a display 235. In one embodiment, the functionality of the chipset 210 is provided by a memory controller 211 and an I/O controller 212. In another embodiment, the memory 215 is coupled directly to the processor 205 instead of the chipset 210. In some embodiments, memory 215 includes high-speed random access memory (RAM), such as DRAM, SRAM, DDR RAM or other random access solid state memory devices.
[0039] The storage device 230 is any non-transitory computer-readable storage medium, such as a hard drive, compact disk read-only memory (CD-ROM), DVD, or a solid-state memory device. The memory 215 holds instructions and data used by the processor 205. The I/O device 225 may be a touch input surface (capacitive or otherwise), a mouse, track ball, or other type of pointing device, a keyboard, or another form of input device. The display 235 displays images and other information from for the computer 200. The network adapter 220 couples the computer 200 to the network 160.
[0040] As is known in the art, a computer 200 can have different and/or other components than those shown in FIG. 2. In addition, the computer 200 can lack certain illustrated components. In one embodiment, a computer 200 acting as server 140 may lack a dedicated I/O device 225, and/or display 218. Moreover, the storage device 230 can be local and/or remote from the computer 200 (such as embodied within a storage area network (SAN)), and, in one embodiment, the storage device 230 is not a CD-ROM device or a DVD device.
[0041] Generally, the exact physical components used in a client device 110, 111 will vary in size, power requirements, and performance from those used in the application server 130 and the database server 140. For example, client devices 110, 111 which will often be home computers, tablet computers, laptop computers, or smart phones, will include relatively small storage capacities and processing power, but will include input devices and displays. These components are suitable for user input of data and receipt, display, and interaction with notifications provided by the application server 130. In contrast, the application server 130 may include many physically separate, locally networked computers each having a significant amount of processing power for carrying out the analyses introduced above. In one embodiment, the processing power of the application server 130 provided by a service such as Amazon Web Services™ or Microsoft Azure™. Also in contrast, the database server 140 may include many, physically separate computers each having a significant amount of persistent storage capacity for storing the data associated with the application server.
[0042] As is known in the art, the computer 200 is adapted to execute computer program modules for providing functionality described herein. A module can be implemented in hardware, firmware, and/or software. In one embodiment, program modules are stored on the storage device 230, loaded into the memory 215, and executed by the processor 205.
III. IMAGE CAPTURE AND CLINICAL FACTORS
III. A. IMAGE CAPTURE DEVICE
[0043] FIGs. 3A-3C depict three views of an exemplary image capture device 120, according to one embodiment. The embodiment depicted is configured for use in a human oral cavity (mouth and if desired upper throat). In other embodiments, the image capture device is configured to capture images of other parts of the body or other objects. For example, in one embodiment, the image capture device 120 is configured to capture images of a subject’s skin. The scanning and detection device can be any desired shape suitable for a given target site, for example a catheter or endoscope or other configuration (e.g., colposcope, laparascope, etc.) shaped to be inserted into or otherwise introduced into or aimed toward the body of a subject.
[0044] In one embodiment, the image capture device 120 comprises a proximal end 4 and a distal end 6, with the distal end 6 configured to introduce into or aim towards an in vivo biological target site suspected of having an infection. Image capture device 120 comprises housing 8 having an excitation light emitter 10 at the distal end 6, the excitation light emitter 10 configured to emit excitation light selected to elicit fluorescent light from the suspected infection at the target site; if desired, multiple excitation light emitters can be provided, each for a different wavelength/wavelength band of excitation light. The image capture device 120 may further comprise a light sensor as well as a heat sensor 14 (refer, e.g., to FIG. 3D and 3F). The light sensor is configured to detect at least fluorescent light emanating from the target site, and heat sensor 14 is configured to at least detect and identify heat levels above ambient body temperature emanating from the infection at the target site.
[0045] In one embodiment, the detection system further comprises operably connected computer-implemented programming configured to accept fluorescent light data associated with the fluorescent light and thermal data associated with the heat levels above ambient body temperature and interpret the data to determine a probability whether the target site contains an infection. Such computer-implemented programming can be contained within housing 8 or can be located externally.
[0046] Image capture device 120 also contains three buttons for user interaction. The first control button 30 controls the illumination LED (white light emitter). The second button 32 initiates an image/scan acquisition procedure such as a fluorescent image/sensing procedure. The third control button 34 initiates a temperature acquisition procedure. Other or fewer buttons can also be provided as desired.
[0047] As shown in FIGs. 3D and FIG. 3F, image capture device 120 can comprise an illumination light emitter 16 and an imaging system 26 comprising a camera 18. One or more filters configured to transmit only desirable wavelengths/indicators of light or heat can also be provided, such as first emanating light filter 20, emanating heat filter 22, and second emanating light filter 24.
[0048] Image capture device 120 further contains a display screen 36, which can display spectrographic results, images of the target site, diagnostic results, false-color representations of the data received from the target site, and the like. The display can also convey other information if desired, such as date, time, subject name, etc. Also shown is an easily removable separable distal element 38 sized and configured to removably attach to the distal end of the housing. The separable distal element 38 can comprise light-blocking sides 40 and if desired a forward-facing window 42, as shown in FIG. 3E, configured to transmit at least the excitation light, the fluorescent light and the heat levels without substantial alteration.
The separable distal element 38 can also comprise recesses 48, 50 to accommodate expected physical structures at a target site, to avoid a side wall from impacting an image/increase scanning/imaging field of view, etc. The distal end 6 of the housing 8 and the separable distal element 38 can be cooperatively configured such that the separable distal element 38 can be snapped on and off the distal end 6 of the housing 8. For example, the distal end 6 of the housing 8 and the separable distal element 38 can comprise cooperative projections 52 and detents 54 configured such that the separable distal element 38 can be snapped on and off the distal end 6 of the housing 8 by cooperatively engaging and releasing such elements. Image capture device 120 can further comprise a plug-port 44 and a battery bay 46.
[0049] In the embodiment depicted in FIGs. 3A-3F, the housing 8 is configured to be held in a single hand of a user, and is configured to fit within a human oral cavity and to scan at least a rear surface of such oral cavity and/or a throat behind such oral cavity.
[0050] FIG. 3G and 3H show further information about the light emitters, light sensors and heat sensors. In this embodiment, all are located at the distal end 6 of the housing 8 (not shown) and are all forward-facing and aimed to substantially cover a same area of the target site, as demonstrated by the overlapping fields of view in the figures. Also in this embodiment, excitation light emitters include red LED 56, green LED 58, and blue LED 60.
[0051] FIG. 31 shows a further embodiment concerning light emitters, light sensors and heat sensors. In this embodiment, the array includes two white light emitting LEDs 62, and two blue LEDs 60, as well as a camera 18 and a radiant heat sensor 14.
[0052] FIG. 3J shows an interaction diagram of the image capture process for providing the set of subject throat images to the detection system 100, according to one embodiment. For white images, the illumination emitter 16 provides light input to the subject’s throat, and the camera 18 simultaneously records a white image of the throat. The white image of the throat may be formed by collecting light reflected from the throat of the subject. For excitation images, also referred to herein as an“blue images,” the excitation emitter 10 provides light input to the subject’s throat at a specific excitation wavelength, and the camera 18 simultaneously records a blue image of the throat, according to one embodiment. The blue image may be formed by collecting light emitted from the throat of the subject as a result of auto-fluorescence, in addition to light reflected by the throat of the subject. In some cases, the light from auto-fluorescence is a different wavelength than the excitation wavelength. In one embodiment, the excitation emitter 10 provides light input to another part of the subject’s body, and the camera 18 simultaneously records an excitation image of the part of the subject’s body. In some embodiments, the excitation emitter 10 provides blue light input, but in other embodiments, the excitation emitter may provide light input at wavelengths corresponding to other colors. In some embodiments, the subject throat images include images other than the white images and the blue images. For example, the captured subject throat images may include images captured in multiple wavelengths and multiple lighting conditions. In embodiments where the detection system 100 targets other diseases and/or medical conditions, the captured subject images may include images other than the white images and the blue images.
[0053] In the case where targeted bacterial pathogens are present in the subject’s throat, fluorescent hosts in the bacteria, for example a porphyrin, cause the bacteria to auto-fluoresce in response to the light input from the excitation emitter 10. The camera 18 will capture this auto-fluorescence as part of the blue image. In some embodiments, viruses or other factors are targeted instead of bacteria. In further embodiments, the excitation emitter 10 causes the targeted virus or other factor to fluoresce in response to the light input from the excitation emitter 10.
[0054] The white image and blue image are included in the set of subject throat images provided to the chained model 150 for use in determining a disease state prediction of the subject. In one embodiment, more than one blue image or white image may be included in the set of subject throat images. In another embodiment, images other than the blue image or white image may be included in the set of subject throat images, for example images with illumination conditions from the image capture device 120. For example, the subject throat images provided to the chained model 150 may include images captured in other colors or other wavelengths of light, according to some embodiments.
[0055] In another example, the detection system 100 may be used to detect diseases and conditions related to skin legions present on a subject. In such a case, the image capture device 120 captures white images and excitation images of the skin legions. In some embodiments, the image capture device 120 only captures white images. In other embodiments, the image capture device 120 only captures excitation images of the skin legions. The captured images are provided to the chained model 150 to determine a disease state prediction related to the skin legion. III.B. COLLECTION OF CLINICAL FACTORS
[0056] Clinical factors for the subject are collected substantially contemporaneously with the capture of subject throat images by the image capture device 120. In one embodiment, the clinical factors for a subject are collected by the medical professional 112 and submitted to the chained model 150 using the client device 111. In another embodiment, the clinical factors are provided by a subject without the aid of or without interacting with the medical professional 112. For example, the subject may report the clinical factors through an application on a client device 110, such as a smartphone. In alternate embodiments, one or more of the clinical factors are not collected contemporaneously to the capture of images by the image capture device 120. For example, if age is a clinical factor for predicting a presence of a disease, the age of the subject may be recorded at a different time than the image capture.
III. C. IMAGE CAPTURE AND PREPROCESSING
[0057] For the case of diagnosing a pharyngitis case related to the presence of Streptococci bacteria, the colored light spectra emitted by porphyrins resulting from Streptococci bacteria metabolism in the oral cavity is fdtered by specific wavelength. It is then captured by the image capture device’s camera 18 as white light and blue light digital images. These images are then curated, centered and cropped by an image pre-processing algorithm that assess the quality and suitability of these images for use in the image model.
[0058] Good image pre-processing leads to a robust AI model for accurate predictions. Pre- Processing techniques that may be performed on the set of subject throat images may include: uniform aspect ratio, rescaling, normalization, segmentation, cropping, object detection, dimensionality deduction/increment, brightness adjustment, data augmentation techniques to increase the data size like: Image Shifting, flipping, zoom in/out, rotation etc., determining quality of the image to exclude bad images from being a part of training dataset, image pixel correction, and performing a FV image florescence brightness algorithm.
IV. CHAINED MODEL
IV. A. IMAGE MODEL TRAINING
[0059] In one embodiment, the chained model 150 includes an image model 400 and a classifier 500. The training of the image model 400 and classifier 500 will be discussed below. [0060] FIG. 4 illustrates a process for training of an image model 400 within a chained model 150, according to one embodiment. The image model 400 is trained on a first set of training throat images associated with a first set of training subjects and a corresponding first set of training labels. In one embodiment, the training images are of sore throats captured under fluorescent light, white light, and ambient light. The fluorescent light may contain blue light at a wavelength for fluorescing porphyrins associated with streptococcal bacteria.
[0061] Each training subject has one of several pre-determined labels. In one embodiment, the pre-determined labels distinguishes the subject as having A) a bacterial pathogen, B) a viral pathogen, or C) an absence of a pathogen. The label may be a categorical label (e.g., A, B, C), or it may be a numerical label (e.g., -1, 0, 1). The first set of training throat images and the associated labels are provided by a training database 415. The first set of training throat images may be captured by the image capture device 120. The labels for the first set of training subjects is provided on the basis that disease states of the first set of training subjects are previously known, for example as determined by traditional cell culturing and evaluation by one or more medical professionals evaluating the training set of subjects.
[0062] The image model 400 is trained by determining image parameter coefficients 430, each associated with a corresponding image parameter, (not shown). Collectively, the image parameter coefficients 430 are determined so as to best represent the relationship between the first set of training subject throat images input into a function of the image model 400 and their associated labels. Generally, the image model 400 is a supervised machine learning technique. In one embodiment, the image model 400 is a convolutional neural network model. In a further embodiment, the convolutional neural network is trained using transfer learning with fine tuning. In other embodiments, the image model 400 is specifically a VGG neural network, a ResNet neural network, or an Inception V4 neural network. In other embodiments, other types of machine learning models and training methods may be used, examples of which include but are not limited to: stochastic gradient descent, transfer learning algorithms, learning rate annealing, cyclic learning rates, differential learning rates, regularization techniques such as batch normalization, ensembling neural networks, etc.
[0063] Once the parameter coefficients are known, the image model 400 may be used for prediction, as discussed in FIG. 5 and FIG. 6 by accessing the image parameter coefficients 430 and the function specified by the model, and inputting input values for the image parameters to generate a prediction of pathogen presence. The prediction generated for a subject by the image model 400 may include one or more of: a probability of a presence of a bacterial pathogen, a probability of a presence of a viral pathogen, and a probability of an absence of a pathogen. The prediction may be output in the form of a vector including one or more of the above numerical values. The prediction may also output a separate numerical confidence in the prediction.
[0064] In one embodiment, the prediction may include one or more of: a probability of a presence of exudate, a probability of a presence of petechiae, a probability of a presence of swollen tonsils, and a probability of a presence of a swollen uvula. In this embodiment, the image model 400 is training with training images and corresponding training labels indicating the presence or absence of these conditions. Again, the prediction may be output in the form of a vector including one or more of the above numerical values, and the prediction may also output a separate numerical confidence in the prediction. In some embodiments, where the detection system 100 is used for diseases and/or medical conditions other than pharyngitis, the prediction may include one or more of: a presence of plaque, a presence of oral mucosa, a presence of cancer, gastroesophageal reflux disease (GERD) detection, and a presence of bacterial pathogens (e.g, e.coli, salmonella, and other pathogens).
[0065] In other embodiments, the image model 400 is any machine learning model that directly or indirectly generates a prediction of a presence of a disease factor such as a pathogen, a presence of or property of a tumor, or a degree of swelling of a body part. In one embodiment, the image model 400 is a machine learning model that performs feature detection on images (e.g., white images, blue images, or images in other wavelengths or lighting conditions) of a subject’s throat, as well as color classification. According to some embodiments the feature detection and color classification may be used to determine targeted feature metrics including, but not limited to: presence/size/shape/location of the oral cavity, oral cavity symmetry, presence/size/shape/location tonsils, tonsil redness, tonsil swelling, a soft or hard palate, presence of red spots on the palate, streaks of pus, white patches, and dry mouth. Each of the feature metrics may correspond to an identified feature in an image. For example, a feature metric may indicate a presence of an identified feature or a property of an identified feature. In some embodiments, feature detection on the white images may complement the feature detection performed on the blue images. [0066] In some embodiments, for the blue images, the feature detection and the color classification determines targeted infection metrics including, but not limited to:
presence/size/shape/location of an infected area, an intensity, and a pattern identification. In some embodiments the feature detection and the color classification is used for images other than the blue images. In some embodiments, one or more of the targeted infection metrics generated by the image model for the blue images indicate characteristics of auto fluorescence in one or more regions of a subject’s throat captured in the blue image, in response to illumination from an excitation light source (e.g., blue light from the image capture device). Each of the infection metrics may correspond to an infection in the subject. For example, an infection metric may indicate a presence of a certain infection (e.g., a viral infection or a bacterial infection) in the subject or a property of an infection. In one embodiment the determined feature metrics and infection metrics may then be provided independently of or alongside the prediction of a presence of a pathogen according to the methods described above to the classifier as inputs for generating a patient’s disease state prediction. In other embodiments, the determined feature metrics and infection metrics may be provided without the prediction of a presence of a pathogen to the classifier as inputs for generating a patient’s disease state. In one embodiment, feature detection and color classification is performed using k-means clustering, however other unsupervised machine learning techniques may also be used.
IV. B. CLASSIFIER TRAINING
[0067] FIG. 5 illustrates a process for training of a classifier within a chained model 150, according to one embodiment. The classifier 500 is trained using a set of training predictions of pathogen presence generated by the pre-trained image model 400 based on a second set of training throat images, training clinical factors associated with a second set of training subjects, and a corresponding second training set of labels. In one embodiment, the classifier 500 is trained using feature metrics and infection metrics generated by the pre-trained image model 400 based on the second set of training throat images, in addition to or independently of the training data described above. As with the first set of training labels, each subject from the second set of training subjects has a corresponding pre-determined label distinguishing the subject as having a bacterial pathogen, a viral pathogen, or an absence of a pathogen. Again, these labels may be determined by traditional cell culturing and evaluation by one or more medical professionals evaluating the training set of subjects. The labels may alternatively be determined by other methods.
[0068] Again, the second set of training subject throat images and the associated labels are provided by the training database 415. The training clinical factors are provided by a training clinical database 515. The training clinical database 515 contains clinical factors for each of the second set of training subjects collected by a medical professional or device 120. These images are generally collected substantially simultaneously with the capture of the corresponding training subject throat images for that subject.
[0069] The classifier 500 is trained by determining classifier parameter coefficients 530, each associated with each classifier parameter (not shown). The coefficients are trained so as to collectively best represent the relationship between the input values (predictions of pathogen presence and clinical factors) of the second set of training subjects and a function of the classifier to the second set of training labels.
[0070] Generally, the classifier 500 is trained using a supervised machine learning technique. In one embodiment, the classifier 500 is a neural network model, trained using trained using stochastic gradient descent. In other embodiments, other types of classifiers and training methods may be used, examples of which include but are not limited to linear, logistic, and other forms of regression (e.g., elastic net, multinomial regression), decision trees (e.g., random forest, gradient boosting), support vector machines, classifiers (e.g. Naive Bayes classifier), fuzzy matching. In other embodiments, the classifier may perform classical statistical analysis methods that include, but are not limited to: correlations, hypothesis tests, and analysis of variance (ANOVA).
[0071] Once the parameter coefficients are known, the classifier model 500 may be used for prediction, as discussed in FIG. 6 by accessing the classifier parameter coefficients 530 and the function specified by the classifier, and inputting input values for the parameters to generate a prediction of disease state. The disease state prediction of the subject generated by the classifier 500 may include one or more of: a probability of bacterial infection, a probability of viral infection, and a probability of no infection. Additionally or alternatively, the disease state prediction may include probabilities indicating the presence of anatomical morphologies or symptoms. In one embodiment, the probabilities indicating the presence of anatomical morphologies or systems include one or more of: a probability of a presence of exudate, a probability of a presence of petechiae, a probability of a presence of swollen tonsils, and a probability of a presence of a swollen uvula. In cases where diseases or conditions other than pharyngitis are targeted, the disease state predictions may indicate probabilities of other morphologies or symptoms.
IV. C. CLINICAL FACTORS
[0072] In one embodiment, the set of clinical factors of the subject used by the classifier 500 in the chained model 150 may include, but are not limited to: an age, a presence or absence of swollen lymph nodes, a subject temperature, a presence or absence of a fever, a presence or absence of coughing symptoms, a presence or absence of a runny nose, a presence or absence of a headache, a presence or absence of body aches, a presence or absence of vomiting, a presence or absence of diarrhea, a presence or absence of fatigue, a presence or absence of chills, a duration of pharyngitis, and a set of symptoms correlated with the Centor procedure. V. MODEL INFERENCE
[0073] FIG. 6 illustrates a process for generating disease state predictions using a chained model 150, according to one embodiment. The chained model 150 receives as input a set of subject throat images from a subject and a set of clinical factors collected by a medical professional 112 substantially contemporaneously to the capture of the set of subject throat images. In one embodiment, the images of the subject are of sore throats captured under fluorescent light, white light, and ambient light. The chained model 150 generates disease state prediction for the subject. In some embodiments, the input set of subject throat images may include only white images captured with white lighting conditions or ambient lighting conditions, or only blue images captured using illumination from an excitation light source for fluorescence. In other embodiments, the input set of subject throat images may include subject throat images captured under other lighting conditions. For example, the input set of subject throat images may include multiple images capturing multiple wavelengths of light.
[0074] The generation of the disease state prediction for the subject is a two-step process. A first step includes inputting the set of subject throat images to the image model 400. The image model 400 accesses the image parameter coefficients 430 and generates a pathogen presence prediction for the subject. The pathogen presence prediction is provided together with the set of clinical factors as inputs to the classifier 500. The classifier 500 accesses the classifier parameter coefficients 530 and together with clinical factors and pathogen presence prediction generates a disease state prediction for the subject. The disease state prediction may then be provided to the client device 110 and displayed to a medical professional or the subject.
[0075] In one embodiment, the chained model 150 can provide a disease state prediction solely using the pathogen presence prediction without accessing the clinical factors for the subject. In this case, the set of subject throat images is sufficient for determining the disease state prediction, and only the output of the image model 400 is used.
[0076] In some embodiments, the image model 400 is trained using blue images, white images, images captured in a different wavelength of light or different lighting conditions, or some combination thereof, but when generating disease state predictions, the chained model 150 may have input subject throat images that are captured in a different wavelength of light or different lighting conditions than the training images. For example, the image model 400 may be trained using a combination of white images and blue images, but only white images may be used as inputs for the chained model 150 when generating disease state predictions for a subject.
[0077] FIG. 7 illustrates example input and output vectors relevant to the chained model 150, according to one embodiment. The input vectors include the set of subject throat images and the clinical factors. The resulting output vector of the chained model is a disease state prediction, which includes probabilities for various types of infections in the subject.
[0078] In the example, shown in FIG. 7, the clinical factors include age, a presence or absence of swollen lymph nodes, a body temperature, and a presence or absence of a cough. The set of subject throat images includes white images and blue images of the subject’s throat captured with the image capture device 120. The disease state prediction includes a probability of a bacterial infection, a probability of a viral infection, and a probability of no infection, as determined by the chained model 150, based on the input vectors. The input vectors and resulting output vectors of the chained model 150 may be different than what is shown in FIG. 7. For example, if a disease or condition other than pharyngitis is targeted, the input vectors and resulting output vectors may be relevant to the targeted disease.
[0079] FIG. 8 is a flowchart 800 of returning a disease state prediction for subject determined by a chained model 150, according to one embodiment. The disease state prediction indicates a probability of a subject having a disease or medical condition, according to some embodiments. The chained model 150 accesses 810 a set of subject images associated with the subject. The subject images depict a part of the subject’s body. For example, the subject image may be an image of the subject’s throat. The chained model 150 accesses 820 a set of clinical factors for the subject. The clinical factors are recorded substantially
contemporaneously with the capture of the subject images. The subject images are inputted 830 into the image model 400 to generate disease metrics. The generated disease metrics and the clinical factors are then inputted into the classifier 500 to determine the disease state prediction for the subject, and the determined disease state prediction is returned 850.
VI. BENEFITS
[0080] The detection system described herein provides for dry in-situ clinical prediction of the presence/absence of bacterial and viral pathogen infections without the need for any pathological or laboratory tests. The detection system, according to some embodiments, may provide subjects with a home diagnostic tool for strep throat. This may effectively reduce the financial burden of treating pharyngitis for both healthcare providers and subjects, as well as reduce the time necessary to determine an accurate diagnosis. Additionally, the detection system may provide accurate predictions for diseases and conditions other than strep throat.
VII. ADDITIONAL CONSIDERATIONS
[0081] Although the discussion above includes examples focusing on pharyngitis and strep throat specifically, all systems and processes described herein are equally applicable to other conditions.
[0082] It is to be understood that the figures and descriptions of the present disclosure have been simplified to illustrate elements that are relevant for a clear understanding of the present disclosure, while eliminating, for the purpose of clarity, many other elements found in a typical system. Those of ordinary skill in the art may recognize that other elements and/or steps are desirable and/or required in implementing the present disclosure. However, because such elements and steps are well known in the art, and because they do not facilitate a better understanding of the present disclosure, a discussion of such elements and steps is not provided herein. The disclosure herein is directed to all such variations and modifications to such elements and methods known to those skilled in the art.
[0083] Some portions of above description describe the embodiments in terms of algorithms and symbolic representations of operations on information. These algorithmic descriptions and representations are commonly used by those skilled in the data processing arts to convey the substance of their work effectively to others skilled in the art. These operations, while described functionally, computationally, or logically, are understood to be implemented by computer programs or equivalent electrical circuits, microcode, or the like. Furthermore, it has also proven convenient at times, to refer to these arrangements of operations as modules, without loss of generality. The described operations and their associated modules may be embodied in software, firmware, hardware, or any combinations thereof.
[0084] As used herein any reference to“one embodiment” or“an embodiment” means that a particular element, feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment. The appearances of the phrase“in one embodiment” in various places in the specification are not necessarily all referring to the same embodiment.
[0085] As used herein, the terms“comprises,”“comprising,”“includes,”“including,”“has,” “having” or any other variation thereof, are intended to cover a non-exclusive inclusion. For example, a process, method, article, or apparatus that comprises a list of elements is not necessarily limited to only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Further, unless expressly stated to the contrary,“or” refers to an inclusive or and not to an exclusive or. For example, a condition A or B is satisfied by any one of the following: A is true (or present) and B is false (or not present), A is false (or not present) and B is true (or present), and both A and B are true (or present).
[0086] In addition, use of the“a” or“an” are employed to describe elements and components of the embodiments herein. This is done merely for convenience and to give a general sense of the invention. This description should be read to include one or at least one and the singular also includes the plural unless it is obvious that it is meant otherwise.
[0087] While particular embodiments and applications have been illustrated and described, it is to be understood that the disclosed embodiments are not limited to the precise construction and components disclosed herein. Various modifications, changes and variations, which will be apparent to those skilled in the art, may be made in the arrangement, operation and details of the method and apparatus disclosed herein without departing from the spirit and scope of the ideas described herein.

Claims

What is claimed is:
1. A method comprising:
accessing a set of subject throat images from a subject capturing an inside of the subject’s throat;
accessing a set of clinical factors from the subject, the clinical factors collected by a device or a medical practitioner substantially contemporaneously with the capture of the subject throat images;
inputting the subject throat images into an image model to generate a prediction regarding a pathogen presence prediction for the subject;
inputting the pathogen presence prediction and the clinical factors into a classifier to determine a disease state prediction; and
returning the disease state prediction.
2. The method of claim 1, wherein the image model comprises:
a set of image parameter coefficients trained using a first set of training throat images and a first set of training labels, each corresponding to a first set of training subjects, the first set of training labels comprising:
a viral label indicating a presence of a viral pathogen,
a bacterial label indicating a presence of a bacterial pathogen, and
a clear label indicating an absence of pathogens; and
a function relating one of the throat images and the image parameter coefficients to the pathogen presence prediction.
3. The method of claim 1, wherein the classifier comprises:
a set of classifier parameter coefficients trained using a set of training pathogen presence predictions, a set of training clinical factors, and a second set of training labels, each corresponding to a second set of training subjects,
the second set of training labels comprising:
a viral label indicating a presence of a viral pathogen, a bacterial label indicating a presence of a bacterial pathogen, and a clear subset indicating an absence of pathogens, the set of training pathogen presence predictions generated by inputting a second set of training throat images corresponding to the second set of subjects into the image model;
the set of training clinical factors collected by a device or a medical practitioner substantially contemporaneously with the capture of the second set of training throat images, and
a function relating the pathogen presence predictions, the clinical factors, and the
classifier parameter coefficients to the disease state prediction.
4. The method of claim 1, wherein the set of training throat images were captured using the same image capture device used to capture the set of subject throat images.
5. The method of claim 4, wherein the set of training throat images and the set of subject throat images each comprise
a plurality of throat images captured under ambient light conditions,
a plurality of throat images captured under fluorescent light, and
a plurality of throat images captured under white light illumination.
6. The method of claim 1, wherein the set of subject throat images is recorded with an image capture device comprising:
a housing;
a light emitter configured to emit excitation light at a wavelength selected to elicit auto fluorescence of a pathogen;
a light sensor configured to detect light emissions or an absence of light emissions
resulting from the auto-fluorescence of the pathogen; and
a display.
7. The method of claim 6, wherein the disease state prediction is displayed on the display of the image capture device.
8. The method of claim 1, wherein the set of subject throat images is recorded by a mobile phone device.
9. The method of claim 8, wherein the disease state prediction is displayed on the mobile phone device.
10. The method of claim 1, wherein the set of subject throat images comprises
at least one blue throat image captured using a blue light emitter; and
and at least one white throat image captured using a white light emitter.
11. The method of claim 1, wherein the subject throat images captures data regarding multiple wavelengths of light.
12. The method of claim 1, wherein at least one of the subject throat images captures infrared light image data.
13. The method of claim 1, wherein the set of subject throat images are pre-processed before being input into the image model, the pre-processing comprising at least one from the group consisting of:
uniform aspect ratio correction,
rescaling,
normalization,
object detection,
segmentation,
cropping,
dimensionality reduction,
dimensionality increment,
brightness adjustment,
image shifting,
image flipping,
zoom in or out, image rotation,
image quality filtering, and
image pixel correction.
14. The method of claim 1, wherein the image model is a convolutional neural network (CNN).
15. The method of claim 1, wherein the classifier is trained using one of: linear regression, logistic regression, multinomial regression, elastic net regression.
16. The method of claim 1, wherein the classifier is one of a random foreign classifier, a gradient boosted classifier, a support vector machine classifier, and a Naive Bayes classifier.
17. The method of claim 1, wherein the pathogen presence prediction comprises at least one of:
a probability of a presence of a viral pathogen,
a probability of a presence of a bacterial pathogen, and
a probability of an absence of a pathogen.
18. The method of claim 1, wherein the pathogen presence prediction comprises at least one of:
a probability of a presence of exudate,
a probability of a presence of petechiae,
a probability of a presence of swollen tonsils, and
a probability of a presence of a swollen uvula.
19. The method of claim 1, the disease state prediction comprises at least one of:
a probability of viral pathogen infection,
a probability of bacterial pathogen infection, and
a probability of no pathogen infection.
20. The method of claim 1, wherein the set of clinical factors comprises at least one from the group consisting of:
age,
a presence or absence of swollen lymph nodes,
subject temperature,
a presence or absence of a fever, and
a presence or absence of a cough.
21. The method of claim 1, wherein the set of clinical factors comprises at least one from the group consisting of:
age,
a presence or absence of swollen lymph nodes,
subject temperature,
a presence or absence of fever
a presence or absence of a cough,
a presence or absence of a runny nose,
a presence or absence of a headache,
a presence or absence of body aches,
a presence or absence of vomiting,
a presence or absence of diarrhea,
a presence or absence of fatigue,
a presence or absence of chills, and
a duration of pharyngitis.
22. A computer system comprising a computer processor and a memory, the memory storing computer program instructions that when executed by the computer processor cause the processor to:
access a set of subject throat images from a subject;
access a set of clinical factors from the subject, the clinical factors collected by a device or a medical practitioner substantially contemporaneously with the capture of the subject throat images; input the subject throat images into an image model to generate a prediction regarding a pathogen presence prediction for the subject;
input the pathogen presence prediction and the clinical factors into a classifier to
determine a disease state prediction; and
return the disease state prediction.
23. A non-transitory computer readable storage medium comprising computer program instructions that when executed by a computer processor cause the processor to:
access a set of subject throat images from a subject;
access a set of clinical factors from the subject, the clinical factors collected by a device or a medical practitioner substantially contemporaneously with the capture of the subject throat images;
input the subject throat images into an image model to generate a prediction regarding a pathogen presence prediction for the subject;
input the pathogen presence prediction and the clinical factors into a classifier to
determine a disease state prediction; and
return the disease state prediction.
24. A method comprising:
accessing a set of subject images, the subject images capturing a part of a subject’s body; accessing a set of clinical factors from the subject, the clinical factors collected by a device or a medical practitioner substantially contemporaneously with the capture of the subject images;
inputting the subject images into an image model to generate disease metrics for disease prediction for the subject;
inputting the disease metrics and the clinical factors into a classifier to determine a
disease state prediction, the disease state prediction relating to a disease or medical condition; and
returning the disease state prediction.
25. The method of claim 24, wherein the disease metrics comprise: feature metrics corresponding to identified features in the subject image, and
infection metrics corresponding to a presence of a bacterial or viral infection in the part of the subject’s body.
26. The method of claim 24, wherein the image model comprises:
a set of image parameter coefficients trained using a first set of training subject images and a first set of training labels, each corresponding to a first set of training subjects, the first set of training labels comprising:
a viral label indicating a presence of a viral pathogen,
a bacterial label indicating a presence of a bacterial pathogen, and
a clear label indicating an absence of pathogens; and
a function relating one of the subject images and the image parameter coefficients to the disease metrics.
27. The method of claim 24, wherein the classifier comprises:
a set of classifier parameter coefficients trained using a set of training disease metrics, a set of training clinical factors, and a second set of training labels, each corresponding to a second set of training subjects,
the second set of training labels comprising:
a viral label indicating a presence of a viral pathogen, a bacterial label indicating a presence of a bacterial pathogen, and a clear subset indicating an absence of pathogens,
the set of training disease metrics generated by inputting a second set of training subject images corresponding to the second set of subjects into the image model;
the set of training clinical factors collected by a device or a medical practitioner substantially contemporaneously with the capture of the second set of training subject images, and
a function relating the disease metrics, the clinical factors, and the classifier parameter coefficients to the disease state prediction.
PCT/CA2019/051401 2018-10-09 2019-10-01 Image processing of streptococcal infection in pharyngitis subjects WO2020073114A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
AU2019357949A AU2019357949B2 (en) 2018-10-09 2019-10-01 Image processing of streptococcal infection in pharyngitis subjects
EP19871128.5A EP3864669A4 (en) 2018-10-09 2019-10-01 Image processing of streptococcal infection in pharyngitis subjects
CA3115738A CA3115738A1 (en) 2018-10-09 2019-10-01 Image processing of streptococcal infection in pharyngitis subjects
IL282169A IL282169A (en) 2018-10-09 2021-04-08 Image processing of streptococcal infection in pharyngitis subjects

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
US201862743245P 2018-10-09 2018-10-09
US62/743,245 2018-10-09
US201962855875P 2019-05-31 2019-05-31
US62/855,875 2019-05-31
US16/589,077 US11369318B2 (en) 2019-05-31 2019-09-30 Image processing of streptococcal infection in pharyngitis subjects
US16/589,077 2019-09-30

Publications (1)

Publication Number Publication Date
WO2020073114A1 true WO2020073114A1 (en) 2020-04-16

Family

ID=70052773

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CA2019/051401 WO2020073114A1 (en) 2018-10-09 2019-10-01 Image processing of streptococcal infection in pharyngitis subjects

Country Status (6)

Country Link
US (1) US11602312B2 (en)
EP (1) EP3864669A4 (en)
AU (1) AU2019357949B2 (en)
CA (1) CA3115738A1 (en)
IL (1) IL282169A (en)
WO (1) WO2020073114A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111968740A (en) * 2020-09-03 2020-11-20 卫宁健康科技集团股份有限公司 Diagnostic label recommendation method and device, storage medium and electronic equipment
WO2021044431A1 (en) * 2019-09-08 2021-03-11 Camdoc Ltd. Methods, systems and software for improved diagnosis of a medical condition
EP4121983A4 (en) * 2020-03-19 2024-04-17 Light AI Inc. Infection detection using image data analysis

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2024059184A1 (en) * 2022-09-16 2024-03-21 The Johns Hopkins University Machine learning systems and related aspects for the detection of disease states
CN116309501B (en) * 2023-03-27 2024-02-02 北京鹰之眼智能健康科技有限公司 Sore surface type prediction method, electronic equipment and storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106295139A (en) * 2016-07-29 2017-01-04 姹ゅ钩 A kind of tongue body autodiagnosis health cloud service system based on degree of depth convolutional neural networks
US20170156597A1 (en) 2015-11-13 2017-06-08 Yes Biotechnology Inc. Devices, systems and methods relating to in situ differentiation between viral and bacterial infections
WO2017214061A1 (en) * 2016-06-07 2017-12-14 The Board Of Trustees Of The Leland Stanford Junior University Methods for diagnosis of bacterial and viral infections
US20180001184A1 (en) * 2016-05-02 2018-01-04 Bao Tran Smart device
US20180045798A1 (en) * 2011-08-25 2018-02-15 Hitachi, Ltd. Medical image imaging device for positioning an imaging slice
US20190110753A1 (en) * 2017-10-13 2019-04-18 Ai Technologies Inc. Deep learning-based diagnosis and referral of ophthalmic diseases and disorders

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CA2724973C (en) * 2008-05-20 2015-08-11 University Health Network Device and method for fluorescence-based imaging and monitoring
US9031632B2 (en) * 2009-10-22 2015-05-12 Koninklijke Philips N.V. Imaging based virus detection
ES2865298T3 (en) * 2015-05-19 2021-10-15 Tyto Care Ltd Throat Imaging Systems and Methods
WO2017055412A1 (en) * 2015-09-30 2017-04-06 Siemens Healthcare Gmbh Method and system for classification of endoscopic images using deep decision networks
US9589374B1 (en) * 2016-08-01 2017-03-07 12 Sigma Technologies Computer-aided diagnosis system for medical images using deep convolutional neural networks
CN107492099B (en) * 2017-08-28 2021-08-20 京东方科技集团股份有限公司 Medical image analysis method, medical image analysis system, and storage medium
US10558887B2 (en) * 2017-12-04 2020-02-11 Adobe Inc. Digital image search based on arbitrary image features

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180045798A1 (en) * 2011-08-25 2018-02-15 Hitachi, Ltd. Medical image imaging device for positioning an imaging slice
US20170156597A1 (en) 2015-11-13 2017-06-08 Yes Biotechnology Inc. Devices, systems and methods relating to in situ differentiation between viral and bacterial infections
US20180001184A1 (en) * 2016-05-02 2018-01-04 Bao Tran Smart device
WO2017214061A1 (en) * 2016-06-07 2017-12-14 The Board Of Trustees Of The Leland Stanford Junior University Methods for diagnosis of bacterial and viral infections
CN106295139A (en) * 2016-07-29 2017-01-04 姹ゅ钩 A kind of tongue body autodiagnosis health cloud service system based on degree of depth convolutional neural networks
US20190110753A1 (en) * 2017-10-13 2019-04-18 Ai Technologies Inc. Deep learning-based diagnosis and referral of ophthalmic diseases and disorders

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP3864669A4

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021044431A1 (en) * 2019-09-08 2021-03-11 Camdoc Ltd. Methods, systems and software for improved diagnosis of a medical condition
EP4121983A4 (en) * 2020-03-19 2024-04-17 Light AI Inc. Infection detection using image data analysis
CN111968740A (en) * 2020-09-03 2020-11-20 卫宁健康科技集团股份有限公司 Diagnostic label recommendation method and device, storage medium and electronic equipment

Also Published As

Publication number Publication date
EP3864669A1 (en) 2021-08-18
US11602312B2 (en) 2023-03-14
AU2019357949B2 (en) 2022-09-29
AU2019357949A1 (en) 2021-06-03
CA3115738A1 (en) 2020-04-16
EP3864669A4 (en) 2022-07-27
IL282169A (en) 2021-05-31
US20220273245A1 (en) 2022-09-01

Similar Documents

Publication Publication Date Title
US11369318B2 (en) Image processing of streptococcal infection in pharyngitis subjects
US11602312B2 (en) Image processing of streptococcal infection in pharyngitis subjects
US11759109B2 (en) Method for automating collection, association, and coordination of multiple medical data sources
US10505935B1 (en) Providing notifications to authorized users
US11593522B2 (en) Systems and methods for a de-identified medical and healthcare data marketplace
US10937160B1 (en) Dental images processed with artificial intelligence
US20120008838A1 (en) System and method for remote melanoma screening
Rahman et al. A multimodal, multimedia point-of-care deep learning framework for COVID-19 diagnosis
US20210295506A1 (en) Infection detection using image data analysis
US7333921B2 (en) Scalable, concurrent, distributed sensor system and method
US9529968B2 (en) System and method of integrating mobile medical data into a database centric analytical process, and clinical workflow
US20170249738A1 (en) Software tool for breast cancer screening
EP3671751A1 (en) Systems and methods for implementing a secure database for storing a patient operational longitudinal record
Makkar et al. SecureFed: federated learning empowered medical imaging technique to analyze lung abnormalities in chest X-rays
Sies et al. Dark corner artefact and diagnostic performance of a market‐approved neural network for skin cancer classification
Van Molle et al. Dermatologist versus artificial intelligence confidence in dermoscopy diagnosis: Complementary information that may affect decision‐making
AU2022200340B2 (en) Digital image screening and/or diagnosis using artificial intelligence
WO2021002847A1 (en) Method for automating collection, association, and coordination of multiple medical data sources
Sakakibara et al. Construction and validation of an image discrimination algorithm to discriminate necrosis from wounds in pressure ulcers
CN108717870A (en) A kind of disease remote diagnosis system
Koren et al. Personal wireless data in formal electronic health records: future potential of Internet of Medical Things data
Chakraborty et al. Artificial Intelligence-based Oral Cancer Screening System using Smartphones: Oral cancer screening system
Scarbrough et al. Designing a use-error robust machine learning model for quantitative analysis of diffuse reflectance spectra
US20240120046A1 (en) Patient data utilization reporting
US20240112772A1 (en) Patient data curation

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19871128

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 3115738

Country of ref document: CA

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 2019871128

Country of ref document: EP

Effective date: 20210510

ENP Entry into the national phase

Ref document number: 2019357949

Country of ref document: AU

Date of ref document: 20191001

Kind code of ref document: A