WO2019153039A1 - Systems and methods for ai-assisted echocardiography - Google Patents

Systems and methods for ai-assisted echocardiography Download PDF

Info

Publication number
WO2019153039A1
WO2019153039A1 PCT/AU2019/050086 AU2019050086W WO2019153039A1 WO 2019153039 A1 WO2019153039 A1 WO 2019153039A1 AU 2019050086 W AU2019050086 W AU 2019050086W WO 2019153039 A1 WO2019153039 A1 WO 2019153039A1
Authority
WO
WIPO (PCT)
Prior art keywords
data
measurement
dataset
records
disease
Prior art date
Application number
PCT/AU2019/050086
Other languages
French (fr)
Inventor
Edward Bordin
Razali Bin MOHAMAD
Original Assignee
Alerte Echo IQ Pty Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from AU2018900364A external-priority patent/AU2018900364A0/en
Application filed by Alerte Echo IQ Pty Ltd filed Critical Alerte Echo IQ Pty Ltd
Publication of WO2019153039A1 publication Critical patent/WO2019153039A1/en

Links

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/08Detecting organic movements or changes, e.g. tumours, cysts, swellings
    • A61B8/0883Detecting organic movements or changes, e.g. tumours, cysts, swellings for diagnosis of the heart
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/72Signal processing specially adapted for physiological signals or for diagnostic purposes
    • A61B5/7235Details of waveform analysis
    • A61B5/7264Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems
    • A61B5/7267Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems involving training the classification device
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/72Signal processing specially adapted for physiological signals or for diagnostic purposes
    • A61B5/7271Specific aspects of physiological measurement analysis
    • A61B5/7275Determining trends in physiological measurement data; Predicting development of a medical condition based on physiological measurements, e.g. determining a risk factor
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/52Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/5215Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of medical diagnostic data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/088Non-supervised learning, e.g. competitive learning
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/20ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/50ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for simulation or modelling of medical disorders
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/70ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for mining of medical data, e.g. analysing previous cases of other patients
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/48Diagnostic techniques
    • A61B8/488Diagnostic techniques involving Doppler signals
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2216/00Indexing scheme relating to additional aspects of information retrieval not explicitly covered by G06F16/00 and subgroups
    • G06F2216/03Data mining
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N7/00Computing arrangements based on specific mathematical models
    • G06N7/01Probabilistic graphical models, e.g. probabilistic networks
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H10/00ICT specially adapted for the handling or processing of patient-related medical or healthcare data
    • G16H10/60ICT specially adapted for the handling or processing of patient-related medical or healthcare data for patient-specific data, e.g. for electronic patient records

Definitions

  • the present invention relates to artificial intelligence and in particular to systems and methods for artificial intelligence integration to analysis of medical data and records.
  • the invention has been developed primarily for use in methods and systems for systems and methods for AI-assisted echocardiography and will be described hereinafter with reference to this application. However, it will be appreciated that the invention is not limited to this particular field of use.
  • Echocardiography (or simply “echo ”) is a sub-specialty of cardiology that makes use of specialised ultrasound equipment to take diagnostic images of the heart. It is particularly valuable as a first-line diagnostic tool due to its ability to non-invasively assess the internal structure and function of the heart in a cost-effective manner.
  • a comprehensive Echo procedure can measure many features of the heart totaling in approximately 150 variables but the full set is rarely measured due to time constraints.
  • NEDA National Echocardiogram Database Australia
  • This unique NEDA resource collates all echocardiographic measurement and report data contained in the echocardiographic database of participating centers. Each database is then remotely transferred into the Master NEDA Database via a“vendor-agnostic” , automated data extraction process that transfers every measurement for each echocardiogram performed into a standardized NEDA data format (according to the NEDA Data Dictionary ). Each individual contributing to NEDA is given a unique identifier along with their demographic profile (date of birth and sex) and all data recorded with their echocardiogram.
  • Each patient record in the database usually does not contain a full set of measurements that can be possibly taken during an Echo procedure due to time constraints. Measurements are targeted to a subset of measurements that may be related to a suspected diagnosis or condition of the heart.
  • NEDA being a sparse data set with each patient echo record consisting of a subset of a full echo procedure leaving“blanks” for measurements that are not measured in a full procedure.
  • Figure 1 shows a few real examples of different measurements present and missing in echo studies for a small randomly selected group of patients, being a typical example of sparse echo data where each row of the table is a record of the echo measurement available for a single patient.
  • Images are acquired 101 by a sonographer using a special ultrasound workstation.
  • ⁇ Specific physical features are measured 103 from the acquired echo images by the sonographer, for example, the diameter of the left ventricle is commonly measured. (Note: many of the measurements will typically be taken during the image acquisition process with the patient present, while others may be measured from the images acquired during the procedure once the patient has left). In this workflow, the sonographer has complete and sole control over whether or not sufficient images of the patient have been acquired or whether additional images are required 104 for a meaningful diagnosis of the patient’s actual or suspected condition.
  • a preliminary report is prepared 107 by the sonographer detailing their interpretation of the study.
  • the cardiologist reads the preliminary report and inspects the manual analyses 109 and measurements.
  • a key point is that the set of images and measurements required to be taken so as to ensure that the sonographer or cardiologist has sufficient data to diagnose the patient’s condition is comprehensive, meaning an echo study is very time-consuming for the sonographer and prone to error, such as particular data being missed during the scan by a sonographer who may be inexperienced or unfamiliar with the requirements for a particular study. In practice, however, not all possible measurements are taken, only a subset related to a suspected condition or disease are recorded by the sonographer. It is dependent upon the sonographer’ s skill and experience to know which measurements are important for subsequent analysis and diagnosis.
  • AS Severe aortic stenosis
  • AS is the most common primary valve disease leading to intervention 1 .
  • AS is associated with progressive myocardial hypertrophy and dysfunction 2 , left atrial dilatation and pulmonary hypertension 1,3 ; the clinical sequalae being heart failure and sudden death.
  • Echocardiography is pivotal to identifying AS and its accompanying adaptive response 1,3 .
  • it is highly operator dependent 4 and requires (often scarce 5 ) expert interpretation. Accordingly, minor errors in measurement of left ventricular outflow tract dimension and velocity time integral are multiplied when calculating the aortic valve area in the continuity equation, a pitfall noted in clinical guidelines 1 .
  • the transaortic gradient a robust marker of severe AS3, is heavily influenced by left ventricular systolic function.
  • a lower transaortic gradient in the setting of impaired left ventricular systolic function impacts on interpretation of AS severity.
  • Artificial Intelligence is a disruptive technology with enormous potential to improve the quality and consistency in echocardiography.
  • Recent work has shown AI has promise in image recognition 7 9 .
  • AI-based interpretation of the comprehensive measurement set performed during echocardiography has not been explored. Consequently, the inventors set-out to develop, for the first-time, a robust AI-based system using echocardiographic measurements to interpret the pathophysiology of AS.
  • the aim was to produce a reliable AI system to assist in the clinical diagnosis of severe AS without the need for potentially unreliable left ventricular outflow tract dimension and velocity measurements, and to introduce a robust quality feedback system that could be routinely applied in clinical practice.
  • AI artificial intelligence
  • One embodiment provides a computer program product for performing a method as described herein.
  • One embodiment provides a non-transitive carrier medium for carrying computer executable code that, when executed on a processor, causes the processor to perform a method as described herein.
  • One embodiment provides a system configured for performing a method as described herein.
  • a method for processing a sparsely populated data source may comprise the step of (a) retrieving data from a sparsely populated data source to form a base dataset.
  • the data source may comprise a plurality of records. Each record may comprise at least one unpopulated data field corresponding to a medical measurement.
  • the method may comprise the further step of (b) dividing the retrieved data into two portions.
  • a first portion may comprise a training dataset being a defined percentage X% of the base dataset.
  • a second portion may comprise a validation dataset being a defined percentage (100%-C%) of the base dataset.
  • the method may comprise the further step of (c) analysing the training data set to jointly model variable relationships using a non-linear function approximation algorithm. The approximation algorithm may be applied iteratively to the records of the training data set to obtain a trained model and measurement prediction protocols for populating unpopulated field in the training data set.
  • the method may comprise the further step of (d) using the measurement prediction protocols, computing prediction values for measurement data for the unpopulated data fields.
  • the method may comprise the further step of (e) imputing the predicted measurement values in the records of the training dataset.
  • the method may comprise the further step of (f) analysing the training dataset on the basis of predefined disease conditions in known patient records of the base data set to form a disease model.
  • the analysing of the training dataset may be adapted to predict a probability of a disease condition in patient records of the trained data set.
  • the method may comprise the further step of (g) imputing the predicted measurement values in the records of the validation dataset.
  • the method may comprise the further step of (h) validating the disease model.
  • Validating the disease model may comprise analysing the validation dataset using the disease model, wherein the records of the validation dataset comprise disease data associated with patient data.
  • Validating the disease model may further comprise determining a validation error comprising a probability of correctly predicting a patient disease state in the records of the validation set.
  • the method may comprise the further step of repeating Steps (c) to (h) to minimise the validation error and computing a prediction of a probable disease state for each patient record in the base dataset.
  • a method for processing a sparsely populated data source comprising: retrieving data from a sparsely populated data source to form a base dataset, the data source comprising a plurality of records, each record comprising at least one unpopulated data field corresponding to a medical measurement; dividing the retrieved data into two portions, a first portion comprising a training dataset being a defined percentage X% of the base dataset; and a second portion comprising a validation dataset being a defined percentage (100%-C%) of the base dataset; analysing the training data set to jointly model variable relationships using a non-linear function approximation algorithm applied iteratively to the records of the training data set to obtain a trained model and measurement prediction protocols for populating unpopulated field in the training data set; using the measurement prediction protocols, computing prediction values for measurement data for the unpopulated data fields; imputing the predicted measurement values in the records of the training dataset; analysing the training dataset on the basis of predefined disease conditions in known
  • unpopulated measurement data may be predicted using the measurement prediction protocols on the basis of data collected by a procedure operator.
  • a patient disease state may be predicted by the disease model on the basis of data collected by a procedure operator and on measurement data predicted using the measurement prediction protocols.
  • the unpopulated measurement data and the patient disease state prediction may be computed in real-time during the measurement procedure.
  • measurements obtained by a procedure operator may be incorporated into the training data set to form an updated data set and analysing the updated training data set using the neural network to compute updated measurement prediction protocols and an updated disease model.
  • the method may comprise the further step of analysing the measurement obtained during the measurement procedure using the updated measurement prediction protocols and updated disease model to predict a probable disease state for a patient undergoing the measurement procedure.
  • the method may comprise the further step of directing the measurement operator to record relevant measurement data to increase the confidence of the probable disease state prediction.
  • the disease condition may comprise aortic stenosis.
  • an apparatus for conducting a measurement procedure may comprise measurement tools relevant to the measurement procedure.
  • the apparatus may further comprise means for recording measurement data.
  • the apparatus may further comprise means for transmitting measurement data to an analysis means.
  • the analysis means may comprise measurement prediction protocols for predicting measurement data for unpopulated measurement fields.
  • the analysis means may further comprise a disease model for predicting a probable disease state for a patient undergoing a measurement procedure.
  • the apparatus may further comprise means for alerting the measurement operator of the predicted measurement data and probable disease state and for providing direction to the measurement operator for relevant measurement data to be collected on the basis of the predicted probable disease state.
  • the means for alerting the operator may comprise a visible notification on a display surface of the apparatus.
  • an apparatus for conducting a measurement procedure comprising:
  • means for recording measurement data means for transmitting measurement data to an analysis means, said analysis means comprising:
  • a disease model for predicting a probable disease state for a patient undergoing a measurement procedure; and means for alerting the measurement operator of the predicted measurement data and probable disease state and for providing direction to the measurement operator for relevant measurement data to be collected on the basis of the predicted probable disease state.
  • a computer program product having a computer readable medium having a computer program recorded therein for processing a sparsely populated data source.
  • the computer program product may comprise computer program code means for retrieving data from a sparsely populated data source to form a base dataset, the data source comprising a plurality of records. Each record may comprise at least one unpopulated data field corresponding to a medical measurement.
  • the computer program product may further comprise computer program code means for dividing the retrieved data into two portions.
  • a first portion may comprise a training dataset being a defined percentage X% of the base dataset.
  • a second portion may comprise a validation dataset being a defined percentage (100%-C%) of the base dataset.
  • the computer program product may further comprise computer program code means for analysing the training data set using a non-linear function approximation algorithm applied iteratively over the records of the training data set to obtain a trained data set and measurement prediction protocols for populating unpopulated field in the training data set.
  • the computer program product may further comprise computer program code means for, using the measurement prediction protocols, computing prediction values for measurement data for the unpopulated data fields.
  • the computer program product may further comprise computer program code means for imputing the predicted measurement values in the records of the training dataset.
  • the computer program product may further comprise computer program code means for analysing the training dataset on the basis of predefined disease conditions in known patient records of the base data set to form a disease model adapted to predict a probability of a disease condition in patient records of the trained data set.
  • the computer program product may further comprise computer program code means for imputing the predicted measurement values in the records of the validation dataset.
  • the computer program product may further comprise computer program code means for validating the disease model.
  • Validating the disease model may comprise analysing the validation dataset using the disease model.
  • the records of the validation dataset may comprise disease data associated with patient data.
  • Validating the disease model may further comprise determining a validation error comprising a probability of correctly predicting a patient disease state in the records of the validation set.
  • the computer program product may further comprise computer program code means for computing a prediction of a probable disease state for each patient record in the base dataset.
  • a computer program product having a computer readable medium having a computer program recorded therein for processing a sparsely populated data source
  • said computer program product comprising: computer program code means for retrieving data from a sparsely populated data source to form a base dataset, the data source comprising a plurality of records, each record comprising at least one unpopulated data field corresponding to a medical measurement; computer program code means for dividing the retrieved data into two portions: a first portion comprising a training dataset being a defined percentage X% of the base dataset; and a second portion comprising a validation dataset being a defined percentage (100%-X%) of the base dataset; computer program code means for analysing the training data set using a non-linear function approximation algorithm applied iteratively over the records of the training data set to obtain a trained data set and measurement prediction protocols for populating unpopulated field in the training data set; computer program code means for using the measurement prediction protocols, computing prediction values for measurement
  • Figure 1 shows an example of patient records in a sparsely populated dataset
  • Figure 2 shows a typical workflow procedure for existing echocardiography analysis methods
  • Figure 3 shows a typical workflow of an AI Assisted Reporting mode of echocardiography analysis according to an embodiment of the invention as disclosed herein;
  • Figure 4 shows a typical workflow of an AI-In-The-Loop mode of echocardiography analysis according to an embodiment of the invention as disclosed herein;
  • Figure 5 shows a schematic representation of the major components of an AI-architecture system adapted for AI-assisted echocardiography as disclosed herein;
  • Figure 6 shows a schematic representation of a network accessible application implementation of the AI-assisted echocardiography methods and systems as disclosed herein;
  • FIG. 7 shows a computing device on which the various embodiments described herein may be implemented in accordance with an embodiment of the present invention
  • Figure 8 shows graphical results of an embodiment of the AI-assisted echocardiography methods and systems as disclosed herein for prediction of severe Aortic Stenosis (AS) in the NEDA records for the general population
  • EF ejection fraction value
  • Figure 10 shows a schematic representation of the overall architecture of the AI model and the training process
  • Figure 11 shows the error distribution plots when the trained AI model was applied to the test subset
  • Figure 12 shows a Table 4 of a comparison of variables for patients with and without severe AS
  • Figure 13 is a schematic depiction of the flow 900 of analyses performed on data derived from the study cohort of Example 2;
  • Figure 14 depicts AUROC curves of AI-augmented diagnosis of severe AS compared with that derived from the continuity equation from the study in Example 2;
  • Figure 15 depicts actuarial 5-year mean survival for the AI diagnosis of non-severe AS from the study in Example 2.
  • the term,“real-time” for example“displaying real-time data” , refers to the display of the data without intentional delay, given the processing limitations of the system and the time required to accurately measure the data.
  • the term“near-real-time” for example“obtaining real-time or near-real-time data” refers to the obtaining of data either without intentional delay (“real-time”) or as close to real-time as practically possible (i.e. with a small, but minimal, amount of delay whether intentional or not within the constraints and processing limitations of the of the system for obtaining and recording or transmitting the data.
  • an“exemplary embodiment” is used in the sense of providing examples, as opposed to indicating quality. That is, an“exemplary embodiment” is an embodiment provided as an example, as opposed to necessarily being an embodiment of exemplary quality for example serving as a desirable model or representing the best of its kind.
  • the various methods or processes outlined herein may be coded as software that is executable on one or more processors that employ any one of a variety of operating systems or platforms. Additionally, such software may be written using any of a number of suitable programming languages and/or programming or scripting tools, and also may be compiled as executable machine language code or intermediate code that is executed on a framework or virtual machine.
  • inventive concepts may be embodied as a computer readable storage medium (or multiple computer readable storage media) (e.g. a computer memory, one or more floppy discs, compact discs, optical discs, magnetic tapes, flash memories, circuit configurations in Field Programmable Gate Arrays or other semiconductor devices, or other non-transitory medium or tangible computer storage medium) encoded with one or more programs that, when executed on one or more computers or other processors, perform methods that implement the various embodiments of the invention discussed above.
  • the computer readable medium or media can be transportable, such that the program or programs stored thereon can be loaded onto one or more different computers or other processors to implement various aspects of the present invention as discussed above.
  • program or“software” are used herein in a generic sense to refer to any type of computer code or set of computer-executable instructions that can be employed to program a computer or other processor to implement various aspects of embodiments as discussed above. Additionally, it should be appreciated that according to one aspect, one or more computer programs that when executed perform methods of the present invention need not reside on a single computer or processor, but may be distributed in a modular fashion amongst a number of different computers or processors to implement various aspects of the present invention.
  • Computer-executable instructions may be in many forms, such as program modules, executed by one or more computers or other devices.
  • program modules include routines, programs, objects, components, data structures, etc., that perform particular tasks or implement particular abstract data types.
  • functionality of the program modules may be combined or distributed as desired in various embodiments.
  • data structures may be stored in computer-readable media in any suitable form.
  • data structures may be shown to have fields that are related through location in the data structure. Such relationships may likewise be achieved by assigning storage for the fields with locations in a computer-readable medium that convey relationship between the fields.
  • any suitable mechanism may be used to establish a relationship between information in fields of a data structure, including through the use of pointers, tags or other mechanisms that establish relationship between data elements.
  • inventive concepts may be embodied as one or more methods, of which an example has been provided.
  • the acts performed as part of the method may be ordered in any suitable way. Accordingly, embodiments may be constructed in which acts are performed in an order different than illustrated, which may include performing some acts simultaneously, even though shown as sequential acts in illustrative embodiments.
  • a reference to“A and/or B" when used in conjunction with open-ended language such as“comprising” can refer, in one embodiment, to A only (optionally including elements other than B); in another embodiment, to B only (optionally including elements other than A); in yet another embodiment, to both A and B (optionally including other elements), etc.
  • “or” should be understood to have the same meaning as“and/or” as defined above.
  • “or” or“and/or” shall be interpreted as being inclusive, i.e. the inclusion of at least one, but also including more than one, of a number or list of elements, and, optionally, additional unlisted items. Only terms clearly indicated to the contrary, such as“only one of” or“exactly one of”, or, when used in the claims,“consisting of” will refer to the inclusion of exactly one element of a number or list of elements.
  • the phrase“at least one” in reference to a list of one or more elements, should be understood to mean at least one element selected from any one or more of the elements in the list of elements, but not necessarily including at least one of each and every element specifically listed within the list of elements and not excluding any combinations of elements in the list of elements.
  • This definition also allows that elements may optionally be present other than the elements specifically identified within the list of elements to which the phrase“at least one” refers, whether related or unrelated to those elements specifically identified.
  • “at least one of A and B” can refer, in one embodiment, to at least one, optionally including more than one, A, with no B present (and optionally including elements other than B); in another embodiment, to at least one, optionally including more than one, B, with no A present (and optionally including elements other than A); in yet another embodiment, to at least one, optionally including more than one, A, and at least one, optionally including more than one, B (and optionally including other elements); etc.
  • AI artificial intelligence
  • a de-identified copy of the NEDA database containing the full range of measurements, but without mortality data, is used for creating an AI model as discussed below.
  • a modified Mixture Density Network 10 is trained to serve as a multiple-imputation model.
  • the model is trained with missing data by augmenting the model with Boolean inputs signaling whether a measurement was present or not to predict severe AS.
  • example data entries from the NEDA data source had measurements randomly held out from the training inputs and used as target outputs for regression.
  • Backpropagation is only applied to model outputs with a target output present.
  • the training examples approximately resemble typical sets of measurements encountered in echo without requiring complete sets of measurements for the training process.
  • the resulting model is designed to be general-purpose and can perform inference using arbitrary sets of available measurements.
  • the atypical backpropagation procedure utilised herein can be viewed as training a family of models with shared weights (this has some similarities to techniques previously applied to restricted Boltzmann machines 14 ).
  • the input holdout process has similarities to the common technique known as dropout (although in this case rather than using the existing method of discarding the“dropped out’’ values they are used to build the sentinel vector and the target outputs) and is seen to have a secondary effect of regularizing the model, thereby encouraging the learning of more generalizable patterns 10 .
  • the continuous rank probability score (CRPS) is chosen as the loss function, since it has a closed-form solution for a mixture -of-Gaussians 11 and encourages convergence to sharp and well-calibrated predictions 12 .
  • the CRPS loss function penalises the model for predicting an incorrect expected value while also penalising over or under confident predicted distributions.
  • Figure 10 depicts the overall architecture 800 of the model and the training process.
  • Inputs 803 [xi ... x n ] are echo measurements with missing values, outputs 805 (m,. s ; where i 6 ⁇ l,...,n ⁇ denote Gaussian prediction densities with mean m, and standard deviation Oi (but the general approach may be applied to any choice of density function with closed-form solution for CRPS including mixture of Gaussians).
  • Magnified section 850 depicts the cumulative density function (CDF) of the prediction x n -2 ⁇ LG(mdress 2, s h 2) plotted for possible values of the measurement z and compared with the target value.
  • CDF cumulative density function
  • the results show minimal bias and low error rates considering the heterogeneous nature of the data and the fact that key information (i.e. LVOT data) has been removed from the studies.
  • the panels on the left demonstrate the imputed vs actual measurements overlaid.
  • the panels on the right demonstrate the imputation error (imputed vs actual measurement), calculated after predicting while holding out each measurement plus any directly dependent variables.
  • LVOT dimension O.OlOcm (-0.165 to 0.165)
  • LVOT velocity time integral -0.669cm (-6.019 to 4.249)
  • Mean transaortic valve gradient 0.068mmHg (-5.639 to 3.133)
  • Aortic Valve Area -0.056 (0.885, 0.664)
  • p ns for each imputed vs actual measurement.
  • the AI predictions for the continuity-derived aortic valve area were then evaluated in the clinical context of classifying severe AS. Initially, the test set data was filtered to only consider studies with a known aortic valve area calculated using the continuity equation and used this to label the studies as“severe AS’’ or“not severe AS’’. Then, all left ventricular outflow tract measurements (velocity, gradient and diameter), the aortic valve area, and the aortic root dimensions were left out from this test set and the model was used to predict a distribution of likely values for the aortic valve area. A predicted probability of severe AS was derived by evaluating the cumulative density function of the predictions to calculate p(AVA ⁇ 1 cm 2 ).
  • VTI LVOT is the velocity time integral of the LVOT velocity trace
  • VTI AV is the velocity time integral of the aortic valve velocity trace.
  • Severe AS is defined as an AVA ⁇ l.0cm 2 (highest measured VTI AV and the mean VTI LV0T ) . Characteristics of Patients Identified as Severe AS by AI
  • Patients identified by the AI system as having severe AS had the expected demographic and clinical characteristics (see Table 4 in Figure 12), with increased aortic valve gradients, impaired left ventricular diastolic function, and increased indexed left ventricular mass, indexed left atrial volume and right ventricular systolic pressure.
  • Patients diagnosed with severe AS by the AI system but not by the continuity equation (CE) had similar characteristics to those with severe AS by continuity, except for a lower transaortic gradient and stroke volume index, consistent with the AG s interpretation of typical cardiac structural changes in response to aortic stenosis.
  • model inputs where known: gender, height, weight, basal 2D dimensions (ventricular septal and posterior wall diastolic thickness, left ventricular internal dimension in systole and diastole), left ventricular ejection fraction measured using the Simpson’s Biplane method, ascending aortic dimension, atrial measurements (left atrial area in 4- and 2-chamber, 4-chamber left atrial length, right atrial area), mitral inflow pulsed wave Doppler data (E velocity, A velocity, E wave pressure half-time, mitral inflow velocity time integral), left ventricular diastolic basal tissue Doppler velocities (E’ septal velocity, E’ lateral velocity), transaortic velocities (aortic peak velocity, aortic velocity time integral), and pulmonary
  • Doppler measurements AV Mean Gradient, AV Mean Velocity, AV Peak Gradient, AV Peak Velocity, AV Velocity Time Integral, Mitral E’ velocity septum, Mitral E’ velocity lateral, Mitral E to E Prime Ratio Septal, Mitral E to LV E Prime Ratio Lateral, Mitral E to MV E Prime Ratio Mean, MV Mean Gradient, MV Mean Velocity, MV Peak Gradient, MV Peak Velocity, MV Pressure Half Time, MV Velocity Time Integral, Mitral A Point Velocity, Mitral E Point Velocity, Mitral E to A Ratio, PV Peak Gradient, PV Peak Velocity.
  • This limited data set takes approximately 10 minutes to acquire when performing a transthoracic echocardiogram. All other variables were held out from the test set, including left ventricular outflow tract measurements, and the AI was evaluated using the same methodology described above. This is defined as the“limited echocardiography’’ system, and the output classification was a probability of severe AS as described above.
  • Echocardiographic studies are time-consuming, and so for reasons of efficiency an echo study will only focus on measurements deemed relevant by the cardiologist. This means that the NEDA database contains an incomplete set of measurements for each patient, which presents as an obstacle to the use of many techniques from the fields of statistics and machine learning.
  • an AI-assisted echocardiography reporting aid In a first embodiment disclosed herein, there is provided an AI-assisted echocardiography reporting aid.
  • an echocardiography examination of a patient is carried out as usual and the AI-model is then used to augment the study with a set of predictions.
  • This process 300 is shown schematically in Figure 3.
  • the major difference between the prior art process 200 of Figure 2 and the AI-assisted reporting procedure 300 presented in Figure 3 is the addition of AI-model predictions 301 to impute any missing measurement data into the patient’s measurement record for the scan, which in turn enables Step 4—“ AI- Assisted Analysis’’ 303.
  • the AI- Assisted Analysis 303 uses actual measurements acquired by the sonographer at Step 1 101 and may, optionally, also include predictions by the AI-model of missing measurement parameters from Step 3 301, to provide computed estimates of the patient’s disease risk to the healthcare professional (e.g. the sonographer or cardiologist) analysing the study.
  • the sonographer then prepares the preliminary report 109 including measurement data recorded during the scan and, if utilised, measurement data imputed into the scan record by the AI system, and the report is forwarded to the cardiologist for further analysis.
  • the cardiologist may also, optionally, use an AI system (either the same as that used by the sonographer or a different AI model) to analyse 305 the scanned and/or imputed measurement data in the context of determining the patient’ s general state of health or their risk of having of possible progressing to a diseased state.
  • Process 300 also has the significant advantage of optionally being able to identify other potential patient abnormalities based on predicted data for missing measurement data imputed into the patient’s scan record, which helps the healthcare professional to pick up on subtle or uncommon conditions which may otherwise be missed by a less experienced sonographer or cardiologist. In these cases, a disease condition may otherwise progress untreated, leading it to only be detected later when more extreme symptoms have manifested. This scenario of the prior art process where subtle or less common disease indicators are missed by the healthcare providers leads to worse patient outcomes and typically also increased costs to healthcare providers.
  • a more advanced application of the AI models disclosed herein is to integrate directly with the measurement process performed by the sonographer whilst taking the measurements of the patient during a scan—“AI-In-The-Loop” .
  • This configuration of the AI model is adapted to provide real-time predictions of various echocardiography measurements to the sonographer whilst they are performing the scan on the patient.
  • Figure 4 provides a schematic representation of the workflow for the AI-In-The-Loop configuration.
  • the major benefit of this approach is that certain measurements may not be required to be taken by the sonographer as they can be predicted 401 by the system in real-time while the scan is in progress.
  • the system may also update a prediction of one or more possible diagnoses 403 of condition(s) the patient may have on the basis of the measurements acquired by the sonographer in conjunction with the predicted measurements 401. If any AI-predicted measurements have a high enough“ confidence” output from the system, they can be used as-is, saving time. Alternatively, if the sonographer is not satisfied of the confidence level of the predicted measurement, then they could make the particular measurement manually whilst the patient is present.
  • the system may also optionally request 405 particular measurements to be acquired by the sonographer on the basis of the predicted diagnoses 403, for example to manually acquire a particular measurement that may be useful in confirming or ruling out a particular predicted diagnosis made on the basis of existing acquired and predicted measurements.
  • the method 400 proposed in Figure 4 is a particularly compelling proposition, particularly in light of the fact that new measurement techniques are continually appearing in the literature and echo specialists (particularly sonographers) are increasingly required to prioritise which measurements to perform in the limited time available for a study.
  • a second benefit of the method 400 is that in the case of a subtle abnormality in patient’s heart, the system is also able to suggest further measurements to the sonographer and the data may be acquired on-the-spot whilst the patient is present. This is a significant improvement over the method 300 of Figure 3 which is only able to flag abnormalities to the healthcare professional after the patient has left the clinic, meaning the patient may be required to return for an expensive and time consuming second physical examination.
  • Figure 5 depicts the major components of an AI-architecture 500 adapted for AI-assisted echocardiography as disclosed herein and the interconnections and interactions between the component parts of system 500. Implementation details of the particular components of system 500 will typically vary by application and are subject to change and so are considered largely out-of-scope for this document.
  • System 500 is particularly adapted to connect with a database 501 of medical records which may be any form of structured medical data, which could include data derived from medical images, measurements taken during a procedure or even the output of a Natural Language Processing (NLP) algorithm from reading existing medical reports.
  • the data source 501 comprises a set of measurements taken from Echocardiogram studies, however it will be readily apparent to the skilled addressee that the methods and systems disclosed below have application in imputing missing measurement data into records of different types of data sets (medical or otherwise) where a relatively large number of possible measurement data is included in each record.
  • a portion of the data records from data source 501 is designated as training data 503 which is used to train the AI system 500. Due to the complex nature of the human body, a large number of examples are typically required. Having a large number of example records (in the order of hundreds of thousands of records) better allows the AI model to distinguish between true patterns in the data and random noise caused by various factors such as human error. Approximately 60%-80%, typically about 70%, of the records from data source 501 is selected to form the training data. The remaining 20%-40%, typically about 30%, is designated as test data 505 which is used later in the process to validate the predictive capabilities of the AI model on a real data set with known data values. As will be appreciated by the skilled addressee. It is important not to use the training data 503 for AI model validation as many AI models are prone to overfitting— which would lead to the undesirable outcome of tailoring the model to patterns in the training data that are non-generalisable“noise” rather than meaningful data.
  • this architecture of system 500 allows the possibility of online-training- as more data is collected at a clinical site and continually added to the data source 501, the AI model may be continually refined using the data records to improve its predictive performance.
  • the AI models in the presently described arrangement of system 500 are provided with an initial model state 507 state which encodes the behaviour of the AI.
  • an Artificial Neural Network (ANN) model has a set of weight coefficients which are learned during the training process.
  • a new model will typically be initialised with a random state, but in the case of online training, a previously trained model may be used as the initial model state.
  • ANN Artificial Neural Network
  • the architecture 500 presented herein allows any AI model to be used in the system 500, provided that the inputs and outputs meet a standard interface specification which is particular to the desired outcomes of the system and taking into consideration the nature of the data records in data source 501.
  • the details of the training process 509 largely depend on the AI algorithm selected, but typically algorithms will iteratively process each example and use the“mistakes” made by the model to “learn” by gradually mutating the model state. A successful training process will trend over time towards more accurate predictions.
  • the AI model arrives at (converges to) a proposed trained model 511 for data analysis, which encodes the learned AI model state.
  • Preferred training procedures 509 are non-parametric in that the training process methodology employed makes no explicit assumptions about the relationships between any of the variables or data elements in the data source 501.
  • a useful analogy of the training process 509 would be training an Al-system (such as for example, system 500) to build a model of a cube across three dimensions (i.e. the available variables or features to be modelled) of width, height and depth.
  • the AI model takes the form of a sparse-data self-organising map imputation model as discussed below.
  • n v 6 Z + be the number of variables per training example.
  • N the number of training epochs.
  • variable j be defined as
  • the imputation algorithm disclosed below provides the system 500 with the ability to fill in missing or blank data measurement fields from records in the data source 501 during model validation 513 and also to predict missing or additional data measurements in production data source 519 whilst the system is in use. Generally speaking, imputation preserves all cases by replacing missing data with an estimated value based on other available information. Once all missing values have been imputed into the data records, the data set can then be analysed using standard techniques for complete data.
  • Imputed vector
  • Equation A2 Using the identity in Equation A2 the modified Euclidean distance function simplifies to the usual definition of Euclidean distance:
  • the AI model takes the form of a sparse-data mixture density network (MDN) imputation model as discussed below (noting that any previous mathematical definitions in this document, in particular those definitions in relation to the Sparse-Data Self-Organising Map imputation model (A) discussed above, should be disregarded in favour of the definitions below).
  • MDN sparse-data mixture density network
  • n v 6 7L + be the number of variables per training example.
  • N the number of training epochs.
  • variable j be defined as:
  • PDF Gaussian probability density function
  • CDF Gaussian cumulative density function
  • n p G 7L + be the number of parameters per mixture component.
  • f nn may define any feedforward neural network with the given input and output dimensions but nominally 4 layers of width 4096 with the“leaky ReLU” activation function is sufficient.
  • the imputation algorithm disclosed below provides the system 500 with the ability to fill in missing or blank data measurement fields from records in the data source 501 during model validation 513 and also to predict missing or additional data measurements in production data source 519 whilst the system is in use. Generally speaking, imputation preserves all cases by replacing missing data with an estimated value based on other available information. In the algorithm below, the estimated values for missing data are provided as probability distributions of likely values.
  • the classification algorithm disclosed below provides the system 500 with the ability to classify the likely severity of disease in a patient based on the imputed density function for a given measurement and established clinical thresholds for that measurement, from records in the data source 501 during model validation 513 and from measurements in production data source 519 whilst the system is in use.
  • T— 1 clinical thresholds for variable j defining T gradings of disease severity e.g. normal, mild, moderate, severe
  • the dataset of a cube (Table 1) is substituted by the NED A database which allows a model (phenotype) of a heart to be built across several variables.
  • This phenotype model data describes the interrelation of variables across many different heart configurations and diseases.
  • Model validation 513 makes use of the test data 505 to evaluate the performance of the trained AI model 511 on new, previously unseen data.
  • the outputs of the validation process 513 largely depend on the particular application and the data source 501 used which are tested against particular performance metrics 515 to evaluate the predictive power (e.g. the probability that a predicted data measurement will be within defined tolerance levels as compared with the actual measurement data).
  • An example of a useful performance metric 515 is Root-Mean-Square Error of predicted measurements compared with known data in the test data 505.
  • the trained model 511 is validated to be able to provide meaningful measurement predictions within pre -defined tolerance levels (defined according to typical measurement error ranges taken from validation studies plus a tolerance of about 5%)
  • the trained model 511 is imported to the primary predictive engine 517 of system 500.
  • the prediction engine 517 is then used to analyse new data e.g. production data set 519, during system operation.
  • production data set 519 may be a database of measurements in an echo reporting software package, or it may be measurements on an echo/ultrasound workstation obtained during working of methods 200 or 300 disclosed above.
  • the outputs from the AI prediction engine 517 comprise predicted measurements 521 and/or predicted diagnoses 523 with respect to the production data source 519, for example, obtained during working of methods 200 or 300 disclosed above.
  • the outputs of the AI prediction engine 517 may be a set of predictions for measurements 521 that were not provided as inputs or alternatively a predicted diagnosis on the basis of a set of pre-defined risk factors for various diseases relevant to the nature of the data source, for example where the data source comprises echocardiography measurements, the AI prediction engine 517 may provide predictions of a particular patient’s probability of possessing, or likely to subsequently possess, heart-related diseases such as, for example arterial stenosis or heart chamber or valve malfunction.
  • AI prediction engine 517 may also utilise a tailored implementation of the AI algorithm to produce predictions in the required operating (e.g. software) environment. For example, an implementation for an echo software package would likely be implemented to run on Microsoft Windows, whereas an implementation for an echo workstation would need to be tailored to the vendor-specific hardware and operating system running on the workstation.
  • an implementation for an echo software package would likely be implemented to run on Microsoft Windows, whereas an implementation for an echo workstation would need to be tailored to the vendor-specific hardware and operating system running on the workstation.
  • a further example embodiment of the systems disclosed above may be implemented as an application which is made available to, for example, healthcare professionals, over a communications network such as, for example, a diagnostic clinic intranet network or a publicly accessible communications network such as the internet.
  • the network-accessible application may be provided as an interactive web application with a simple workflow, such as depicted in an example wireframe schematic representation 600 shown in Figure 6.
  • a user enters a set of echo measurements (or alternatively connects the input to a database of echo measurements e.g. for a plurality of patients) into an input interface 601 of web interface 610 (where the application 600 is accessible over the worldwide web/internet).
  • the backend of application 600 (not shown) is connected to a AI system such as system 500 depicted in Figure 5.
  • Application 600 feeds the inputted user data to the backend analysis system so as to return outputs comprising predicted measurement data 620 to fill in any blank or missing data measurements from the input data and further to output predicted disease risk factor 630.
  • Predicted measurements and disease risk factors are then presented to the user via interface 610 to a user display means 640, which may present the output predictions in any useful manner for interpretation by the user, for example the outputs may be presented in a graphical form for easy interpretation by the user.
  • the methods of training and operating an AI-assisted echocardiography system as disclosed herein may be implemented using a computer system 700, such as that shown in Figure 7 wherein the processes of Figures 2 to 5 may be implemented as software, such as one or more application programs executable within the computing device 700.
  • the steps of method(s) 200, 300, 400 and 500 are effected by instructions in the software that are carried out within the computer system 700.
  • the instructions may be formed as one or more code modules, each for performing one or more particular tasks.
  • the software may also be divided into two separate parts, in which a first part and the corresponding code modules performs the described methods and a second part and the corresponding code modules manage a user interface between the first part and the user.
  • the software may be stored in a computer readable medium, including the storage devices described below, for example.
  • the software is loaded into the computer system 700 from the computer readable medium, and then executed by the computer system 700.
  • a computer readable medium having such software or computer program recorded on it is a computer program product.
  • the use of the computer program product in the computer system 700 preferably effects an advantageous apparatus for AI assisted echocardiography.
  • the exemplary computing device 700 can include, but is not limited to, one or more central processing units (CPUs) 701 comprising one or more processors 702, a system memory 703, and a system bus 704 that couples various system components including the system memory 703 to the processing unit 701.
  • the system bus 704 may be any of several types of bus structures including a memory bus or memory controller, a peripheral bus, and a local bus using any of a variety of bus architectures.
  • the computing device 700 also typically includes computer readable media, which can include any available media that can be accessed by computing device 700 and includes both volatile and non-volatile media and removable and non-removable media.
  • computer readable media may comprise computer storage media and communication media.
  • Computer storage media includes media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data.
  • Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by the computing device 700.
  • Communication media typically embodies computer readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media.
  • communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media. Combinations of the any of the above should also be included within the scope of computer readable media.
  • the system memory 703 includes computer storage media in the form of volatile and/or non-volatile memory such as read only memory (ROM) 705 and random-access memory (RAM) 706.
  • ROM read only memory
  • RAM random-access memory
  • a basic input/output system 707 (BIOS) containing the basic routines that help to transfer information between elements within computing device 700, such as during start-up, is typically stored in ROM 705.
  • BIOS basic input/output system 707
  • RAM 706- typically contains data and/or program modules that are immediately accessible to and/or presently being operated on by processing unit 701.
  • Figure 7 illustrates an operating system 708, other program modules 709, and program data 710.
  • the computing device 700 may also include other removable/non-removable, volatile/non-volatile computer storage media.
  • Figure 7 illustrates a hard disk drive 711 that reads from or writes to non-removable, non-volatile magnetic media.
  • Other removable/non-removable, volatile/non-volatile computer storage media that can be used with the exemplary computing device include, but are not limited to, magnetic tape cassettes, flash memory cards, digital versatile disks, digital video tape, solid state RAM, solid state ROM, and the like.
  • the hard disk drive 711 is typically connected to the system bus 704 through a non-removable memory interface such as interface 712.
  • the drives and their associated computer storage media discussed above and illustrated in Figure 7, provide storage of computer readable instructions, data structures, program modules and other data for the computing device 700.
  • hard disk drive 711 is illustrated as storing an operating system 713, other program modules 714, and program data 715. Note that these components can either be the same as or different from operating system 708, other program modules 709 and program data 710.
  • Operating system 3013, other program modules 714 and program data 715 are given different numbers hereto illustrate that, at a minimum, they are different copies.
  • the computing device also includes one or more input/output (I/O) interfaces 730 connected to the system bus 704 including an audio-video interface that couples to output devices including one or more of a video display 734 and loudspeakers 735.
  • I/O interface(s) 730 also couple(s) to one or more input devices including, for example a mouse 731, keyboard 732 or touch sensitive device 733 such as for example a smartphone or tablet device.
  • input interface 730 may also comprise an echocardiography/ultrasound hand piece and computing device 700 may comprise or be integrated with an echo/ultrasound workstation.
  • the computing device 700 may operate in a networked environment using logical connections to one or more remote computers.
  • the computing device 700 is shown in Figure 7 to be connected to a network 720 that is not limited to any particular network or networking protocols, but which may include, for example Ethernet, Bluetooth or IEEE 802.X wireless protocols.
  • the logical connection depicted in Figure 7 is a general network connection 721 that can be a local area network (LAN), a wide area network (WAN) or other network, for example, the internet.
  • the computing device 700 is connected to the general network connection 721 through a network interface or adapter 722 which is, in turn, connected to the system bus 704.
  • program modules depicted relative to the computing device 700, or portions or peripherals thereof, may be stored in the memory of one or more other computing devices that are communicatively coupled to the computing device 700 through the general network connection 721. It will be appreciated that the network connections shown are exemplary and other means of establishing a communications link between computing devices may be used.
  • the following examples demonstrate the utility of the AI-assisted echocardiography methods and systems disclosed above.
  • the methods are used to predict the incidence of aortic stenosis while completely removing the need for left ventricular outflow tract measurements by the use of artificial intelligence. It is observed that the above disclosed systems and methods improve the consistency of echo, in addition to saving a significant amount of scanning time for the sonographer performing echocardiography studies.
  • LVOT diameter is a circular approximation of cross-sectional area measured by the sonographer during an echocardiogram study of a patient used in evaluation of the continuity equation (CE) above.
  • the estimated AV Area is relatively independent of flow and so in patients with low flow, it becomes a critical parameter. However, it is not without its own difficulties due to its dependence on the estimated LVOT Area. To quote the current guidelines:
  • AV Area LVOT Area * LVOT VTI / AV VTI
  • a snapshot of the NEDA database (approx. 435,000 patients) was taken and split into a 70% training set and a 30% test set.
  • the training set was used to train an imputation model using the above -disclosed Self-Organising Map algorithm.
  • test set was subsampled to only include patients with a complete set of measurements for:
  • the AV Area was calculated for each patient in the subsampled test set using Equation (4) above and a“ground truth” binary label was generated for each patient in the subsampled test set, labelling them as either“Severe AS” or“Not Severe AS”.
  • the patients were labelled as“Severe AS” if and only if AV Area ⁇ 1cm 2 .
  • Predicted AV Area was calculated for each patient using inputs oTPredicted LVOT DiameterPredicted LVOT VT Measured AV VTI.
  • the measurements included in the imputation model include AR Peak Velocity; AR Pressure Half Time; AV Mean Gradient; AV Mean Velocity; AV Peak Velocity; AV Velocity Time Integral; Aorta at Sinotubular Diameter; Aorta at Sinuses Diameter; Aortic Root Diameter; Aortic Root Diameter MM; Ascending Aorta Diameter; Body Weight; Diastolic Slope; Distal Transverse Aorta Diameter; EFtext; Heart Rate; IVC Diameter Expiration; IVS Diastolic Thickness; IVS Diastolic Thickness MM; IVS Systolic Thickness MM; LA Area 4C View; LA Length 4C; LA Systolic Diameter LX; LA Systolic Diameter MM; LA Systolic Diameter Transverse; LV Diameter; LV Diastolic Area 2C; LV Diastolic Area 4C; LV Diastolic Area PSAX; LV Dias
  • Figure 8 shows the Receiver Operating Characteristic and Precision-Recall curve area for this example.
  • the total patients in the sample was 24748 and the diagnosis was based on the following measurements: AV Area (VTI), Holding out: [“LVOT Mean Velocity’’,“LVOT Peak Velocity’’,“LVOT Velocity Time Integral’’ , and“LVOT Diameter’’].
  • the number of patients predicted with severe AS was 1834 (7.410700%) and the number of patients predicted without severe was 22914 (92.589300%).
  • Example IB AS Prediction in Patients with Reduced Ejection Fraction
  • Step 1 for ejection fractions of 40%, 35%, 30%.
  • the total number of patients in the sample was 1391.
  • the number of patients predicted with severe AS was 143 (10.280374%) and the number of patients predicted without severe was 1248 (89.719626%)
  • the total patients in the sample was 861.
  • the number of patients predicted with severe AS was 96 (11.149826%) and the number of patients predicted without severe AS was 765 (88.850174%).
  • the number of patients predicted with severe AS was 58 (11.026616%) and the number of patients predicted without severe AS was 468 (88.973384%).
  • the total patients in the sample was 426.
  • the number of patients predicted with severe AS was 45 (10.563380%) and the number of patients predicted without severe AS was 381 (89.436620%).
  • Average precision is generally a more informative metric than AUC for evaluating binary classifier performance with imbalanced data sets (see Appendix A for explanation). With this in mind, the results are interpreted as follows.
  • ROC Receiver Operating Characteristic
  • ⁇ Recall/TPR is the percentage of patients with severe AS that the system correctly identifies.
  • ⁇ FPR is the percentage of patients without severe AS that the system incorrectly diagnoses with severe AS
  • ⁇ Precision is the fraction of positive diagnoses for severe AS generated by the system that are correct.
  • the FPR metric is relative to the number of patients without severe AS, which in this setting is the majority of patients. Hence an FPR that seems low actually translates to a large number of false positive diagnoses.
  • the precision metric is relative to the number of positive diagnoses generated by the algorithm. This gives a better sense of how the system would be perceived in actual usage. For example, a precision of 80% means that if the system diagnoses a patient with severe AS then there is an 80% chance it is correct.
  • Example 2 Severe Aortic Stenosis
  • Performance of the AI system was assessed using receiver operating characteristic (ROC) curves, with the prediction of severe AS by AI against the aortic valve area calculated using the continuity equation.
  • the area under the ROC (AUROC) curve was calculated for all patients and for those with impaired left ventricular function. All fatal events were identified, and the last recorded echocardiogram for those patients with complete follow up used to calculate actuarial 5-year mortality survival curves.
  • Cox proportional hazard ratios adjusted for age and gender were calculated, along with further adjustments for the aortic valve area and mean aortic gradient. To establish potential differences in survival of those with AI-diagnosed severe AS compared those with severe AS using calculated aortic valve area, actuarial 5-year survival curves were examined.
  • the phenotypic diagnosis of severe AS were compared using the AI system to that of the continuity-derived severe AS.
  • three groups of patients (full system and limited data-set) were compared: 1) All patients in the test set, 2) those with an ejection fraction ⁇ 50% and 3) those with an ejection fraction ⁇ 30%.
  • the AI prediction of severe AS used all available data (minus the left ventricular outflow tract data), with missing measurements ignored by the AI.
  • Figure 13 shows the flow 900 of analyses performed on data derived from the study cohort of 171,571 males & 158,404 females (aged 61.5+17.6 years) with a median 4.1 (IQR 2.2, 7.1 years) follow. There were no differences in the baseline characteristics between the 70% test set and the 30% training set, nor for those with a complete set of data available for aortic valve area calculation using the continuity equation (Table 4). A total of 2382/32,574 individuals (7.38% 95% Cl 7.10 to 7.67%) from the 30% test with a known continuity-derived aortic valve area had severe AS. An output probability cut-off of 0.065425 provided an identical sensitivity and specificity of 91.4% for severe AS diagnosed by AI when compared with the calculated aortic valve area.
  • the AUCROC curve 951 of AI-augmented diagnosis of severe AS compared with that derived from the continuity equation was 0.9696 (see Figure 14) with a positive predictive value of 45.9%.
  • the model performed almost as well 953 in those with an ejection fraction ⁇ 50% ( Figure 14) with an AUCROC of 0.9462 (2308 patients, 11% with severe AS).
  • the AI also performed 955 very well in those with an ejection fraction ⁇ 30% (491 studies, 13%- many with low gradient, low-output severe AS) with an AUCROC of 0.9200 ( Figure 14).
  • Actuarial 5-year mean (+/-standard error of mean) survival for the AI diagnosis of non-severe AS was 1536.0+/-8.8 days vs 1072.5+/-23.3 days for severe AS, p ⁇ 0.00001, representing a mean survival difference of 463.5 days (Figure 15) which shows the actuarial survival curves for diagnosis of severe aortic stenosis using artificial intelligence vs a traditional continuity diagnosis.
  • the patients were matched across the 30% test cohort.
  • the upper line 961 in each of panels (A) and (B) represents the number of individuals at risk without severe aortic stenosis at each time period.
  • the lower line 963 in each of panels (A) and (B) represents number of individuals at risk diagnosed with severe aortic stenosis at each time period.
  • the mean survival was 1489.0+/-8.9 days vs 1086.0+/-31.6 days, a difference of 403 days.
  • AI can robustly augment the diagnosis of severe AS by interpreting the entire echocardiographic phenotype without reliance on left ventricular outflow tract measurements in a very large cohort of individuals subject to prolonged follow-up. Moreover, consistent with its multi-parameter approach to interpretation, an AI-augmented diagnosis of severe AS remained a significant predictor of long-term mortality even after adjustment for traditional AS severity measures. Accordingly, the purpose-built AI systems disclosed herein also introduce the first potential quality system for echocardiography by providing automatic measurement and disease predictions in real-time. These disclosed systems can provide a known statistical outcome for a defined set of measurements. Importantly, the fully trained AI system takes minimal computing power to operate and can be installed on both echocardiography machines and imaging reading software to improve diagnostic consistency in the absence of expert review.
  • AI offers an extremely useful clinical tool; particularly in low-resource settings where specialist cardiologists are scarce.
  • AS complex interactions present in AS require evaluation by a subspecialist-trained echocardiographer.
  • under-diagnosis of severe AS may occur in some individuals 15 .
  • the AI evaluated in this study consistently examines the entire echocardiography phenotype; taking into account the known pathophysiologic changes 19 22 such as left ventricular diastolic and systolic dysfunction, left atrial enlargement and pulmonary hypertension 23 .
  • the AI systems disclosed herein are trained on data from NEDA, a very large echo database linked with mortality. Because of the nature and scope of NEDA, quality control against individual images was not feasible and data obtained from individual laboratories is assumed to be correct. However, systematic bias is unlikely because NEDA is sourced from different tertiary hospital laboratories across Australia rather than a single source. Although the AI accurately identified severe AS, the model needs to be tested across the whole range of aortic valve disease; noting the error for imputation of the aortic valve area was small (see Table 4), with no systematic bias and a mean imputation error (95% Cl) of-0.056 (-0.885, 0.664)cm 2 .
  • the AI may identify some individuals without severe AS, but with similar cardiac phenotypic changes. However, these patients had a similar mortality trajectory to those with traditional severe AS, highlighting the capability of the AI to identify those at high risk.
  • the systems disclosed herein have not yet been validated in populations outside of Australia, although Australia is a multicultural nation broadly representative of the world’s population, with over 300 different ancestries and 28% of the resident population born overseas. Also clinical linkage associated with the data sets used was not available for inclusion in the system validation examples discussed above. Potential contributors to the phenotypic changes such as hypertension or valve intervention may also need to be considered in the clinical context of the individual.
  • AI can augment and improve the diagnosis of severe AS and associated risk mortality predictions, by analyzing the entire echo phenotype and without the need for left ventricular outflow tract measurements.
  • the AI systems disclosed herein were able to predict patients at higher mortality risk due to AS, independently of aortic valve gradients. Decisions for intervention in severe aortic stenosis (AS) depend on reliable echocardiographic interpretation.
  • the artificial intelligence (AI) systems disclosed herein are particularly designed to augment the diagnosis of severe AS using echocardiographic measurement data, thus providing a significant input into development of effective patient care outcomes by healthcare professionals.
  • bus and its derivatives, while being described in a preferred embodiment as being a communication bus subsystem for interconnecting various devices including by way of parallel connectivity such as Industry Standard Architecture (ISA), conventional Peripheral Component Interconnect (PCI) and the like or serial connectivity such as PCI Express (PCIe), Serial Advanced Technology Attachment (Serial AT A) and the like, should be construed broadly herein as any system for communicating data.
  • parallel connectivity such as Industry Standard Architecture (ISA), conventional Peripheral Component Interconnect (PCI) and the like
  • serial connectivity such as PCI Express (PCIe), Serial Advanced Technology Attachment (Serial AT A) and the like
  • a computer implemented method should not necessarily be inferred as being performed by a single computing device such that the steps of the method may be performed by more than one cooperating computing devices.
  • objects as used herein such as ‘web server’, ‘server’, ‘client computing device’, ‘computer readable medium’ and the like should not necessarily be construed as being a single object, and may be implemented as a two or more objects in cooperation, such as, for example, a web server being construed as two or more web servers in a server farm cooperating to achieve a desired goal or a computer readable medium being distributed in a composite manner, such as program code being provided on a compact disk activatable by a license key downloadable from a computer network.
  • database and its derivatives may be used to describe a single database, a set of databases, a system of databases or the like.
  • the system of databases may comprise a set of databases wherein the set of databases may be stored on a single implementation or span across multiple implementations.
  • database is also not limited to refer to a certain database format rather may refer to any database format.
  • database formats may include MySQL, MySQLi , XML or the like.
  • the invention may be embodied using devices conforming to other network standards and for other applications, including, for example other WLAN standards and other wireless standards.
  • Applications that can be accommodated include IEEE 802.11 wireless LANs and links, and wireless Ethernet.
  • the term“ wireless” and its derivatives may be used to describe circuits, devices, systems, methods, techniques, communications channels, etc., that may communicate data through the use of modulated electromagnetic radiation through a non-solid medium. The term does not imply that the associated devices do not contain any wires, although in some embodiments they might not.
  • the term“wired” and its derivatives may be used to describe circuits, devices, systems, methods, techniques, communications channels, etc., that may communicate data through the use of modulated electromagnetic radiation through a solid medium. The term does not imply that the associated devices are coupled by electrically conductive wires.
  • processor may refer to any device or portion of a device that processes electronic data, e.g. from registers and/or memory to transform that electronic data into other electronic data that, e.g. may be stored in registers and/or memory.
  • A“ computer” or a“computing device” or a“computing machine” or a“computing platform” may include one or more processors.
  • the methodologies described herein are, in one embodiment, performable by one or more processors that accept computer-readable (also called machine-readable) code containing a set of instructions that when executed by one or more of the processors carry out at least one of the methods described herein.
  • Any processor capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken are included.
  • a typical processing system that includes one or more processors.
  • the processing system further may include a memory subsystem including main RAM and/or a static RAM, and/or ROM.
  • a computer-readable carrier medium may form, or be included in a computer program product.
  • a computer program product can be stored on a computer usable carrier medium, the computer program product comprising a computer readable program means for causing a processor to perform a method as described herein.
  • the one or more processors operate as a standalone device or may be connected, e.g. networked to other processor(s), in a networked deployment, the one or more processors may operate in the capacity of a server or a client machine in server-client network environment, or as a peer machine in a peer-to-peer or distributed network environment.
  • the one or more processors may form a web appliance, a network router, switch or bridge, or any machine capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that machine.
  • each of the methods described herein is in the form of a computer-readable carrier medium carrying a set of instructions, e.g. a computer program that are for execution on one or more processors.
  • a computer-readable carrier medium carrying computer readable code including a set of instructions that when executed on one or more processors cause a processor or processors to implement a method.
  • aspects of the present invention may take the form of a method, an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects.
  • the present invention may take the form of carrier medium (e.g. a computer program product on a computer-readable storage medium) carrying computer-readable program code embodied in the medium.
  • some of the embodiments are described herein as a method or combination of elements of a method that can be implemented by a processor or a processor device, computer system, or by other means of carrying out the function.
  • a processor with the necessary instructions for carrying out such a method or element of a method forms a means for carrying out the method or element of a method.
  • an element described herein of an apparatus embodiment is an example of a means for carrying out the function performed by the element for the purpose of carrying out the invention.
  • a device A connected to a device B should not be limited to devices or systems wherein an output of device A is directly connected to an input of device B. It means that there exists a path between an output of A and an input of B which may be a path including other devices or means.
  • Connected may mean that two or more elements are either in direct physical or electrical contact, or that two or more elements are not in direct contact with each other but yet still co-operate or interact with each other.
  • references throughout this specification to “one embodiment’’, “an embodiment’’ , “one arrangement’’ or“an arrangement’’ means that a particular feature, structure or characteristic described in connection with the embodiment/arrangement is included in at least one embodiment/arrangement of the present invention.
  • appearances of the phrases “in one embodiment/arrangement’’ or “in an embodiment/arrangement’’ in various places throughout this specification are not necessarily all referring to the same embodiment/arrangement, but may.
  • the particular features, structures or characteristics may be combined in any suitable manner, as would be apparent to one of ordinary skill in the art from this disclosure, in one or more embodiments/arrangements.
  • Valvular aortic stenosis and coronary atherosclerosis pathophysiology and clinical consequences. European Heart Journal 1988 ;9 Suppl E:83-6.
  • Pibarot P Dumesnil JG. New concepts in valvular hemodynamics: implications for diagnosis and treatment of aortic stenosis.
  • Mutlak D Aronson D, Carasso S, Lessick J, Reisner SA, Agmon Y.

Abstract

A method for processing a sparsely populated data source comprising: retrieving data from a sparsely populated data source whose records comprise at least one unpopulated data field corresponding to a medical measurement; dividing the data into a training data set and a validation data set; analysing the training data set using a non-linear function approximation algorithm to obtain a trained model and measurement prediction protocols for populating unpopulated fields in the training data set; using the measurement prediction protocols to predict data for the unpopulated data fields; analysing the training dataset on the basis of predefined disease conditions in known patient records to form a disease model which predicts a probability of a disease condition; validating the disease model by analysing the validation dataset and determining a validation error; repeating steps to minimise the validation error and predict a probable disease state for each patient record.

Description

SYSTEMS AND METHODS FOR AI- ASSISTED ECHOCARDIOGRAPHY
FIELD OF THE INVENTION
[0001] The present invention relates to artificial intelligence and in particular to systems and methods for artificial intelligence integration to analysis of medical data and records.
[0002] The invention has been developed primarily for use in methods and systems for systems and methods for AI-assisted echocardiography and will be described hereinafter with reference to this application. However, it will be appreciated that the invention is not limited to this particular field of use.
BACKGROUND OF THE INVENTION
[0003] Any discussion of the background art throughout the specification should in no way be considered as an admission that such background art is prior art, nor that such background art is widely known or forms part of the common general knowledge in the field in Australia or worldwide.
[0004] All references, including any patents or patent applications, cited in this specification are hereby incorporated by reference. No admission is made that any reference constitutes prior art. The discussion of the references states what their authors assert, and the applicants reserve the right to challenge the accuracy and pertinence of the cited documents. It will be clearly understood that, although a number of prior art publications are referred to herein, this reference does not constitute an admission that any of these documents forms part of the common general knowledge in the art, in Australia or in any other country.
[0005] Echocardiography (or simply “echo ") is a sub-specialty of cardiology that makes use of specialised ultrasound equipment to take diagnostic images of the heart. It is particularly valuable as a first-line diagnostic tool due to its ability to non-invasively assess the internal structure and function of the heart in a cost-effective manner.
[0006] A comprehensive Echo procedure can measure many features of the heart totaling in approximately 150 variables but the full set is rarely measured due to time constraints.
[0007] Aggregate echocardiogram data from a wide range of clinics Australia-wide are stored in the National Echocardiogram Database Australia (NEDA). In its current form, NEDA consists of measurements taken from echo procedures and report texts that are an analysis from that procedure for many patients. Currently, the database has over 500k patients.
[0008] This unique NEDA resource collates all echocardiographic measurement and report data contained in the echocardiographic database of participating centers. Each database is then remotely transferred into the Master NEDA Database via a“vendor-agnostic” , automated data extraction process that transfers every measurement for each echocardiogram performed into a standardized NEDA data format (according to the NEDA Data Dictionary ). Each individual contributing to NEDA is given a unique identifier along with their demographic profile (date of birth and sex) and all data recorded with their echocardiogram.
[0009] Individual linkage to mortality outcomes of individuals contributing to NEDA (including the date and listed causes of death) is periodically performed via the validated National Deaths Index14. NEDA has received Human Research Ethics Committee clearance by local and national ethics committees, and is registered by the Australasian Clinical Trials Registry (htp://www.anzctr.org.ao/ACTRN12617001387314.aspx). All data storage and analyses are maintained and performed on a de-identified basis to protect participant anonymity.
[0010] Each patient record in the database usually does not contain a full set of measurements that can be possibly taken during an Echo procedure due to time constraints. Measurements are targeted to a subset of measurements that may be related to a suspected diagnosis or condition of the heart. This results in NEDA being a sparse data set with each patient echo record consisting of a subset of a full echo procedure leaving“blanks” for measurements that are not measured in a full procedure. Figure 1 shows a few real examples of different measurements present and missing in echo studies for a small randomly selected group of patients, being a typical example of sparse echo data where each row of the table is a record of the echo measurement available for a single patient.
[0011] The workflow for a typical prior art echocardiography study process is depicted in Figure 2. In summary, a typical workflow consists of the following steps:
Images are acquired 101 by a sonographer using a special ultrasound workstation.
Specific physical features are measured 103 from the acquired echo images by the sonographer, for example, the diameter of the left ventricle is commonly measured. (Note: many of the measurements will typically be taken during the image acquisition process with the patient present, while others may be measured from the images acquired during the procedure once the patient has left). In this workflow, the sonographer has complete and sole control over whether or not sufficient images of the patient have been acquired or whether additional images are required 104 for a meaningful diagnosis of the patient’s actual or suspected condition.
The images and measurements are manually interpreted 105 by the sonographer.
A preliminary report is prepared 107 by the sonographer detailing their interpretation of the study.
The cardiologist reads the preliminary report and inspects the manual analyses 109 and measurements.
The cardiologist creates a final report 111 with their remarks and conclusions from the study.
[0012] A key point is that the set of images and measurements required to be taken so as to ensure that the sonographer or cardiologist has sufficient data to diagnose the patient’s condition is comprehensive, meaning an echo study is very time-consuming for the sonographer and prone to error, such as particular data being missed during the scan by a sonographer who may be inexperienced or unfamiliar with the requirements for a particular study. In practice, however, not all possible measurements are taken, only a subset related to a suspected condition or disease are recorded by the sonographer. It is dependent upon the sonographer’ s skill and experience to know which measurements are important for subsequent analysis and diagnosis.
[0013] Accordingly, there is a need for methods of obtaining complete echocardiography datasets for diagnosis of abnormal or disease conditions in patients within time and cost constraints of physical echocardiography procedures. Also, there is a need for methods and systems for determining meaningful data to provide a complete echocardiography record for past echocardiography patients.
[0014] Severe aortic stenosis (AS) is the most common primary valve disease leading to intervention1. Without timely intervention, AS is associated with progressive myocardial hypertrophy and dysfunction2, left atrial dilatation and pulmonary hypertension1,3; the clinical sequalae being heart failure and sudden death. Echocardiography is pivotal to identifying AS and its accompanying adaptive response1,3. However, it is highly operator dependent4 and requires (often scarce5) expert interpretation. Accordingly, minor errors in measurement of left ventricular outflow tract dimension and velocity time integral are multiplied when calculating the aortic valve area in the continuity equation, a pitfall noted in clinical guidelines1.
[0015] The transaortic gradient, a robust marker of severe AS3, is heavily influenced by left ventricular systolic function. A lower transaortic gradient in the setting of impaired left ventricular systolic function impacts on interpretation of AS severity. These technical challenges have resulted in calls for improved automation and objective echo interpretation systems6. Artificial Intelligence (AI) is a disruptive technology with enormous potential to improve the quality and consistency in echocardiography. Recent work has shown AI has promise in image recognition7 9. However, AI-based interpretation of the comprehensive measurement set performed during echocardiography has not been explored. Consequently, the inventors set-out to develop, for the first-time, a robust AI-based system using echocardiographic measurements to interpret the pathophysiology of AS. Specifically, the aim was to produce a reliable AI system to assist in the clinical diagnosis of severe AS without the need for potentially unreliable left ventricular outflow tract dimension and velocity measurements, and to introduce a robust quality feedback system that could be routinely applied in clinical practice.
SUMMARY OF THE INVENTION
[0016] It is an object of the present invention to overcome or ameliorate at least one or more of the disadvantages of the prior art, or to provide a useful alternative.
[0017] The systems and methods disclosed herein provide artificial intelligence (AI) systems for predicting missing measurements from echocardiography data records and providing risk assessments for diseases from incomplete data records with AI populated data.
[0018] One embodiment provides a computer program product for performing a method as described herein.
[0019] One embodiment provides a non-transitive carrier medium for carrying computer executable code that, when executed on a processor, causes the processor to perform a method as described herein.
[0020] One embodiment provides a system configured for performing a method as described herein.
[0021] According to a first aspect of the present invention, there is provided a method for processing a sparsely populated data source. The method may comprise the step of (a) retrieving data from a sparsely populated data source to form a base dataset. The data source may comprise a plurality of records. Each record may comprise at least one unpopulated data field corresponding to a medical measurement.
[0022] The method may comprise the further step of (b) dividing the retrieved data into two portions. A first portion may comprise a training dataset being a defined percentage X% of the base dataset. A second portion may comprise a validation dataset being a defined percentage (100%-C%) of the base dataset. [0023] The method may comprise the further step of (c) analysing the training data set to jointly model variable relationships using a non-linear function approximation algorithm. The approximation algorithm may be applied iteratively to the records of the training data set to obtain a trained model and measurement prediction protocols for populating unpopulated field in the training data set.
[0024] The method may comprise the further step of (d) using the measurement prediction protocols, computing prediction values for measurement data for the unpopulated data fields.
[0025] The method may comprise the further step of (e) imputing the predicted measurement values in the records of the training dataset.
[0026] The method may comprise the further step of (f) analysing the training dataset on the basis of predefined disease conditions in known patient records of the base data set to form a disease model. The analysing of the training dataset may be adapted to predict a probability of a disease condition in patient records of the trained data set.
[0027] The method may comprise the further step of (g) imputing the predicted measurement values in the records of the validation dataset.
[0028] The method may comprise the further step of (h) validating the disease model. Validating the disease model may comprise analysing the validation dataset using the disease model, wherein the records of the validation dataset comprise disease data associated with patient data. Validating the disease model may further comprise determining a validation error comprising a probability of correctly predicting a patient disease state in the records of the validation set.
[0029] The method may comprise the further step of repeating Steps (c) to (h) to minimise the validation error and computing a prediction of a probable disease state for each patient record in the base dataset.
[0030] According to a particular arrangement of the first aspect, there is provided a method for processing a sparsely populated data source comprising: retrieving data from a sparsely populated data source to form a base dataset, the data source comprising a plurality of records, each record comprising at least one unpopulated data field corresponding to a medical measurement; dividing the retrieved data into two portions, a first portion comprising a training dataset being a defined percentage X% of the base dataset; and a second portion comprising a validation dataset being a defined percentage (100%-C%) of the base dataset; analysing the training data set to jointly model variable relationships using a non-linear function approximation algorithm applied iteratively to the records of the training data set to obtain a trained model and measurement prediction protocols for populating unpopulated field in the training data set; using the measurement prediction protocols, computing prediction values for measurement data for the unpopulated data fields; imputing the predicted measurement values in the records of the training dataset; analysing the training dataset on the basis of predefined disease conditions in known patient records of the base data set to form a disease model adapted to predict a probability of a disease condition in patient records of the trained data set; imputing the predicted measurement values in the records of the validation dataset; validating the disease model comprising analysing the validation dataset using the disease model, wherein the records of the validation dataset comprise disease data associated with patient data, and determining a validation error comprising a probability of correctly predicting a patient disease state in the records of the validation set; repeating Steps (c) to (h) to minimise the validation error and computing a prediction of a probable disease state for each patient record in the base dataset. [0031] The sparsely populated data source may comprise a plurality of medical records comprising measurement data obtained from a medical study procedure. The medical study procedure may comprise an echocardiography procedure.
[0032] During a measurement procedure, unpopulated measurement data may be predicted using the measurement prediction protocols on the basis of data collected by a procedure operator.
[0033] During a measurement procedure, a patient disease state may be predicted by the disease model on the basis of data collected by a procedure operator and on measurement data predicted using the measurement prediction protocols.
[0034] The unpopulated measurement data and the patient disease state prediction may be computed in real-time during the measurement procedure.
[0035] During a measurement procedure, measurements obtained by a procedure operator may be incorporated into the training data set to form an updated data set and analysing the updated training data set using the neural network to compute updated measurement prediction protocols and an updated disease model. The method may comprise the further step of analysing the measurement obtained during the measurement procedure using the updated measurement prediction protocols and updated disease model to predict a probable disease state for a patient undergoing the measurement procedure.
[0036] As a result of the probable disease state prediction, the method may comprise the further step of directing the measurement operator to record relevant measurement data to increase the confidence of the probable disease state prediction. The disease condition may comprise aortic stenosis.
[0037] According to a second aspect of the invention, there is provided an apparatus for conducting a measurement procedure. The apparatus may comprise measurement tools relevant to the measurement procedure. The apparatus may further comprise means for recording measurement data. The apparatus may further comprise means for transmitting measurement data to an analysis means. The analysis means may comprise measurement prediction protocols for predicting measurement data for unpopulated measurement fields. The analysis means may further comprise a disease model for predicting a probable disease state for a patient undergoing a measurement procedure. The apparatus may further comprise means for alerting the measurement operator of the predicted measurement data and probable disease state and for providing direction to the measurement operator for relevant measurement data to be collected on the basis of the predicted probable disease state. The means for alerting the operator may comprise a visible notification on a display surface of the apparatus.
[0038] According to a particular arrangement of the second aspect, there is provided an apparatus for conducting a measurement procedure comprising:
measurement tools relevant to the measurement procedure;
means for recording measurement data; means for transmitting measurement data to an analysis means, said analysis means comprising:
measurement prediction protocols for predicting measurement data for unpopulated measurement fields; and
a disease model for predicting a probable disease state for a patient undergoing a measurement procedure; and means for alerting the measurement operator of the predicted measurement data and probable disease state and for providing direction to the measurement operator for relevant measurement data to be collected on the basis of the predicted probable disease state.
[0039] According to a third aspect of the invention, there is provided a computer program product having a computer readable medium having a computer program recorded therein for processing a sparsely populated data source. The computer program product may comprise computer program code means for retrieving data from a sparsely populated data source to form a base dataset, the data source comprising a plurality of records. Each record may comprise at least one unpopulated data field corresponding to a medical measurement.
[0040] The computer program product may further comprise computer program code means for dividing the retrieved data into two portions. A first portion may comprise a training dataset being a defined percentage X% of the base dataset. A second portion may comprise a validation dataset being a defined percentage (100%-C%) of the base dataset.
[0041] The computer program product may further comprise computer program code means for analysing the training data set using a non-linear function approximation algorithm applied iteratively over the records of the training data set to obtain a trained data set and measurement prediction protocols for populating unpopulated field in the training data set.
[0042] The computer program product may further comprise computer program code means for, using the measurement prediction protocols, computing prediction values for measurement data for the unpopulated data fields.
[0043] The computer program product may further comprise computer program code means for imputing the predicted measurement values in the records of the training dataset.
[0044] The computer program product may further comprise computer program code means for analysing the training dataset on the basis of predefined disease conditions in known patient records of the base data set to form a disease model adapted to predict a probability of a disease condition in patient records of the trained data set.
[0045] The computer program product may further comprise computer program code means for imputing the predicted measurement values in the records of the validation dataset.
[0046] The computer program product may further comprise computer program code means for validating the disease model. Validating the disease model may comprise analysing the validation dataset using the disease model. The records of the validation dataset may comprise disease data associated with patient data. Validating the disease model may further comprise determining a validation error comprising a probability of correctly predicting a patient disease state in the records of the validation set.
[0047] The computer program product may further comprise computer program code means for computing a prediction of a probable disease state for each patient record in the base dataset.
[0048] According to a particular arrangement of the third aspect, there is provided a computer program product having a computer readable medium having a computer program recorded therein for processing a sparsely populated data source, said computer program product comprising: computer program code means for retrieving data from a sparsely populated data source to form a base dataset, the data source comprising a plurality of records, each record comprising at least one unpopulated data field corresponding to a medical measurement; computer program code means for dividing the retrieved data into two portions: a first portion comprising a training dataset being a defined percentage X% of the base dataset; and a second portion comprising a validation dataset being a defined percentage (100%-X%) of the base dataset; computer program code means for analysing the training data set using a non-linear function approximation algorithm applied iteratively over the records of the training data set to obtain a trained data set and measurement prediction protocols for populating unpopulated field in the training data set; computer program code means for using the measurement prediction protocols, computing prediction values for measurement data for the unpopulated data fields; computer program code means for imputing the predicted measurement values in the records of the training dataset; computer program code means for analysing the training dataset on the basis of predefined disease conditions in known patient records of the base data set to form a disease model adapted to predict a probability of a disease condition in patient records of the trained data set; computer program code means for imputing the predicted measurement values in the records of the validation dataset; computer program code means for validating the disease model comprising analysing the validation dataset using the disease model, wherein the records of the validation dataset comprise disease data associated with patient data, and determining a validation error comprising a probability of correctly predicting a patient disease state in the records of the validation set; and computer program code means for computing a prediction of a probable disease state for each patient record in the base dataset.
BRIEF DESCRIPTION OF THE DRAWINGS
[0049] Notwithstanding any other forms which may fall within the scope of the present invention, preferred embodiments of the invention will now be described, by way of example only, with reference to the accompanying drawings in which:
Figure 1 shows an example of patient records in a sparsely populated dataset;
Figure 2 shows a typical workflow procedure for existing echocardiography analysis methods; and
Figure 3 shows a typical workflow of an AI Assisted Reporting mode of echocardiography analysis according to an embodiment of the invention as disclosed herein;
Figure 4 shows a typical workflow of an AI-In-The-Loop mode of echocardiography analysis according to an embodiment of the invention as disclosed herein;
Figure 5 shows a schematic representation of the major components of an AI-architecture system adapted for AI-assisted echocardiography as disclosed herein;
Figure 6 shows a schematic representation of a network accessible application implementation of the AI-assisted echocardiography methods and systems as disclosed herein;
Figure 7 shows a computing device on which the various embodiments described herein may be implemented in accordance with an embodiment of the present invention;
Figure 8 shows graphical results of an embodiment of the AI-assisted echocardiography methods and systems as disclosed herein for prediction of severe Aortic Stenosis (AS) in the NEDA records for the general population; Figures 9A, 9B, 9C and 9D shows graphical results of the prediction accuracy of an embodiment of the AI-assisted echocardiography methods and systems as disclosed herein for prediction of severe Aortic Stenosis (AS) in the NEDA records for subsets of the general population, respectively for ejection fraction value (EF) of <= 50%, EF <=40%, EF <= 35%, and EF <=30%;
Figure 10 shows a schematic representation of the overall architecture of the AI model and the training process;
Figure 11 shows the error distribution plots when the trained AI model was applied to the test subset;
Figure 12 shows a Table 4 of a comparison of variables for patients with and without severe AS;
Figure 13 is a schematic depiction of the flow 900 of analyses performed on data derived from the study cohort of Example 2;
Figure 14 depicts AUROC curves of AI-augmented diagnosis of severe AS compared with that derived from the continuity equation from the study in Example 2; and
Figure 15 depicts actuarial 5-year mean survival for the AI diagnosis of non-severe AS from the study in Example 2.
DEFINITIONS
[0050] The following definitions are provided as general definitions and should in no way limit the scope of the present invention to those terms alone, but are put forth for a better understanding of the following description.
[0051] Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by those of ordinary skill in the art to which the invention belongs. It will be further understood that terms used herein should be interpreted as having a meaning that is consistent with their meaning in the context of this specification and the relevant art and will not be interpreted in an idealized or overly formal sense unless expressly so defined herein. For the purposes of the present invention, additional terms are defined below. Furthermore, all definitions, as defined and used herein, should be understood to control over dictionary definitions, definitions in documents incorporated by reference, and/or ordinary meanings of the defined terms unless there is doubt as to the meaning of a particular term, in which case the common dictionary definition and/or common usage of the term will prevail.
[0052] For the purposes of the present invention, the following terms are defined below.
[0053] The articles“a" and“an" are used herein to refer to one or to more than one (i.e. to at least one) of the grammatical object of the article. By way of example,“an element” refers to one element or more than one element.
[0054] The term“about" is used herein to refer to quantities that vary by as much as 30%, preferably by as much as 20%, and more preferably by as much as 10% to a reference quantity. The use of the word ‘about’ to qualify a number is merely an express indication that the number is not to be construed as a precise value. [0055] Throughout this specification, unless the context requires otherwise, the words “comprise”, “comprises” and“comprising” will be understood to imply the inclusion of a stated step or element or group of steps or elements but not the exclusion of any other step or element or group of steps or elements.
[0056] Any one of the terms:“including” or“which includes” or“that includes” as used herein is also an open term that also means including at least the elements/features that follow the term, but not excluding others. Thus,“including” is synonymous with and means“comprising”.
[0057] In the claims, as well as in the summary above and the description below, all transitional phrases such as “comprising” , “including” , “carrying”, “having”, “containing” , “involving” , “holding” , “ composed of”, and the like are to be understood to be open-ended, i.e. to mean“including but not limited to”. Only the transitional phrases“consisting of” and“consisting essentially of” alone shall be closed or semi -closed transitional phrases, respectively.
[0058] The term,“real-time” , for example“displaying real-time data” , refers to the display of the data without intentional delay, given the processing limitations of the system and the time required to accurately measure the data.
[0059] The term“near-real-time” , for example“obtaining real-time or near-real-time data” refers to the obtaining of data either without intentional delay (“real-time”) or as close to real-time as practically possible (i.e. with a small, but minimal, amount of delay whether intentional or not within the constraints and processing limitations of the of the system for obtaining and recording or transmitting the data.
[0060] Although any methods and materials similar or equivalent to those described herein can be used in the practice or testing of the present invention, preferred methods and materials are described. It will be appreciated that the methods, apparatus and systems described herein may be implemented in a variety of ways and for a variety of purposes. The description here is by way of example only.
[0061] As used herein, the term“exemplary” is used in the sense of providing examples, as opposed to indicating quality. That is, an“exemplary embodiment” is an embodiment provided as an example, as opposed to necessarily being an embodiment of exemplary quality for example serving as a desirable model or representing the best of its kind.
[0062] The various methods or processes outlined herein may be coded as software that is executable on one or more processors that employ any one of a variety of operating systems or platforms. Additionally, such software may be written using any of a number of suitable programming languages and/or programming or scripting tools, and also may be compiled as executable machine language code or intermediate code that is executed on a framework or virtual machine.
[0063] In this respect, various inventive concepts may be embodied as a computer readable storage medium (or multiple computer readable storage media) (e.g. a computer memory, one or more floppy discs, compact discs, optical discs, magnetic tapes, flash memories, circuit configurations in Field Programmable Gate Arrays or other semiconductor devices, or other non-transitory medium or tangible computer storage medium) encoded with one or more programs that, when executed on one or more computers or other processors, perform methods that implement the various embodiments of the invention discussed above. The computer readable medium or media can be transportable, such that the program or programs stored thereon can be loaded onto one or more different computers or other processors to implement various aspects of the present invention as discussed above. [0064] The terms“program” or“software” are used herein in a generic sense to refer to any type of computer code or set of computer-executable instructions that can be employed to program a computer or other processor to implement various aspects of embodiments as discussed above. Additionally, it should be appreciated that according to one aspect, one or more computer programs that when executed perform methods of the present invention need not reside on a single computer or processor, but may be distributed in a modular fashion amongst a number of different computers or processors to implement various aspects of the present invention.
[0065] Computer-executable instructions may be in many forms, such as program modules, executed by one or more computers or other devices. Generally, program modules include routines, programs, objects, components, data structures, etc., that perform particular tasks or implement particular abstract data types. Typically, the functionality of the program modules may be combined or distributed as desired in various embodiments.
[0066] Also, data structures may be stored in computer-readable media in any suitable form. For simplicity of illustration, data structures may be shown to have fields that are related through location in the data structure. Such relationships may likewise be achieved by assigning storage for the fields with locations in a computer-readable medium that convey relationship between the fields. However, any suitable mechanism may be used to establish a relationship between information in fields of a data structure, including through the use of pointers, tags or other mechanisms that establish relationship between data elements.
[0067] Also, various inventive concepts may be embodied as one or more methods, of which an example has been provided. The acts performed as part of the method may be ordered in any suitable way. Accordingly, embodiments may be constructed in which acts are performed in an order different than illustrated, which may include performing some acts simultaneously, even though shown as sequential acts in illustrative embodiments.
[0068] The phrase“and/or", as used herein in the specification and in the claims, should be understood to mean“either or both" of the elements so conjoined, i.e. elements that are conjunctively present in some cases and disjunctively present in other cases. Multiple elements listed with“and/or" should be construed in the same fashion, i.e.“one or more " of the elements so conjoined. Other elements may optionally be present other than the elements specifically identified by the“and/or" clause, whether related or unrelated to those elements specifically identified. Thus, as a non-limiting example, a reference to“A and/or B", when used in conjunction with open-ended language such as“comprising” can refer, in one embodiment, to A only (optionally including elements other than B); in another embodiment, to B only (optionally including elements other than A); in yet another embodiment, to both A and B (optionally including other elements), etc.
[0069] As used herein in the specification and in the claims,“or" should be understood to have the same meaning as“and/or" as defined above. For example, when separating items in a list,“or" or“and/or" shall be interpreted as being inclusive, i.e. the inclusion of at least one, but also including more than one, of a number or list of elements, and, optionally, additional unlisted items. Only terms clearly indicated to the contrary, such as“only one of" or“exactly one of", or, when used in the claims,“consisting of" will refer to the inclusion of exactly one element of a number or list of elements. In general, the term“or" as used herein shall only be interpreted as indicating exclusive alternatives (i.e.“one or the other but not both”) when preceded by terms of exclusivity, such as“either” ,“one of” ,“only one of” , or“exactly one of’ . “Consisting essentially of”, when used in the claims, shall have its ordinary meaning as used in the field of patent law.
[0070] As used herein in the specification and in the claims, the phrase“at least one” , in reference to a list of one or more elements, should be understood to mean at least one element selected from any one or more of the elements in the list of elements, but not necessarily including at least one of each and every element specifically listed within the list of elements and not excluding any combinations of elements in the list of elements. This definition also allows that elements may optionally be present other than the elements specifically identified within the list of elements to which the phrase“at least one” refers, whether related or unrelated to those elements specifically identified. Thus, as a non-limiting example, “at least one of A and B” (or, equivalently,“at least one of A or B”, or, equivalently“at least one of A and/or B”) can refer, in one embodiment, to at least one, optionally including more than one, A, with no B present (and optionally including elements other than B); in another embodiment, to at least one, optionally including more than one, B, with no A present (and optionally including elements other than A); in yet another embodiment, to at least one, optionally including more than one, A, and at least one, optionally including more than one, B (and optionally including other elements); etc.
[0071] For the purpose of this specification, where method steps are described in sequence, the sequence does not necessarily mean that the steps are to be carried out in chronological order in that sequence, unless there is no other logical manner of interpreting the sequence.
[0072] In addition, where features or aspects of the invention are described in terms of Markush groups, those skilled in the art will recognise that the invention is also thereby described in terms of any individual member or subgroup of members of the Markush group.
DETAILED DESCRIPTION
[0073] It should be noted in the following description that like or the same reference numerals in different embodiments denote the same or similar features.
[0074] As discussed above, the systems and methods disclosed herein provide artificial intelligence (AI) systems and methods for predicting missing measurements from echocardiography data records and providing risk assessments for diseases from incomplete data records with AI populated data.
[0075] A de-identified copy of the NEDA database containing the full range of measurements, but without mortality data, is used for creating an AI model as discussed below. Using a random 70% subset of patients, a modified Mixture Density Network10 is trained to serve as a multiple-imputation model. The model is trained with missing data by augmenting the model with Boolean inputs signaling whether a measurement was present or not to predict severe AS.
[0076] In particular, example data entries from the NEDA data source had measurements randomly held out from the training inputs and used as target outputs for regression. Backpropagation is only applied to model outputs with a target output present. In this way, the training examples approximately resemble typical sets of measurements encountered in echo without requiring complete sets of measurements for the training process. The resulting model is designed to be general-purpose and can perform inference using arbitrary sets of available measurements. The atypical backpropagation procedure utilised herein can be viewed as training a family of models with shared weights (this has some similarities to techniques previously applied to restricted Boltzmann machines14). The input holdout process has similarities to the common technique known as dropout (although in this case rather than using the existing method of discarding the“dropped out’’ values they are used to build the sentinel vector and the target outputs) and is seen to have a secondary effect of regularizing the model, thereby encouraging the learning of more generalizable patterns10. The continuous rank probability score (CRPS) is chosen as the loss function, since it has a closed-form solution for a mixture -of-Gaussians 11 and encourages convergence to sharp and well-calibrated predictions12. Intuitively, the CRPS loss function penalises the model for predicting an incorrect expected value while also penalising over or under confident predicted distributions. Figure 10 depicts the overall architecture 800 of the model and the training process. The overall training process consisting of random input holdout 801 followed by backpropagation from the target outputs. Inputs 803 [xi ... xn] are echo measurements with missing values, outputs 805 (m,. s ; where i 6 { l,...,n} denote Gaussian prediction densities with mean m, and standard deviation Oi (but the general approach may be applied to any choice of density function with closed-form solution for CRPS including mixture of Gaussians). Magnified section 850 depicts the cumulative density function (CDF) of the prediction xn-2 ~ LG(m„ 2, sh 2) plotted for possible values of the measurement z and compared with the target value. CRPS is calculated by integrating the squared height of the shaded area over all values of z.
[0077] The model was tested by applying it to the remaining 30% test subset of patients not used for training. As an initial diagnostic, selected groups of related measurements were withheld and the AI predictions were evaluated against the known measured values. These results indicated the predicted measurements had minimal bias and surprisingly low error bounds considering the heterogeneous nature of the data and the fact that key information (i.e. left ventricular outflow tract data) had been removed from the studies. Figure 11 shows the error distribution plots when the trained model was applied to the test subset. Demonstration of the error plots for prediction of measurements in the 30% test set.
[0078] In Figure 11, the results show minimal bias and low error rates considering the heterogeneous nature of the data and the fact that key information (i.e. LVOT data) has been removed from the studies. The panels on the left demonstrate the imputed vs actual measurements overlaid. The panels on the right demonstrate the imputation error (imputed vs actual measurement), calculated after predicting while holding out each measurement plus any directly dependent variables. The mean errors (95% confidence interval) were as follows: LVOT dimension = O.OlOcm (-0.165 to 0.165), LVOT velocity time integral =-0.669cm (-6.019 to 4.249), Mean transaortic valve gradient = 0.068mmHg (-5.639 to 3.133), Aortic Valve Area =-0.056 (0.885, 0.664), p=ns for each imputed vs actual measurement.
[0079] The AI predictions for the continuity-derived aortic valve area were then evaluated in the clinical context of classifying severe AS. Initially, the test set data was filtered to only consider studies with a known aortic valve area calculated using the continuity equation and used this to label the studies as“severe AS’’ or“not severe AS’’. Then, all left ventricular outflow tract measurements (velocity, gradient and diameter), the aortic valve area, and the aortic root dimensions were left out from this test set and the model was used to predict a distribution of likely values for the aortic valve area. A predicted probability of severe AS was derived by evaluating the cumulative density function of the predictions to calculate p(AVA < 1 cm2). ROC analysis was then performed to quantify classifier performance by comparing the AI predicted probabilities with the original classification label using the standard, continuity-derived AVA. It will be readily appreciated by the skilled addressee that this method lends itself to the application of this type of model to existing workflows, processes and guidelines. Severe Aortic Stenosis
[0080] AS was evaluated using the peak aortic jet velocity, aortic mean gradient, and the aortic valve area defined by the Continuity equation13 (CE):
Figure imgf000014_0001
where:
Figure imgf000014_0002
VTILVOT is the velocity time integral of the LVOT velocity trace; and
VTIAV is the velocity time integral of the aortic valve velocity trace.
[0081] Severe AS is defined as an AVA < l.0cm2 (highest measured VTIAV and the mean VTILV0T ). Characteristics of Patients Identified as Severe AS by AI
[0082] Patients identified by the AI system as having severe AS had the expected demographic and clinical characteristics (see Table 4 in Figure 12), with increased aortic valve gradients, impaired left ventricular diastolic function, and increased indexed left ventricular mass, indexed left atrial volume and right ventricular systolic pressure. Patients diagnosed with severe AS by the AI system but not by the continuity equation (CE) had similar characteristics to those with severe AS by continuity, except for a lower transaortic gradient and stroke volume index, consistent with the AG s interpretation of typical cardiac structural changes in response to aortic stenosis.
Application to a Limited Data Set
[0083] To determine which echocardiographic parameters most influenced the AI in making its prediction, measurements were chosen which classically represent a pressure-loaded left ventricle in severe AS. These were then used to create a limited test set which contained only the following variables as model inputs (where known): gender, height, weight, basal 2D dimensions (ventricular septal and posterior wall diastolic thickness, left ventricular internal dimension in systole and diastole), left ventricular ejection fraction measured using the Simpson’s Biplane method, ascending aortic dimension, atrial measurements (left atrial area in 4- and 2-chamber, 4-chamber left atrial length, right atrial area), mitral inflow pulsed wave Doppler data (E velocity, A velocity, E wave pressure half-time, mitral inflow velocity time integral), left ventricular diastolic basal tissue Doppler velocities (E’ septal velocity, E’ lateral velocity), transaortic velocities (aortic peak velocity, aortic velocity time integral), and pulmonary valve peak velocity.
(a) Complete Set of Measured Variables
[0084] The full set of variables used for training the artificial intelligence in the“comprehensive” exam were as follows.
[0085] General— Body Height, Body Weight, Body Mass Index, Body Surface Area, Gender, Age,
Heart Rate [0086] 2-Dimensional Measurements— IVS Diastolic Thickness, LVPW Diastolic Thickness, LA Area, LA Length 4C, LA Systolic Diameter LX, LA Volume, LA Volume Index, LV Diastolic Area 2C, LV Diastolic Area 4C, LV Diastolic Area PSAX, LV Diastolic Diameter PLAX, LV Systolic Diameter MM, LV Systolic Diameter PLAX, LV Diastolic Length 2C, LV Diastolic Length 4C, LV Diastolic Volume 2C, LV Diastolic Volume 2D Teich, LV Diastolic Volume 4C, LV Diastolic Volume BP, LV Ejection Fraction 2C, LV Ejection Fraction 4C, LV Ejection Fraction BP, LV Stroke Volume 2C AL, LV Stroke Volume 2D Teich, LV Stroke Volume 4C AL, LV Stroke Volume MOD 2C, LV Stroke Volume MOD 4C, LV Stroke Volume MOD BP, LV Stroke Volume SIM, LV Systolic Area 2C, LV Systolic Area 4C, LV Systolic Diameter 4C, LV Systolic Length 2C, LV Systolic Length 4C, LV Systolic Volume 2C AL, LV Systolic Volume 2D Teich, LV Systolic Volume 4C AL, LV Systolic Volume MOD 2C, LV Systolic Volume MOD 4C, LV Systolic Volume MOD BP, Ejection Fraction (any method), Aorta at Sinotubular Diameter, Aorta at Sinuses Diameter, Ascending Aorta Diameter, Distal Transverse Aorta Diameter, IVC Diameter Expiration, IVC Diameter Inspiration, LV Epi Diastolic Area PSAX, LV Mass AL, LV Mass AL Index, RA Area, Right Atrial Volume, Right Atrial Volume Index.
[0087] Doppler measurements— MV Mean Gradient, MV Mean Velocity, MV Peak Gradient, MV Peak Velocity, MV Pressure Half Time, MV Velocity Time Integral, MV Area PHT, MV Deceleration Slope, MV Deceleration Time, Mitral A wave Velocity, Mitral E wave Velocity, Mitral E to A Ratio, MV E’ Septal Velocity, LV E’ Lateral Velocity, LV A’ Velocity, Mitral E to E Prime Septal Ratio, Mitral E to E Prime Lateral Ratio, Mitral mean E to E’ Ratio, AR Peak Velocity, AR Pressure Half Time, AV Mean Gradient, AV Mean Velocity, AV Peak Gradient, AV Peak Velocity, AV Velocity Time Integral, Pulm Vein Atrial Duration, Pulm Vein Atrial Reversal Velocity, Pulmonary Vein A Velocity, Pulmonary Vein Diastolic Velocity, Pulmonary Vein S/D Ratio, Pulmonary Vein Systolic Velocity, TR Peak Gradient, TR Peak Velocity, Right Ventricular Systolic Pressure, PV Mean Gradient, PV Mean Velocity, PV Peak Gradient, PV Peak Velocity, PV Velocity Time Integral, RVOT Mean Gradient, RVOT Mean Velocity, RVOT Peak Gradient, RVOT Peak Velocity, RVOT Velocity Time Integral.
(b) Reduced/Limited set of Measured Variables
[0088] The set of variables included in the“limited’’ exam were as follows.
[0089] General— Gender, Body Height, Body Weight, Body Mass Index, Body Surface Area
[0090] 2-dimensional measurements— Ejection Fraction (by any method), LV Systolic Diameter Base PLAX, LV Diastolic Diameter base PLAX, LV Systolic Diameter PLAX, LVPW Diastolic Thickness, IVS Diastolic Thickness, LA Area, RA Area, Aorta at Sinotubular Diameter, Aorta at Sinuses Diameter, Ascending Aorta Diameter.
[0091] Doppler measurements— AV Mean Gradient, AV Mean Velocity, AV Peak Gradient, AV Peak Velocity, AV Velocity Time Integral, Mitral E’ velocity septum, Mitral E’ velocity lateral, Mitral E to E Prime Ratio Septal, Mitral E to LV E Prime Ratio Lateral, Mitral E to MV E Prime Ratio Mean, MV Mean Gradient, MV Mean Velocity, MV Peak Gradient, MV Peak Velocity, MV Pressure Half Time, MV Velocity Time Integral, Mitral A Point Velocity, Mitral E Point Velocity, Mitral E to A Ratio, PV Peak Gradient, PV Peak Velocity.
[0092] This limited data set takes approximately 10 minutes to acquire when performing a transthoracic echocardiogram. All other variables were held out from the test set, including left ventricular outflow tract measurements, and the AI was evaluated using the same methodology described above. This is defined as the“limited echocardiography’’ system, and the output classification was a probability of severe AS as described above.
Imputation
[0093] Echocardiographic studies are time-consuming, and so for reasons of efficiency an echo study will only focus on measurements deemed relevant by the cardiologist. This means that the NEDA database contains an incomplete set of measurements for each patient, which presents as an obstacle to the use of many techniques from the fields of statistics and machine learning.
[0094] A naive solution to this problem is“complete case analysis’’, in which a subset of measurements are selected and any patient data with an incomplete set of measurements is discarded. This approach is flawed for two reasons:
1. It results in large amounts of useful information being discarded.
2. It introduces sampling bias- the fact that a subpopulation of patients have a certain set of measurements is likely to be associated with a specific family of diseases. In statistical terms, the measurements are Missing Not At Random (MNAR).
[0095] A better alternative is to“fill in the blanks’’ or, in statistical terms, to impute the missing values.
AI- Assisted Reporting
[0096] In a first embodiment disclosed herein, there is provided an AI-assisted echocardiography reporting aid. In the present embodiment, an echocardiography examination of a patient is carried out as usual and the AI-model is then used to augment the study with a set of predictions. This process 300 is shown schematically in Figure 3.
[0097] The major difference between the prior art process 200 of Figure 2 and the AI-assisted reporting procedure 300 presented in Figure 3 is the addition of AI-model predictions 301 to impute any missing measurement data into the patient’s measurement record for the scan, which in turn enables Step 4—“ AI- Assisted Analysis’’ 303. The AI- Assisted Analysis 303 uses actual measurements acquired by the sonographer at Step 1 101 and may, optionally, also include predictions by the AI-model of missing measurement parameters from Step 3 301, to provide computed estimates of the patient’s disease risk to the healthcare professional (e.g. the sonographer or cardiologist) analysing the study. The sonographer then prepares the preliminary report 109 including measurement data recorded during the scan and, if utilised, measurement data imputed into the scan record by the AI system, and the report is forwarded to the cardiologist for further analysis. In addition to the AI-assisted analysis undertaken by the sonographer in Step 4 303, the cardiologist may also, optionally, use an AI system (either the same as that used by the sonographer or a different AI model) to analyse 305 the scanned and/or imputed measurement data in the context of determining the patient’ s general state of health or their risk of having of possible progressing to a diseased state.
[0098] The key efficiency gain from process 300 of prior art process 200 is in the use of AI-assisted analysis techniques as discussed herein to reduce of the time spent by both the sonographer and the cardiologist in manually checking the measurements for the presence or absence of abnormalities in the patient’s scan results which would lead to a particular diagnosis of the patient’s state of health. Process 300 also has the significant advantage of optionally being able to identify other potential patient abnormalities based on predicted data for missing measurement data imputed into the patient’s scan record, which helps the healthcare professional to pick up on subtle or uncommon conditions which may otherwise be missed by a less experienced sonographer or cardiologist. In these cases, a disease condition may otherwise progress untreated, leading it to only be detected later when more extreme symptoms have manifested. This scenario of the prior art process where subtle or less common disease indicators are missed by the healthcare providers leads to worse patient outcomes and typically also increased costs to healthcare providers.
AI in the Loop
[0099] A more advanced application of the AI models disclosed herein is to integrate directly with the measurement process performed by the sonographer whilst taking the measurements of the patient during a scan—“AI-In-The-Loop” . This configuration of the AI model is adapted to provide real-time predictions of various echocardiography measurements to the sonographer whilst they are performing the scan on the patient. Figure 4 provides a schematic representation of the workflow for the AI-In-The-Loop configuration.
[0100] The major benefit of this approach is that certain measurements may not be required to be taken by the sonographer as they can be predicted 401 by the system in real-time while the scan is in progress. The system may also update a prediction of one or more possible diagnoses 403 of condition(s) the patient may have on the basis of the measurements acquired by the sonographer in conjunction with the predicted measurements 401. If any AI-predicted measurements have a high enough“ confidence” output from the system, they can be used as-is, saving time. Alternatively, if the sonographer is not satisfied of the confidence level of the predicted measurement, then they could make the particular measurement manually whilst the patient is present. The system may also optionally request 405 particular measurements to be acquired by the sonographer on the basis of the predicted diagnoses 403, for example to manually acquire a particular measurement that may be useful in confirming or ruling out a particular predicted diagnosis made on the basis of existing acquired and predicted measurements.
[0101] The method 400 proposed in Figure 4 is a particularly compelling proposition, particularly in light of the fact that new measurement techniques are continually appearing in the literature and echo specialists (particularly sonographers) are increasingly required to prioritise which measurements to perform in the limited time available for a study.
[0102] A second benefit of the method 400 is that in the case of a subtle abnormality in patient’s heart, the system is also able to suggest further measurements to the sonographer and the data may be acquired on-the-spot whilst the patient is present. This is a significant improvement over the method 300 of Figure 3 which is only able to flag abnormalities to the healthcare professional after the patient has left the clinic, meaning the patient may be required to return for an expensive and time consuming second physical examination.
System Architecture
[0103] Figure 5 depicts the major components of an AI-architecture 500 adapted for AI-assisted echocardiography as disclosed herein and the interconnections and interactions between the component parts of system 500. Implementation details of the particular components of system 500 will typically vary by application and are subject to change and so are considered largely out-of-scope for this document.
[0104] System 500 is particularly adapted to connect with a database 501 of medical records which may be any form of structured medical data, which could include data derived from medical images, measurements taken during a procedure or even the output of a Natural Language Processing (NLP) algorithm from reading existing medical reports. In the embodiments disclosed herein, the data source 501 comprises a set of measurements taken from Echocardiogram studies, however it will be readily apparent to the skilled addressee that the methods and systems disclosed below have application in imputing missing measurement data into records of different types of data sets (medical or otherwise) where a relatively large number of possible measurement data is included in each record.
[0105] A portion of the data records from data source 501 is designated as training data 503 which is used to train the AI system 500. Due to the complex nature of the human body, a large number of examples are typically required. Having a large number of example records (in the order of hundreds of thousands of records) better allows the AI model to distinguish between true patterns in the data and random noise caused by various factors such as human error. Approximately 60%-80%, typically about 70%, of the records from data source 501 is selected to form the training data. The remaining 20%-40%, typically about 30%, is designated as test data 505 which is used later in the process to validate the predictive capabilities of the AI model on a real data set with known data values. As will be appreciated by the skilled addressee. It is important not to use the training data 503 for AI model validation as many AI models are prone to overfitting— which would lead to the undesirable outcome of tailoring the model to patterns in the training data that are non-generalisable“noise” rather than meaningful data.
[0106] It is important to note that this architecture of system 500 allows the possibility of online-training- as more data is collected at a clinical site and continually added to the data source 501, the AI model may be continually refined using the data records to improve its predictive performance.
[0107] The AI models in the presently described arrangement of system 500 are provided with an initial model state 507 state which encodes the behaviour of the AI. For example, an Artificial Neural Network (ANN) model has a set of weight coefficients which are learned during the training process. A new model will typically be initialised with a random state, but in the case of online training, a previously trained model may be used as the initial model state.
[0108] As would be appreciated, the architecture 500 presented herein allows any AI model to be used in the system 500, provided that the inputs and outputs meet a standard interface specification which is particular to the desired outcomes of the system and taking into consideration the nature of the data records in data source 501. The details of the training process 509 largely depend on the AI algorithm selected, but typically algorithms will iteratively process each example and use the“mistakes” made by the model to “learn” by gradually mutating the model state. A successful training process will trend over time towards more accurate predictions.
[0109] After the training process 509 has completed, the AI model arrives at (converges to) a proposed trained model 511 for data analysis, which encodes the learned AI model state. Preferred training procedures 509 are non-parametric in that the training process methodology employed makes no explicit assumptions about the relationships between any of the variables or data elements in the data source 501. A useful analogy of the training process 509 would be training an Al-system (such as for example, system 500) to build a model of a cube across three dimensions (i.e. the available variables or features to be modelled) of width, height and depth.
[0110] This is in contrast to traditional modelling techniques which assume an explicit parametric relationship or probabilistic model. Where, human trial-and-error is required to find relationships that are effective approximations. As the required degree of model fidelity increases this becomes difficult to the point of near-impossibility for a human when the number of variables is high and there are complex interdependences- the number of possibilities become too numerous for a human to try.
A. Sparse-Data Self-Organising Map Imputation Model
[0111] In particular embodiments of system 500 the AI model takes the form of a sparse-data self-organising map imputation model as discussed below.
A.1 Definitions
[0112] Let mt 6 Z+bc the number of training examples.
[0113] Let nv 6 Z + be the number of variables per training example.
[0114] Let N be the number of training epochs.
[0115] Let the“indicator matrix’’ be defined as:
Figure imgf000019_0001
[0116] Let the“retrieval function’’ for example t, variable j be defined as
Figure imgf000019_0002
[0117] Let the z-score normalisation constants be:
Figure imgf000019_0003
[0118] Let the z-score transformation function for variable j be:
Figure imgf000019_0004
[0119] Let a normalised example vector be: [0120] Let an example indicator vector be:
Figure imgf000020_0004
[0121] Let the map dimensions be defined as:
Figure imgf000020_0001
[0122] Let the“map weights” at iteration t be
Figure imgf000020_0005
that is, a real- valued tensor of dimensions
Figure imgf000020_0012
[0123] Let a“weight vector’’ at grid coordinates (x, y) and iteration t be:
Figure imgf000020_0002
[0124] Let the binary operator O represent element-wise vector multiplication:
Figure imgf000020_0006
[0125] Let the modified Euclidean vector distance function be:
Figure imgf000020_0007
[0126] Let the neighbourhood radius at time t be:
Figure imgf000020_0008
Figure imgf000020_0009
[0127] Let the learning rate at time t be:
Figure imgf000020_0010
[0128] Let the variable observation fractions be:
Figure imgf000020_0003
[0129] Let the variable observation correction factors be:
Figure imgf000020_0011
[0130] Let the toroidal squared distance function between two points
Figure imgf000021_0005
grid of size m X n be:
Figure imgf000021_0004
A.2 Training Algorithm
A.2.1 Inputs
1. A set of normalised example vectors as defined in Equation (Al):
Figure imgf000021_0003
2. Chosen values for parameters:
Figure imgf000021_0009
A.2.2 Initialisation
1. Duplicate the normalised example vectors N times and randomly shuffle to create a new sequence of vectors and corresponding indicator vectors:
Figure imgf000021_0001
2. Initialise every element of the weight map tensor to contain a random number drawn independently and identically from a gaussian distribution with a mean value zero and variance of one:
Figure imgf000021_0006
3. Set the integer variable t = 0
A.2.3 Main Algorithm
[0131] For each example vector in sequence
Figure imgf000021_0008
Figure imgf000021_0007
Calculate the update vector
Figure imgf000021_0002
4. Calculate the updated weights map such that:
Figure imgf000022_0001
5. Update the iteration counter:
Figure imgf000022_0009
A.2.4 Outputs
1. Trained map
Figure imgf000022_0011
2. Normalisation constants
Figure imgf000022_0010
A.3 Imputation Algorithm
[0132] The imputation algorithm disclosed below provides the system 500 with the ability to fill in missing or blank data measurement fields from records in the data source 501 during model validation 513 and also to predict missing or additional data measurements in production data source 519 whilst the system is in use. Generally speaking, imputation preserves all cases by replacing missing data with an estimated value based on other available information. Once all missing values have been imputed into the data records, the data set can then be analysed using standard techniques for complete data.
A.3.I Inputs
1. Trained map:
Figure imgf000022_0002
2. Normalisation constants:
Figure imgf000022_0003
3. Normalised output of the retrieval function for a set of measurements (as per Equation Al):
Figure imgf000022_0004
4. The indicator vector for the set of measurements:
Figure imgf000022_0005
Figure imgf000022_0006
A.3.2 Main Algorithm
Find the grid coordinates
Figure imgf000022_0008
of the“best-matching unit’
Figure imgf000022_0007
2. Calculate the output vector:
Figure imgf000023_0006
A.3.3 Outputs
1. Imputed vector:
Comparison to Previous Works
Figure imgf000023_0007
[0133] The training algorithm process disclosed above provides a significant improvement of existing Self-Organising Map processes by making them applicable to databases comprising random missing data values rather than, as previously known, only to be applicable in the case of complete measurement of all variables where:
Figure imgf000023_0005
[0134] Under complete measurement of all variables, by definition, every example indicator vector has the value:
Figure imgf000023_0004
[0135] Now, note that the O operator has the properties:
Figure imgf000023_0003
[0136] Under complete measurement of all variables, the normalisation constants simplify to their usual definitions since
Figure imgf000023_0001
Figure imgf000023_0002
[0137] Similarly, the observation correction factors simplify to:
Figure imgf000023_0008
[0138] Using the identity in Equation A2 the modified Euclidean distance function simplifies to the usual definition of Euclidean distance:
Figure imgf000023_0009
[0139] Hence, the update steps simplify to their usual definitions. The best-matching unit is found simply with the expression:
Figure imgf000023_0010
[0140] Finally, the update vector expression simplifies to
Figure imgf000024_0006
B. Sparse-Data Mixture Density Network (MDN) Imputation Model
[0141] In alternative particular embodiments of system 500 the AI model takes the form of a sparse-data mixture density network (MDN) imputation model as discussed below (noting that any previous mathematical definitions in this document, in particular those definitions in relation to the Sparse-Data Self-Organising Map imputation model (A) discussed above, should be disregarded in favour of the definitions below).
[0142] Although the formulation discussed here for the sparse-data MDN may be applied to any choice of parametric mixture-model, for the purposes of demonstrating a concrete embodiment this section will provide additional details on the formulation of the algorithm using the choice of a Gaussian Mixture Model (GMM).
B.1 Definitions
[0143] Let mt 6 7L+ be the number of training examples.
[0144] Let nv 6 7L+ be the number of variables per training example.
[0145] Let N be the number of training epochs.
[0146] Let the“indicator matrix’’ be defined as:
0 no recorded measurement for example i, variable j
Figure imgf000024_0001
1 measurement was recorded for example i, variable j
[0147] Let the“retrieval function” for example t, variable j be defined as:
Figure imgf000024_0002
[0148] Let the z-score normalisation constants be:
Figure imgf000024_0003
[0149] Let the z-score transformation function for variable j be:
Figure imgf000024_0004
[0150] Let a normalised example vector be:
Figure imgf000024_0005
[0151] Let be the probability that a variable is held out for use as a target output during the
Figure imgf000024_0007
training process.
[0152] Let the random input partition vector for example i be:
Figure imgf000025_0005
[0153] Let each 6j be drawn independent and identically distributed (i.i.d) from a single-trial binomial distribution:
Figure imgf000025_0006
[0154] Let the negated partition vector be:
Figure imgf000025_0001
[0155] Let the binary operator O represent element-wise vector multiplication:
[0156] Let the training inputs be:
Figure imgf000025_0007
[0157] Let the target outputs be:
Figure imgf000025_0008
[0158] Let the neural net input sentinel vector be:
Figure imgf000025_0009
[0159] Let the mini -batch size be B.
[0160] Let the Gaussian probability density function (PDF) be defined as:
Figure imgf000025_0002
[0161] Let the Gaussian cumulative density function (CDF) be defined as:
Figure imgf000025_0003
[0162] Let the softmax function be defined as:
Figure imgf000025_0004
[0163] Let the neural network inputs for example t be:
Figure imgf000025_0010
[0164] Let a mini-batch of training data be:
Figure imgf000026_0001
where 1 < a £ mt— B is the starting index of the batch.
[0165] Let a mini-batch of target outputs be:
Figure imgf000026_0002
[0166] Let the batch partition matrix and negated batch partition matrix, respectively, be:
Figure imgf000026_0003
[0167] Let c £ Z+ be the number of model mixture components.
[0168] Let np G 7L+ be the number of parameters per mixture component.
[0169] Let the overall feedforward neural network function from inputs to outputs be:
Figure imgf000026_0004
noting that fnn may define any feedforward neural network with the given input and output dimensions but nominally 4 layers of width 4096 with the“leaky ReLU” activation function is sufficient.
[0170] Let the neural network output logits be:
Figure imgf000026_0008
[0171] Let a parametric probability density function representing the chosen mixture model be:
Figure imgf000026_0005
[0172] Let a parametric probability density function representing a probabilistic prediction for batch t, variable j be:
Figure imgf000026_0006
[0173] Let the cumulative density function for the prediction for batch t, variable j be:
Figure imgf000026_0007
[0174] Let the continuous rank probability score (CRPS) of a probabilistic prediction for variable x with PDF / (x) against the true value v be:
Figure imgf000027_0001
[0175] Let the CRPS of the prediction for batch t, variable j be:
Figure imgf000027_0002
[0176] Let the variable observation fractions be:
Figure imgf000027_0003
[0177] Let the variable observation correction factors be:
Figure imgf000027_0007
[0178] Let the batch loss function be:
Figure imgf000027_0004
[0179] For the case where the chosen mixture model is a GMM, let and the neural network output
Figure imgf000027_0008
logits be:
Figure imgf000027_0005
[0180] For the case where the chosen mixture model is a GMM, let eff > 0 be a small constant setting the minimum standard deviation on a mixture component to avoid issues caused by limited numeric precision.
[0181] For the case where the chosen mixture model is a GMM, let the output predictions be:
Figure imgf000027_0006
[0182] For the case where the chosen mixture model is a GMM, let:
Figure imgf000028_0001
[0183] For the case where the chosen mixture model is a GMM, note that:
Figure imgf000028_0002
[0184] For the case where the chosen mixture model is a GMM, note that there exists a closed form solution:
Figure imgf000028_0003
where
Figure imgf000028_0005
B.2 Training Algorithm
B.2.1 Inputs
1. A partially observed set of measurements.
Figure imgf000028_0007
2. Chosen values for parameters:
Figure imgf000028_0006
B.2.2 Initialisation
1. Duplicate the normalised example vectors N times and randomly shuffle to create a new sequence of vectors and corresponding indicator vectors:
Figure imgf000028_0004
2. Initialise the neural network weights and biases using the desired technique (nominally the Glorot uniform initializer performs well).
3. Initialize the optimizer algorithm (nominally“Adam" is a good general choice)
B.2.3 Main Algorithm
[0185] For each batch xbatciu xbatch_target taken in sequence from
Figure imgf000028_0008
1. Calculate loss(xbatch,xbatch target)
2. Using automatic differentiation calculate the loss gradient with respect to each neural network weight:
Figure imgf000029_0001
3. Update the neural network using backpropagation with the chosen optimizer algorithm.
B.2.4 Outputs
1. Trained neural network.
2. Normalisation constants:
Figure imgf000029_0003
B.3 Imputation Algorithm
[0186] The imputation algorithm disclosed below provides the system 500 with the ability to fill in missing or blank data measurement fields from records in the data source 501 during model validation 513 and also to predict missing or additional data measurements in production data source 519 whilst the system is in use. Generally speaking, imputation preserves all cases by replacing missing data with an estimated value based on other available information. In the algorithm below, the estimated values for missing data are provided as probability distributions of likely values.
B.3 A Inputs
1. Trained neural network.
2. Normalisation constants:
Figure imgf000029_0002
3. Normalised output of the retrieval function for a set of measurements (as provided equation (B3) in Paragraph [0150] (above):
Figure imgf000029_0004
4. The indicator matrix for the batch of measurements:
Figure imgf000029_0005
R.3.2 Main Algorithm
1. Redefine the partition vector at inference time to be a constant:
Figure imgf000029_0006
2. Build Xbatch using x and I as per the definition provided in equation (B5) in Paragraph [0164] above.
3. Run forward inference on the neural network to derive probabilistic imputations:
Figure imgf000030_0004
B.3.3 Outputs
1. Imputed density functions:
Figure imgf000030_0001
[0187] The classification algorithm disclosed below provides the system 500 with the ability to classify the likely severity of disease in a patient based on the imputed density function for a given measurement and established clinical thresholds for that measurement, from records in the data source 501 during model validation 513 and from measurements in production data source 519 whilst the system is in use.
B.4 Classification Algorithm
B.4.1 Inputs
1. Imputed density functions:
Figure imgf000030_0002
2. A choice of threshold variable j.
3. A set of T— 1 clinical thresholds for variable j defining T gradings of disease severity (e.g. normal, mild, moderate, severe):
Figure imgf000030_0005
B.4.2 Main Algorithm
[0188] For each set of measurements at index t:
1. Calculate the probability of each severity grading according to:
Figure imgf000030_0007
Figure imgf000030_0003
B.4.3 Outputs
1. Calculated probabilities:
Figure imgf000030_0006
EXAMPLES
[0189] An example of a dataset that replicates a sparsely populated data source 501 is shown in Table 1 below. Table 1
Example Sparsely Populated Data Source
Figure imgf000031_0001
[0190] Using data source of Table 1 as a training set 503, the system is able to develop a model of a cube which is constrained by the fact that width=height=depth. Once the model is learnt then predictions can be made on similarly sparsely populated data sets. For example, if the width is known then both the height and depth of the cube can be predicted. This simple example demonstrates how the blank data fields of the cube records are filled in the original sparse dataset i.e. see Table 2 below comprising the initial data set (normal font typeset) and predicted measurements (boldface typeset) output from the trained AI model.
Table 2
Example Sparsely Populated Data Source With Imputed Data
Figure imgf000031_0002
[0191] In the embodiments described herein, the dataset of a cube (Table 1) is substituted by the NED A database which allows a model (phenotype) of a heart to be built across several variables. This phenotype model data describes the interrelation of variables across many different heart configurations and diseases.
[0192] Model validation 513 makes use of the test data 505 to evaluate the performance of the trained AI model 511 on new, previously unseen data. The outputs of the validation process 513 largely depend on the particular application and the data source 501 used which are tested against particular performance metrics 515 to evaluate the predictive power (e.g. the probability that a predicted data measurement will be within defined tolerance levels as compared with the actual measurement data). An example of a useful performance metric 515 is Root-Mean-Square Error of predicted measurements compared with known data in the test data 505.
[0193] Once the trained model 511 is validated to be able to provide meaningful measurement predictions within pre -defined tolerance levels (defined according to typical measurement error ranges taken from validation studies plus a tolerance of about 5%) the trained model 511 is imported to the primary predictive engine 517 of system 500. The prediction engine 517 is then used to analyse new data e.g. production data set 519, during system operation. For example, e.g. production data set 519, may be a database of measurements in an echo reporting software package, or it may be measurements on an echo/ultrasound workstation obtained during working of methods 200 or 300 disclosed above. The outputs from the AI prediction engine 517 comprise predicted measurements 521 and/or predicted diagnoses 523 with respect to the production data source 519, for example, obtained during working of methods 200 or 300 disclosed above. As will be readily appreciated, the outputs of the AI prediction engine 517 may be a set of predictions for measurements 521 that were not provided as inputs or alternatively a predicted diagnosis on the basis of a set of pre-defined risk factors for various diseases relevant to the nature of the data source, for example where the data source comprises echocardiography measurements, the AI prediction engine 517 may provide predictions of a particular patient’s probability of possessing, or likely to subsequently possess, heart-related diseases such as, for example arterial stenosis or heart chamber or valve malfunction.
[0194] AI prediction engine 517 may also utilise a tailored implementation of the AI algorithm to produce predictions in the required operating (e.g. software) environment. For example, an implementation for an echo software package would likely be implemented to run on Microsoft Windows, whereas an implementation for an echo workstation would need to be tailored to the vendor-specific hardware and operating system running on the workstation.
[0195] A further example embodiment of the systems disclosed above may be implemented as an application which is made available to, for example, healthcare professionals, over a communications network such as, for example, a diagnostic clinic intranet network or a publicly accessible communications network such as the internet. The network-accessible application may be provided as an interactive web application with a simple workflow, such as depicted in an example wireframe schematic representation 600 shown in Figure 6. In an exemplary workflow embodiment of application 600 in use, a user enters a set of echo measurements (or alternatively connects the input to a database of echo measurements e.g. for a plurality of patients) into an input interface 601 of web interface 610 (where the application 600 is accessible over the worldwide web/internet). The backend of application 600 (not shown) is connected to a AI system such as system 500 depicted in Figure 5. Application 600 feeds the inputted user data to the backend analysis system so as to return outputs comprising predicted measurement data 620 to fill in any blank or missing data measurements from the input data and further to output predicted disease risk factor 630. Predicted measurements and disease risk factors are then presented to the user via interface 610 to a user display means 640, which may present the output predictions in any useful manner for interpretation by the user, for example the outputs may be presented in a graphical form for easy interpretation by the user.
[0196] The methods of training and operating an AI-assisted echocardiography system as disclosed herein (e.g. methods 200, 300, 400 and 500 depicted in Figures 2, 3, 4 and 5 respectively may be implemented using a computer system 700, such as that shown in Figure 7 wherein the processes of Figures 2 to 5 may be implemented as software, such as one or more application programs executable within the computing device 700. In particular, the steps of method(s) 200, 300, 400 and 500 are effected by instructions in the software that are carried out within the computer system 700. The instructions may be formed as one or more code modules, each for performing one or more particular tasks. The software may also be divided into two separate parts, in which a first part and the corresponding code modules performs the described methods and a second part and the corresponding code modules manage a user interface between the first part and the user. The software may be stored in a computer readable medium, including the storage devices described below, for example. The software is loaded into the computer system 700 from the computer readable medium, and then executed by the computer system 700. A computer readable medium having such software or computer program recorded on it is a computer program product. The use of the computer program product in the computer system 700 preferably effects an advantageous apparatus for AI assisted echocardiography.
[0197] With reference to Figure 7, an exemplary computing device 700 is illustrated. The exemplary computing device 700 can include, but is not limited to, one or more central processing units (CPUs) 701 comprising one or more processors 702, a system memory 703, and a system bus 704 that couples various system components including the system memory 703 to the processing unit 701. The system bus 704 may be any of several types of bus structures including a memory bus or memory controller, a peripheral bus, and a local bus using any of a variety of bus architectures.
[0198] The computing device 700 also typically includes computer readable media, which can include any available media that can be accessed by computing device 700 and includes both volatile and non-volatile media and removable and non-removable media. By way of example, and not limitation, computer readable media may comprise computer storage media and communication media. Computer storage media includes media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by the computing device 700. Communication media typically embodies computer readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media. By way of example, and not limitation, communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media. Combinations of the any of the above should also be included within the scope of computer readable media.
[0199] The system memory 703 includes computer storage media in the form of volatile and/or non-volatile memory such as read only memory (ROM) 705 and random-access memory (RAM) 706. A basic input/output system 707 (BIOS), containing the basic routines that help to transfer information between elements within computing device 700, such as during start-up, is typically stored in ROM 705. RAM 706- typically contains data and/or program modules that are immediately accessible to and/or presently being operated on by processing unit 701. By way of example, and not limitation, Figure 7 illustrates an operating system 708, other program modules 709, and program data 710.
[0200] The computing device 700 may also include other removable/non-removable, volatile/non-volatile computer storage media. By way of example only, Figure 7 illustrates a hard disk drive 711 that reads from or writes to non-removable, non-volatile magnetic media. Other removable/non-removable, volatile/non-volatile computer storage media that can be used with the exemplary computing device include, but are not limited to, magnetic tape cassettes, flash memory cards, digital versatile disks, digital video tape, solid state RAM, solid state ROM, and the like. The hard disk drive 711 is typically connected to the system bus 704 through a non-removable memory interface such as interface 712.
[0201] The drives and their associated computer storage media discussed above and illustrated in Figure 7, provide storage of computer readable instructions, data structures, program modules and other data for the computing device 700. In Figure 7, for example, hard disk drive 711 is illustrated as storing an operating system 713, other program modules 714, and program data 715. Note that these components can either be the same as or different from operating system 708, other program modules 709 and program data 710. Operating system 3013, other program modules 714 and program data 715 are given different numbers hereto illustrate that, at a minimum, they are different copies.
[0202] The computing device also includes one or more input/output (I/O) interfaces 730 connected to the system bus 704 including an audio-video interface that couples to output devices including one or more of a video display 734 and loudspeakers 735. Input/output interface(s) 730 also couple(s) to one or more input devices including, for example a mouse 731, keyboard 732 or touch sensitive device 733 such as for example a smartphone or tablet device. In the embodiments disclosed herein, input interface 730 may also comprise an echocardiography/ultrasound hand piece and computing device 700 may comprise or be integrated with an echo/ultrasound workstation.
[0203] Of relevance to the descriptions below, the computing device 700 may operate in a networked environment using logical connections to one or more remote computers. For simplicity of illustration, the computing device 700 is shown in Figure 7 to be connected to a network 720 that is not limited to any particular network or networking protocols, but which may include, for example Ethernet, Bluetooth or IEEE 802.X wireless protocols. The logical connection depicted in Figure 7 is a general network connection 721 that can be a local area network (LAN), a wide area network (WAN) or other network, for example, the internet. The computing device 700 is connected to the general network connection 721 through a network interface or adapter 722 which is, in turn, connected to the system bus 704. In a networked environment, program modules depicted relative to the computing device 700, or portions or peripherals thereof, may be stored in the memory of one or more other computing devices that are communicatively coupled to the computing device 700 through the general network connection 721. It will be appreciated that the network connections shown are exemplary and other means of establishing a communications link between computing devices may be used.
[0204] All analyses discussed herein are derived from echocardiograms performed from April 11, 2000 to June 13, 2017. They involved 530,871 echocardiograms from 331,344 adults aged > 18 years (52% male) with follow up mortality data up until 30/10/2017 (study census).
Example 1— Aortic Stenosis Diagnosis
[0205] The following examples demonstrate the utility of the AI-assisted echocardiography methods and systems disclosed above. The methods are used to predict the incidence of aortic stenosis while completely removing the need for left ventricular outflow tract measurements by the use of artificial intelligence. It is observed that the above disclosed systems and methods improve the consistency of echo, in addition to saving a significant amount of scanning time for the sonographer performing echocardiography studies.
[0206] Comprehensive evaluation of the aortic valve is a standard part of every echo examination, requiring measurements performed from multiple echo windows and the use of 2-dimensional measurements and spectral Doppler. Measurement of aortic velocities using continuous wave Doppler is accurate and reproducible, but the same measurements in the LVOT are prone to error. Any error in 2D measurement of the LVOT is magnified by multiplying and squaring the measurement as part of the continuity equation (CE). The time required for aortic valve area calculation is approximately 7 minutes per patient. [0207] The goal of the models is to produce a comprehensive echo interpretation system using artificial intelligence to provide efficient, fast, reproducible echo examinations with accurate and reliable interpretation. Within this larger project it was evaluated whether AI could impute the aortic valve area from other echo data, without the need for any LVOT measurements with the aim of producing a system just as accurate as the traditional aortic valve area calculation, but more reproducible and faster with less images and measurements.
Guidelines
[0208] The current clinically accepted guidelines for diagnosis of severe aortic stenosis (AS) in a patient recommend that the following three main criteria are used:
1. AS Jet Velocity > 4.0 m/s
2. AV Mean Gradient > 40 mmHg
3. AV Area < lcm2
[0209] Due to the complex geometry of the aortic valve, rather than directly measuring its cross-sectional area, an estimate of its effective orifice area is calculated using the following equations:
LVOT Area = p x (½ x LVOT Diameter)2
[0210] Where LVOT diameter is a circular approximation of cross-sectional area measured by the sonographer during an echocardiogram study of a patient used in evaluation of the continuity equation (CE) above.
Difficulties in Diagnosing AS
[0211] It is important to note that AS Jet Velocity and AV Mean Gradient are dependent on flow rate and so in patients with low flow, a low value for these measurements does not rule out the possibility of Aortic Stenosis.
[0212] The estimated AV Area is relatively independent of flow and so in patients with low flow, it becomes a critical parameter. However, it is not without its own difficulties due to its dependence on the estimated LVOT Area. To quote the current guidelines:
“[..J the measurement variability for LVOT diameter ranges from 5 to 8%. When LVOT diameter is squared for the calculation of CS A, it becomes the greatest potential source of measurement error in the continuity equation”.
[0213] For example, in a patient with an LVOT diameter of 2cm, pi * (2/2)L2 = pi = 3. l4cm2 and pi * 2.1*2.1/4 = 1.1 pi = 3.45cm2, therefore LVOT VTI = 20cm and AV VTI = 40cm. The Aortic valve (AV) cross sectional area is then given by:
AV Area = LVOT Area * LVOT VTI / AV VTI
such that Al = 3.14*20/40 = l.57cm2 and A2 = 3.45 * 20/40 = l.72cm2. Now, say LVOT VTI = l5cm, AV VTI = 45cm, therefore:
Al = 3.14 / 3 = 1.05 moderate; and
A2 = 3.45 / 3 = 1.15 mild Example 1A— AS Prediction in General Population
Procedure:
1. A snapshot of the NEDA database (approx. 435,000 patients) was taken and split into a 70% training set and a 30% test set.
2. The training set was used to train an imputation model using the above -disclosed Self-Organising Map algorithm.
3. The test set was subsampled to only include patients with a complete set of measurements for:
AS Jet Velocity (A.K.A. AV Peak Velocity)
AV Mean Gradient
LVOT Diameter
LVOT VTI
AV VTI
4. The AV Area was calculated for each patient in the subsampled test set using Equation (4) above and a“ground truth” binary label was generated for each patient in the subsampled test set, labelling them as either“Severe AS” or“Not Severe AS”. The patients were labelled as“Severe AS” if and only if AV Area < 1cm2.
5. The measured values for“LVOT Diameter” and“LVOT VTP’ were deleted from each patient. In addition, the values for“LVOT Mean Velocity” and“LVOT Peak Velocity” were deleted as these are highly correlated with LVOT VTI
6. The imputation model disclosed above was used to predict values for“LVOT Diameter” and“LVOT VTI” in place of the deleted values.
7. Predicted AV Area was calculated for each patient using inputs oTPredicted LVOT DiameterPredicted LVOT VT Measured AV VTI.
8. A Receiver Operating Characteristic (ROC) plot and a Precision-Recall plot were generated to study the effectiveness of the predicted AV Area as a predictor of Severe AS at different cut-off values. The“ground truth” labels from Step 5 were used.
[0214] The measurements included in the imputation model include AR Peak Velocity; AR Pressure Half Time; AV Mean Gradient; AV Mean Velocity; AV Peak Velocity; AV Velocity Time Integral; Aorta at Sinotubular Diameter; Aorta at Sinuses Diameter; Aortic Root Diameter; Aortic Root Diameter MM; Ascending Aorta Diameter; Body Weight; Diastolic Slope; Distal Transverse Aorta Diameter; EFtext; Heart Rate; IVC Diameter Expiration; IVS Diastolic Thickness; IVS Diastolic Thickness MM; IVS Systolic Thickness MM; LA Area 4C View; LA Length 4C; LA Systolic Diameter LX; LA Systolic Diameter MM; LA Systolic Diameter Transverse; LV Diameter; LV Diastolic Area 2C; LV Diastolic Area 4C; LV Diastolic Area PSAX; LV Diastolic Diameter 4C; LV Diastolic Diameter MM; LV Diastolic Diameter PLAX; LV Diastolic Length 2C; LV Diastolic Length 4C; LV E’ Lateral Velocity; LV Epi Diastolic Area PSAX; LV Mass(C)d; LV Relative Wall Thickness; LV Systolic Area 2C; LV Systolic Area 4C; LV Systolic Diameter 4C; LV Systolic Diameter Base LX; LV Systolic Diameter MM; LV Systolic Diameter PLAX; LV Systolic Length 2C; LV Systolic Length 4C; LVOT Diameter; LVOT Mean Velocity; LVOT Peak Velocity; LVOT Velocity Time Integral; LVPW Diastolic Thickness; LVPW Diastolic Thickness MM; LVPW Systolic Thickness MM; MV A’ Velocity; MV Deceleration Time; MV E’ Velocity; MV Mean Velocity; MV Peak Velocity; MV Pressure Half Time; MV Velocity Time Integral; Mitral A Point Velocity; Mitral E Point Velocity; PV Peak Velocity; PV Velocity Time Integral; PatientAge; Pulm Vein Atrial Duration; Pulm Vein Atrial Reversal Velocity; Pulmonary Vein A Velocity; Pulmonary Vein Diastolic Velocity; Pulmonary Vein Systolic Velocity; RA Area; RA Systolic Diameter LX; RA Systolic Diameter Transverse; RApressuretext; RVOT Peak Velocity; Right Atrial Pressure; TR Peak Velocity; and Thoracic Aorta Diameter.
Example I A— Results
[0215] Figure 8 shows the Receiver Operating Characteristic and Precision-Recall curve area for this example. The total patients in the sample was 24748 and the diagnosis was based on the following measurements: AV Area (VTI), Holding out: [“LVOT Mean Velocity’’,“LVOT Peak Velocity’’,“LVOT Velocity Time Integral’’ , and“LVOT Diameter’’]. The number of patients predicted with severe AS was 1834 (7.410700%) and the number of patients predicted without severe was 22914 (92.589300%).
Example IB— AS Prediction in Patients with Reduced Ejection Fraction
[0216] Because of the potential the general population results may have been influenced by the larger number of patients with normal LV systolic function, and possibly masked a poorer performance in the setting of impaired LV function, the experiment above was repeated for those patients having an EF < 50% and <30% as discussed below.
Example IB— Procedure
1. Follow all instructions for Experiment 1 , except at Step 3, subsample to only include patients with Ejection fraction <= 50%.
2. Repeat Step 1 for ejection fractions of 40%, 35%, 30%.
Example IB— Results
[0217] Figure 9A shows the results of the model using an ejection fraction value of <= 50%, and Diagnosing based on: AV Area (VTI), Holding out: [“LVOT Mean Velocity’’,“LVOT Peak Velocity’’, “LVOT Velocity Time Integral’’,“LVOT Diameter’’] and removing patients with EF > 50.000000. The total number of patients in the sample was 1391. The number of patients predicted with severe AS was 143 (10.280374%) and the number of patients predicted without severe was 1248 (89.719626%)
[0218] Figure 9B shows the results of the model using an ejection fraction value of <= 40%, and diagnosing based on: AV Area (VTI), Holding out: [“LVOT Mean Velocity’’,“LVOT Peak Velocity’’, “LVOT Velocity Time Integral’’,“LVOT Diameter’’] and removing patients with EF > 40.000000. The total patients in the sample was 861. The number of patients predicted with severe AS was 96 (11.149826%) and the number of patients predicted without severe AS was 765 (88.850174%).
[0219] Figure 9C shows the results of the model using an ejection fraction value of <= 35%, and diagnosing based on: AV Area (VTI), Holding out: [“LVOT Mean Velocity’’,“LVOT Peak Velocity’’, “LVOT Velocity Time Integral’’,“LVOT Diameter’’] and removing patients with EF > 35.000000. The number of patients predicted with severe AS was 58 (11.026616%) and the number of patients predicted without severe AS was 468 (88.973384%).
[0220] Figure 9D shows the results of the model using an ejection fraction value of <= 30%, and diagnosing based on: AV Area (VTI), Holding out: [“LVOT Mean Velocity’’, “LVOT Peak Velocity’’, “LVOT Velocity Time Integral’’,“LVOT Diameter’’] and removing patients with EF > 30.000000. The total patients in the sample was 426. The number of patients predicted with severe AS was 45 (10.563380%) and the number of patients predicted without severe AS was 381 (89.436620%).
Example 1— Results Summary
[0221] Table 3 below summarises the results of Examples 1A and 1B above.
Table 3
Result Summary— Prediction of Aortic Stenosis
Figure imgf000038_0001
Example 1— Discussion
[0222] Average precision is generally a more informative metric than AUC for evaluating binary classifier performance with imbalanced data sets (see Appendix A for explanation). With this in mind, the results are interpreted as follows.
[0223] From Table 3, it can be seen that Severe AS can be predicted in the general population with AUC=0.96 and average precision of 73% without measurements for LVOT Diameter or LVOT VTI. In the subpopulation of patients with ejection fraction <= 50%, the average precision improves to 80% and the other subpopulations of patients with reduced EF have similar results. This is particularly significant, given that in these subpopulations the imputation model cannot necessarily rely on simple associations with elevated AV Mean Gradient / AV Peak Velocity. Future work will investigate which variables are the major contributors to the surprising performance of the imputation model in this setting.
[0224] Some important points to note:
The sample sizes for the reduced-EF subpopulations are an order of magnitude smaller than the general population. Hence, there is more uncertainty associated with the resulting AUC and avg. precision statistics. However, a sample size of 1391 patients is still reasonably large. The prevalence of Severe AS is higher in the reduced-EF subpopulations. Precision-Recall is relatively insensitive to a shift in class distribution so the average precisions may be meaningfully compared. But more investigation is still required to determine the exact reasons for the improved performance on this subpopulation. For example, it could be that the “decision boundary’’ in this subpopulation is more pronounced due to more advanced progression of disease.
Evaluation of AS severity in severe EV dysfunction is notoriously difficult because of low flows and resultant low gradients. It was found that in patients with EF<50% the ROC area was 0.96, P-R area 0.80 (in 1391 studies, 10% with severe AS). In EF<30% the ROC area was 0.85, P-R area 0.77 (in 426 studies, 10% with severe AS).
This confirmed that the model identified severe AS just as well in impaired LV function as in normal LV function.
Example 1— Comparison Between ROC and Precision-Recall Curves on Data With Imbalanced Classes
[0225] The Receiver Operating Characteristic (ROC) curve plots true-positive -rate (TPR) versus false-positive-rate. The Precision-Recall curve plots precision versus recall. The relevant equations are:
TPR = Recall = TP / (TP + FN)
FPR = FP / (FP + TN)
Precision = TP / (TP + FP)
where:
TP = True positives
FP = False positives
TN = True Negatives
FN = False Negatives
[0226] In the setting of the examples above, the metrics can be described as follows:
Recall/TPR is the percentage of patients with severe AS that the system correctly identifies.
FPR is the percentage of patients without severe AS that the system incorrectly diagnoses with severe AS
Precision is the fraction of positive diagnoses for severe AS generated by the system that are correct.
[0227] Hence, Precision-Recall is a better metric with an imbalanced test set for the following reasons:
The FPR metric is relative to the number of patients without severe AS, which in this setting is the majority of patients. Hence an FPR that seems low actually translates to a large number of false positive diagnoses.
The precision metric is relative to the number of positive diagnoses generated by the algorithm. This gives a better sense of how the system would be perceived in actual usage. For example, a precision of 80% means that if the system diagnoses a patient with severe AS then there is an 80% chance it is correct. Example 2— Severe Aortic Stenosis
[0228] Given the size and scope of available data no formal analyses of statistical power were performed. After training the AI system, the two AS severity predictions from the 30% test data set (comprehensive and limited echocardiography), along with the de -identified copy of the NEDA database, was imported into the Statistical Package for Social Sciences (SPSS) version 25.0, IBM Corporation, for comparisons with the continuity-derived aortic valve area. Statistical significance was set at a P-value <0.05. A probability cut-off was chosen for the severe AS prediction using AI for both the comprehensive and limited echocardiography, optimized for approximately equal sensitivity and specificity, and used for the remainder of the analysis. Discrete variables were presented as numbers and percentages. Between-group comparisons were assessed by Student’s t-tests, Mann Whitney U test (for non-normally distributed continuous data), and Chi-squared tests (for categorical data).
[0229] Big data from the National Echocardiography Database of Australia was used to create a modified mixture density network AI system. 530,000 echocardiograms on 331,344 adult individuals with follow-up mortality data were randomly split into 363,887 studies (70%) used to train the AI, and 155,967 studies (30%) used to verify the performance of the trained model. Two models were trained to predict severe AS without the need for left ventricular outflow tract data, one using the remainder of the echocardiographic measurements and the other using a limited data set, likely to change with aortic stenosis.
[0230] Performance of the AI system was assessed using receiver operating characteristic (ROC) curves, with the prediction of severe AS by AI against the aortic valve area calculated using the continuity equation. The area under the ROC (AUROC) curve was calculated for all patients and for those with impaired left ventricular function. All fatal events were identified, and the last recorded echocardiogram for those patients with complete follow up used to calculate actuarial 5-year mortality survival curves. Cox proportional hazard ratios adjusted for age and gender were calculated, along with further adjustments for the aortic valve area and mean aortic gradient. To establish potential differences in survival of those with AI-diagnosed severe AS compared those with severe AS using calculated aortic valve area, actuarial 5-year survival curves were examined. The phenotypic diagnosis of severe AS were compared using the AI system to that of the continuity-derived severe AS. To assess whether the results were predominantly a result of patients with classic high gradient, normal ejection fraction severe AS, three groups of patients (full system and limited data-set) were compared: 1) All patients in the test set, 2) those with an ejection fraction < 50% and 3) those with an ejection fraction < 30%. Patients without a full set of measurements available for calculation of the continuity equation were excluded from statistical analysis (n=361,067 patients without a valid left ventricular outflow tract measurements). The AI prediction of severe AS used all available data (minus the left ventricular outflow tract data), with missing measurements ignored by the AI. For the limited data set, only the measurements specified above were visible to the AI (and only if measured). The aortic valve area was then imputed by the AI system based on the remainder of the available echo measurements, followed by calculation of a probability of severe AS, as described above.
Example 2— Results
[0231] In the test set, 2382 of 32,574 individuals (7.38%) with a complete set of measurements had severe aortic stenosis, predicted by the AI system with an area under the receiver operating characteristic curve (AUCROC) of 0.97. The AUCROC was 0.95 in patients with a left ventricular ejection fraction <50%, and 0.92 in those with ejection fraction <30%. The performance of the AI was maintained using the limited data set (AUCROC 0.97, 0.94, and 0.93 respectively).
[0232] Figure 13 shows the flow 900 of analyses performed on data derived from the study cohort of 171,571 males & 158,404 females (aged 61.5+17.6 years) with a median 4.1 (IQR 2.2, 7.1 years) follow. There were no differences in the baseline characteristics between the 70% test set and the 30% training set, nor for those with a complete set of data available for aortic valve area calculation using the continuity equation (Table 4). A total of 2382/32,574 individuals (7.38% 95% Cl 7.10 to 7.67%) from the 30% test with a known continuity-derived aortic valve area had severe AS. An output probability cut-off of 0.065425 provided an identical sensitivity and specificity of 91.4% for severe AS diagnosed by AI when compared with the calculated aortic valve area.
Agreement Between the AI and Continuity-Equation-Derived Severe AS
[0233] The AUCROC curve 951 of AI-augmented diagnosis of severe AS compared with that derived from the continuity equation was 0.9696 (see Figure 14) with a positive predictive value of 45.9%. The model performed almost as well 953 in those with an ejection fraction <50% (Figure 14) with an AUCROC of 0.9462 (2308 patients, 11% with severe AS). The AI also performed 955 very well in those with an ejection fraction <30% (491 studies, 13%- many with low gradient, low-output severe AS) with an AUCROC of 0.9200 (Figure 14).
[0234] In the limited echo cohort, the AI prediction of severe AS remained robust. When applied to the test data set, the AUCROC 952 was 0.9648 (Figure 14). Consistent with the comprehensive data set, the AI performed almost as well in those with an ejection fraction <50% 954 and <30% 956 (AUCROC 0.9450 and 0.9269, respectively - Figure 14).
Long Term Survival of Severe AS Diagnosed by AI Versus Continuity Equation
[0235] Actuarial 5-year mean (+/-standard error of mean) survival for the AI diagnosis of non-severe AS was 1536.0+/-8.8 days vs 1072.5+/-23.3 days for severe AS, p< 0.00001, representing a mean survival difference of 463.5 days (Figure 15) which shows the actuarial survival curves for diagnosis of severe aortic stenosis using artificial intelligence vs a traditional continuity diagnosis. The patients were matched across the 30% test cohort. The upper line 961 in each of panels (A) and (B) represents the number of individuals at risk without severe aortic stenosis at each time period. The lower line 963 in each of panels (A) and (B) represents number of individuals at risk diagnosed with severe aortic stenosis at each time period. For continuity-derived severe AS, the mean survival was 1489.0+/-8.9 days vs 1086.0+/-31.6 days, a difference of 403 days. The non-severe AS group identified by the AI lived 47 days longer (95% Cl 23.3-71.7, p=0.02) than those without severe AS identified by the continuity equation. Alternatively, those with AI-identified severe AS had a similar lifespan to those identified with continuity-derived severe AS (13.5 days, 95% CI-90.4- 63.4 days, p=0.7). The AI diagnosis remained highly predictive of death after adjustment for age and gender (adjusted HR = 1.37, 95% Cl 1.22 - 1.54, p<0.0001), whereas the continuity diagnosis was only weakly predictive after adjustment for age and gender (HR = 1.19, 95% Cl 1.04 to 1.37, p=0.025). On an adjusted basis, the AI continued to predict risk of future death (HR = 1.34, 95% Cl 1.13 to 1.59; p=0.001), consistent with a multi-parameter approach to an AI diagnosis, not just the transaortic valve gradients. Adjusting for clinical outcome such as aortic valve replacement did not significantly change the predictive value of either the AI or continuity equation (CE) above. Example 2— Discussion
[0236] As demonstrated in Example 2 discussed above, AI can robustly augment the diagnosis of severe AS by interpreting the entire echocardiographic phenotype without reliance on left ventricular outflow tract measurements in a very large cohort of individuals subject to prolonged follow-up. Moreover, consistent with its multi-parameter approach to interpretation, an AI-augmented diagnosis of severe AS remained a significant predictor of long-term mortality even after adjustment for traditional AS severity measures. Accordingly, the purpose-built AI systems disclosed herein also introduce the first potential quality system for echocardiography by providing automatic measurement and disease predictions in real-time. These disclosed systems can provide a known statistical outcome for a defined set of measurements. Importantly, the fully trained AI system takes minimal computing power to operate and can be installed on both echocardiography machines and imaging reading software to improve diagnostic consistency in the absence of expert review.
[0237] If proven valid and reliable, AI offers an extremely useful clinical tool; particularly in low-resource settings where specialist cardiologists are scarce. Currently, the complex interactions present in AS require evaluation by a subspecialist-trained echocardiographer. Moreover, even after expert review, under-diagnosis of severe AS may occur in some individuals15. Despite well-developed quality guidelines for diagnosis of AS1,3 rigorous application may not be routinely practiced16 18 and errors may not be identified. The AI evaluated in this study consistently examines the entire echocardiography phenotype; taking into account the known pathophysiologic changes19 22 such as left ventricular diastolic and systolic dysfunction, left atrial enlargement and pulmonary hypertension23. Reliance on left ventricular outflow tract measures in the continuity equation introduces potential error24 26 and potential mis-classification of AS severity, with implications on follow-up echocardiography and timing of intervention. In contrast to current clinical practice, the AI is consistent, completely removing the need for measurement of left ventricular outflow tract dimension or velocity, relevant to both the diagnosis of AS and consistency and timing of follow-up16.
[0238] Outside of guidelines, there is no commonly accepted quality metric for clinical echocardiography27. The rigor applied to echocardiographic interpretation in clinical trials28, 29 may not be consistently applied in real-world environments, with implications on outcome and decisions for intervention30, 31. Quality improvement programs decrease error rates17, universal application requires automation and real-time feedback 6. AI is ideally suited to this task, since its analysis is consistent and phenotype -based. Critically, the AI systems disclosed herein perform equally well on a comprehensive echocardiogram as with a limited data set that takes only 10 minutes to acquire, with implications on efficiency, consistency and cost when applied in specific scenarios (such as follow-up echocardiography for known AS), but this requires further evaluation.
[0239] The AI systems disclosed herein are trained on data from NEDA, a very large echo database linked with mortality. Because of the nature and scope of NEDA, quality control against individual images was not feasible and data obtained from individual laboratories is assumed to be correct. However, systematic bias is unlikely because NEDA is sourced from different tertiary hospital laboratories across Australia rather than a single source. Although the AI accurately identified severe AS, the model needs to be tested across the whole range of aortic valve disease; noting the error for imputation of the aortic valve area was small (see Table 4), with no systematic bias and a mean imputation error (95% Cl) of-0.056 (-0.885, 0.664)cm2. The real-world nature of NEDA data involves missing data points, and a complete set of echocardiography measurements was not obtained in every patient. It is possible some patients with AS did not have sufficient measurements obtained to be included in the AI evaluation, and testing of the AI system in clinical and reference echo laboratories is needed.
[0240] The AI may identify some individuals without severe AS, but with similar cardiac phenotypic changes. However, these patients had a similar mortality trajectory to those with traditional severe AS, highlighting the capability of the AI to identify those at high risk. The systems disclosed herein have not yet been validated in populations outside of Australia, although Australia is a multicultural nation broadly representative of the world’s population, with over 300 different ancestries and 28% of the resident population born overseas. Also clinical linkage associated with the data sets used was not available for inclusion in the system validation examples discussed above. Potential contributors to the phenotypic changes such as hypertension or valve intervention may also need to be considered in the clinical context of the individual.
[0241] In summary, a prototype AI echocardiography measurement interpretation system has been developed which augments the diagnosis of severe AS by evaluating the entire phenotype. This process completely removes the need for left ventricular outflow tract measurements. The AI performs equally in normal and impaired systolic function, including severe left ventricular dysfunction and accurately predicts future mortality risk independent of AS gradients. The systems disclosed herein also provide real-time feedback during echocardiographic examinations with benefits in efficiency, consistency, and follow-up recommendations. There are important implications on study duration, cost and risk of sonographer injury. Overall, a consistent diagnosis of AS severity requires a shift away from the current practice of fully manual interpretation towards a more automated and objective process. The role of AI in these goals requires rigorous clinical evaluation, in differing patient groups, and with different diseases identified by echocardiography.
[0242] The validation examples discussed above strongly indicate that AI can augment and improve the diagnosis of severe AS and associated risk mortality predictions, by analyzing the entire echo phenotype and without the need for left ventricular outflow tract measurements. Particularly, the AI systems disclosed herein were able to predict patients at higher mortality risk due to AS, independently of aortic valve gradients. Decisions for intervention in severe aortic stenosis (AS) depend on reliable echocardiographic interpretation. The artificial intelligence (AI) systems disclosed herein are particularly designed to augment the diagnosis of severe AS using echocardiographic measurement data, thus providing a significant input into development of effective patient care outcomes by healthcare professionals.
INTERPRETATION
Bus
[0243] In the context of this document, the term“bus” and its derivatives, while being described in a preferred embodiment as being a communication bus subsystem for interconnecting various devices including by way of parallel connectivity such as Industry Standard Architecture (ISA), conventional Peripheral Component Interconnect (PCI) and the like or serial connectivity such as PCI Express (PCIe), Serial Advanced Technology Attachment (Serial AT A) and the like, should be construed broadly herein as any system for communicating data. In Accordance With
[0244] As described herein,‘in accordance with’ may also mean‘as a function of’ and is not necessarily limited to the integers specified in relation thereto.
Composite Items
[0245] As described herein,‘a computer implemented method’ should not necessarily be inferred as being performed by a single computing device such that the steps of the method may be performed by more than one cooperating computing devices.
[0246] Similarly objects as used herein such as ‘web server’, ‘server’, ‘client computing device’, ‘computer readable medium’ and the like should not necessarily be construed as being a single object, and may be implemented as a two or more objects in cooperation, such as, for example, a web server being construed as two or more web servers in a server farm cooperating to achieve a desired goal or a computer readable medium being distributed in a composite manner, such as program code being provided on a compact disk activatable by a license key downloadable from a computer network.
Database
[0247] In the context of this document, the term“database” and its derivatives may be used to describe a single database, a set of databases, a system of databases or the like. The system of databases may comprise a set of databases wherein the set of databases may be stored on a single implementation or span across multiple implementations. The term“ database” is also not limited to refer to a certain database format rather may refer to any database format. For example, database formats may include MySQL, MySQLi , XML or the like.
Wireless
[0248] The invention may be embodied using devices conforming to other network standards and for other applications, including, for example other WLAN standards and other wireless standards. Applications that can be accommodated include IEEE 802.11 wireless LANs and links, and wireless Ethernet.
[0249] In the context of this document, the term“ wireless” and its derivatives may be used to describe circuits, devices, systems, methods, techniques, communications channels, etc., that may communicate data through the use of modulated electromagnetic radiation through a non-solid medium. The term does not imply that the associated devices do not contain any wires, although in some embodiments they might not. In the context of this document, the term“wired" and its derivatives may be used to describe circuits, devices, systems, methods, techniques, communications channels, etc., that may communicate data through the use of modulated electromagnetic radiation through a solid medium. The term does not imply that the associated devices are coupled by electrically conductive wires.
Processes
[0250] Unless specifically stated otherwise, as apparent from the following discussions, it is appreciated that throughout the specification discussions utilizing terms such as “processing" , “computing", “calculating",“determining” ,“ analysing” or the like, refer to the action and/or processes of a computer or computing system, or similar electronic computing device, that manipulate and/or transform data represented as physical, such as electronic, quantities into other data similarly represented as physical quantities.
Processor
[0251] In a similar manner, the term“ processor” may refer to any device or portion of a device that processes electronic data, e.g. from registers and/or memory to transform that electronic data into other electronic data that, e.g. may be stored in registers and/or memory. A“ computer” or a“computing device” or a“computing machine” or a“computing platform” may include one or more processors.
[0252] The methodologies described herein are, in one embodiment, performable by one or more processors that accept computer-readable (also called machine-readable) code containing a set of instructions that when executed by one or more of the processors carry out at least one of the methods described herein. Any processor capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken are included. Thus, one example is a typical processing system that includes one or more processors. The processing system further may include a memory subsystem including main RAM and/or a static RAM, and/or ROM.
Computer-Readable Medium
[0253] Furthermore, a computer-readable carrier medium may form, or be included in a computer program product. A computer program product can be stored on a computer usable carrier medium, the computer program product comprising a computer readable program means for causing a processor to perform a method as described herein.
Networked or Multiple Processors
[0254] In alternative embodiments, the one or more processors operate as a standalone device or may be connected, e.g. networked to other processor(s), in a networked deployment, the one or more processors may operate in the capacity of a server or a client machine in server-client network environment, or as a peer machine in a peer-to-peer or distributed network environment. The one or more processors may form a web appliance, a network router, switch or bridge, or any machine capable of executing a set of instructions (sequential or otherwise) that specify actions to be taken by that machine.
[0255] Note that while some diagram(s) only show(s) a single processor and a single memory that carries the computer-readable code, those in the art will understand that many of the components described above are included, but not explicitly shown or described in order not to obscure the inventive aspect. For example, while only a single machine is illustrated, the term“machine” shall also be taken to include any collection of machines that individually or jointly execute a set (or multiple sets) of instructions to perform any one or more of the methodologies discussed herein.
Additional Embodiments
[0256] Thus, one embodiment of each of the methods described herein is in the form of a computer-readable carrier medium carrying a set of instructions, e.g. a computer program that are for execution on one or more processors. Thus, as will be appreciated by those skilled in the art, embodiments of the present invention may be embodied as a method, an apparatus such as a special purpose apparatus, an apparatus such as a data processing system, or a computer-readable carrier medium. The computer-readable carrier medium carries computer readable code including a set of instructions that when executed on one or more processors cause a processor or processors to implement a method. Accordingly, aspects of the present invention may take the form of a method, an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present invention may take the form of carrier medium (e.g. a computer program product on a computer-readable storage medium) carrying computer-readable program code embodied in the medium.
Implementation
[0257] It will be understood that the steps of methods discussed are performed in one embodiment by an appropriate processor (or processors) of a processing (i.e. computer) system executing instructions (computer-readable code) stored in storage. It will also be understood that the invention is not limited to any particular implementation or programming technique and that the invention may be implemented using any appropriate techniques for implementing the functionality described herein. The invention is not limited to any particular programming language or operating system.
Means for Carrying Out a Method or Function
[0258] Furthermore, some of the embodiments are described herein as a method or combination of elements of a method that can be implemented by a processor or a processor device, computer system, or by other means of carrying out the function. Thus, a processor with the necessary instructions for carrying out such a method or element of a method forms a means for carrying out the method or element of a method. Furthermore, an element described herein of an apparatus embodiment is an example of a means for carrying out the function performed by the element for the purpose of carrying out the invention.
Connected
[0259] Similarly, it is to be noticed that the term connected, when used in the claims, should not be interpreted as being limitative to direct connections only. Thus, the scope of the expression a device A connected to a device B should not be limited to devices or systems wherein an output of device A is directly connected to an input of device B. It means that there exists a path between an output of A and an input of B which may be a path including other devices or means. “ Connected” may mean that two or more elements are either in direct physical or electrical contact, or that two or more elements are not in direct contact with each other but yet still co-operate or interact with each other.
Embodiments
[0260] Reference throughout this specification to “one embodiment’’, “an embodiment’’ , “one arrangement’’ or“an arrangement’’ means that a particular feature, structure or characteristic described in connection with the embodiment/arrangement is included in at least one embodiment/arrangement of the present invention. Thus, appearances of the phrases “in one embodiment/arrangement’’ or “in an embodiment/arrangement’’ in various places throughout this specification are not necessarily all referring to the same embodiment/arrangement, but may. Furthermore, the particular features, structures or characteristics may be combined in any suitable manner, as would be apparent to one of ordinary skill in the art from this disclosure, in one or more embodiments/arrangements.
[0261] Similarly, it should be appreciated that in the above description of example embodiments/arrangements of the invention, various features of the invention are sometimes grouped together in a single embodiment/arrangement, figure, or description thereof for the purpose of streamlining the disclosure and aiding in the understanding of one or more of the various inventive aspects. This method of disclosure, however, is not to be interpreted as reflecting an intention that the claimed invention requires more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive aspects lie in less than all features of a single foregoing disclosed embodiment/arrangement. Thus, the claims following the Detailed Description are hereby expressly incorporated into this Detailed Description, with each claim standing on its own as a separate embodiment/arrangement of this invention.
[0262] Furthermore, while some embodiments/arrangements described herein include some but not other features included in other embodiments/arrangements, combinations of features of different embodiments/arrangements are meant to be within the scope of the invention, and form different embodiments/arrangements, as would be understood by those in the art. For example, in the following claims, any of the claimed embodiments/arrangements can be used in any combination.
Specific Details
[0263] In the description provided herein, numerous specific details are set forth. However, it is understood that embodiments of the invention may be practiced without these specific details. In other instances, well-known methods, structures and techniques have not been shown in detail in order not to obscure an understanding of this description.
Terminology
[0264] In describing the preferred embodiment of the invention illustrated in the drawings, specific terminology will be resorted to for the sake of clarity. However, the invention is not intended to be limited to the specific terms so selected, and it is to be understood that each specific term includes all technical equivalents which operate in a similar manner to accomplish a similar technical purpose. Terms such as “forward”,“rearward”,“radially” ,“ peripherally”,“upwardly” ,“downwardly”, and the like are used as words of convenience to provide reference points and are not to be construed as limiting terms.
Different Instances of Objects
[0265] As used herein, unless otherwise specified the use of the ordinal adjectives“first”, “second”, “third” , etc., to describe a common object, merely indicate that different instances of like objects are being referred to, and are not intended to imply that the objects so described must be in a given sequence, either temporally, spatially, in ranking, or in any other manner.
Comprising and Including
[0266] In the claims which follow and in the preceding description of the invention, except where the context requires otherwise due to express language or necessary implication, the word “ comprise” or variations such as“comprises” or“comprising” are used in an inclusive sense, i.e. to specify the presence of the stated features but not to preclude the presence or addition of further features in various embodiments of the invention.
[0267] Any one of the terms“including” or“which includes” or“that includes” as used herein is also an open term that also means “including at least” the elements/features that follow the term, but not excluding others. Thus, including is synonymous with and means comprising. Scope of Invention
[0268] Thus, while there has been described what are believed to be the preferred arrangements of the invention, those skilled in the art will recognize that other and further modifications may be made thereto without departing from the spirit of the invention, and it is intended to claim all such changes and modifications as fall within the scope of the invention. Functionality may be added or deleted from the block diagrams and operations may be interchanged among functional blocks. Steps may be added or deleted to methods described within the scope of the present invention.
[0269] Although the invention has been described with reference to specific examples, it will be appreciated by those skilled in the art that the invention may be embodied in many other forms.
Industrial Applicability
[0270] It is apparent from the above, that the arrangements described are applicable to the mobile device industries, specifically for methods and systems for distributing digital media via mobile devices.
[0271] It will be appreciated that the methods/apparatus/devices/systems described/illustrated above at least substantially provide systems and methods for AI-assisted echocardiography.
[0272] The AI-assisted echocardiography systems and methods described herein, and/or shown in the drawings, are presented by way of example only and are not limiting as to the scope of the invention. Unless otherwise specifically stated, individual aspects and components of the systems and methods may be modified, or may have been substituted therefore known equivalents, or as yet unknown substitutes such as may be developed in the future or such as may be found to be acceptable substitutes in the future. The systems and methods may also be modified for a variety of applications while remaining within the scope and spirit of the claimed invention, since the range of potential applications is great, and since it is intended that the present systems and methods be adaptable to many such variations.
REFERENCES
1. Baumgartner H, Falk V, Bax JJ, et al. 2017 ESC/EACTS Guidelines for the management of valvular heart disease. European Fleart Journal 2017;38:2739-91.
2. Badiani S, van Zalen J, Treibel TA, Bhattacharyya S, Moon JC, Lloyd G. Aortic Stenosis, a Left Ventricular Disease: Insights from Advanced Imaging. Current Cardiology Reports 20l6;l8:80-.
3. Nishimura RA, Otto CM, Bonow RO, et al. 2014 AHA/ ACC Guideline for the Management of Patients With Valvular Heart Disease: A Report of the American College of Cardiology/ American Heart Association Task Force on Practice Guidelines. Journal of the American College of Cardiology 20l4;63:e57-el85.
4. Mo Y, Penicka M, Di Gioia G, et al. Resolving Apparent Inconsistencies Between Area, Flow, and Gradient Measurements in Patients With Aortic Valve Stenosis and Preserved Left Ventricular Ejection Fraction. The American Journal Of Cardiology 2018;121 :751-7. Bennet CS, Abeya FC, Hoffman A, et al. Performance and Interpretation Training of Transthoracic Echocardiography in Resource -Limited Settings. Journal Of The American College Of Cardiology 2017;70:1940-1. Pellikka PA, Douglas PS, Miller JG, et al. American Society of Echocardiography Cardiovascular Technology and Research Summit: a roadmap for 2020. Journal Of The American Society Of Echocardiography: Official Publication Of The American Society Of Echocardiography 2013;26:325-38. Gandhi S, Mosleh W, Shen J, Chow C-M. Automation, machine learning, and artificial intelligence in echocardiography: A brave new world. Echocardiography (Mount Kisco, NY) 2018. Narula S, Shameer K, Salem Omar AM, Dudley JT, Sengupta PP. Machine-Learning Algorithms to Automate Morphological and Functional Assessments in 2D Echocardiography. Journal Of The American College Of Cardiology 2016;68:2287-95. Zhang J, Gajjala S, Agrawal P, et al. Fully Automated Echocardiogram Interpretation in Clinical Practice. Circulation 2018;138:1623-35. Bishop C. Mixture density networks. Technical Report. Birmingham, UK: Aston University; 1994. Srivastava NH, G; Krizhevsky, A; Sutskever, I; Salakhutdinov, R. Dropout: A Simple Way to Prevent Neural Networks from Overfitting. Journal of Machine Learning Research 2014;15:1929-58. Grimit EP, Gneiting T, Berrocal VJ, Johnson NA. The continuous ranked probability score for circular variables and its application to mesoscale forecast ensemble verification. Quarterly Journal of the Royal Meteorological Society 2007;132:2925-42. D’Isanto A. Uncertain Photometric Redshifts with Deep Learning Methods. Proceedings of the International Astronomical Union 2016;12:209-12. Baumgartner H, Hung J, Bermejo J, et al. Recommendations on the Echocardiographic Assessment of Aortic Valve Stenosis: A Focused Update from the European Association of Cardiovascular Imaging and the American Society of Echocardiography. Journal of the American Society of Echocardiography 2017;30:372-92. Magliano D, Liew D, Pater H, et al. Accuracy of the Australian National Death Index: comparison with adjudicated fatal outcomes among Australian participants in the Long-term Intervention with Pravastatin in Ischaemic Disease (LIPID) study. Aust N Z J Public Health 2003;27:649-53. Malouf J, Le Tourneau T, Pellikka P, et al. Aortic valve stenosis in community medical practice: determinants of outcome and implications for aortic valve replacement. The Journal Of Thoracic And Cardiovascular Surgery 2012;144:1421-7. Chan RH, Shaw JL, Hauser TH, Markson LJ, Manning WJ. Guideline Adherence for Echocardiographic Follow-Up in Outpatients with at Least Moderate Valvular Disease. Journal Of The American Society Of Echocardiography: Official Publication Of The American Society Of Echocardiography 2015;28:795-801. Fanari Z, Choudhry UI, Reddy VK, et al. The Value of Quality Improvement Process in the Detection and Correction of Common Errors in Echocardiographic Hemodynamic Parameters in a Busy Echocardiography Laboratory. Echocardiography (Mount Kisco, NY) 2015;32:1778-89. Nagueh SF, Farrell MB, Bremer ML, Dunsiger SI, Gorman BL, Tilkemeier PL. Predictors of Delayed Accreditation of Echocardiography Laboratories: An Analysis of the Inter societal Accreditation Commission Database. Journal Of The American Society Of Echocardiography: Official Publication Of The American Society Of Echocardiography 20l5;28: 1062-9. e7. Delaye J, Chevalier P, Delahaye F, Didier B. Valvular aortic stenosis and coronary atherosclerosis: pathophysiology and clinical consequences. European Heart Journal 1988 ;9 Suppl E:83-6. Pibarot P, Dumesnil JG. New concepts in valvular hemodynamics: implications for diagnosis and treatment of aortic stenosis. The Canadian Journal Of Cardiology 2007;23 Suppl B:40B-7B. Dweck MR, Boon NA, Newby DE. Calcific aortic stenosis: a disease of the valve and the myocardium. Journal Of The American College Of Cardiology 2012;60:1854-63. Mutlak D, Aronson D, Carasso S, Lessick J, Reisner SA, Agmon Y. Frequency, determinants and outcome of pulmonary hypertension in patients with aortic valve stenosis. The American Journal Of The Medical Sciences 2012;343:397-401. Bartel T, Miiller S. Preserved ejection fraction can accompany low gradient severe aortic stenosis: impact of pathophysiology on diagnostic imaging. European Heart Journal 2013;34:1862-3. Michelena HI, Margaryan E, Miller FA, et al. Inconsistent echocardiographic grading of aortic stenosis: is the left ventricular outflow tract important? Heart (British Cardiac Society) 2013;99:921-31. Oh JK, Kane GC. The Echo Manual: Lippincott Williams & Wilkins Lippincott Williams & Wilkins, Attn: Sharon Kimmel, 14700 Citicorp Dr, Bldg 3, Hagerstown, MD, 21742; 2018. Otto CM, Otto CM. The practice of clinical echocardiography: Elsevier Science Health Science Elsevier Science Health Science, Attn Customer Service, 11830 Westline Industrial Dr, Saint Louis, MO, 63146; 2016. Crowley AL, Yow E, Barnhart HX, et al. Critical Review of Current Approaches for Echocardiographic Reproducibility and Reliability Assessment in Clinical Research. Journal of the American Society of Echocardiography 20l6;29:l l44-54.e7. Douglas PS, DeCara JM, Devereux RB, et al. Echocardiographic Imaging in Clinical Trials: American Society of Echocardiography Standards for Echocardiography Core Laboratories. Journal of the American Society of Echocardiography 2009;22:755-65. Khouri MG, Ky B, Dunn G, et al. Echocardiography Core Laboratory Reproducibility of Cardiac Safety Assessments in Cardio-Oncology. Journal Of The American Society Of Echocardiography: Official Publication Of The American Society Of Echocardiography 2018;31:361-71. e3. Khouri MG, Ky B, Dunn G, et al. Echocardiography Core Laboratory Reproducibility of Cardiac Safety Assessments in Cardio-Oncology. Journal of the American Society of Echocardiography 2018;31:361-71. e3. Rigolin VH. Quality Matters. Journal of the American Society of Echocardiography 2017;30:A17-A8

Claims

THE CLAIMS DEFINING THE INVENTION ARE AS FOLLOWS:
1. A method for processing a sparsely populated data source comprising:
(a) retrieving data from a sparsely populated data source to form a base dataset, the data source comprising a plurality of records, each record comprising at least one unpopulated data field corresponding to a medical measurement;
(b) dividing the retrieved data into two portions:
a first portion comprising a training dataset being a defined percentage X% of the base dataset; and
a second portion comprising a validation dataset being a defined percentage (100%-C%) of the base dataset;
(c) analysing the training data set to jointly model variable relationships using a non-linear function approximation algorithm applied iteratively to the records of the training data set to obtain a trained model and measurement prediction protocols for populating unpopulated field in the training data set;
(d) using the measurement prediction protocols, computing prediction values for measurement data for the unpopulated data fields;
(e) imputing the predicted measurement values in the records of the training dataset;
(f) analysing the training dataset on the basis of predefined disease conditions in known patient records of the base data set to form a disease model adapted to predict a probability of a disease condition in patient records of the trained data set;
(g) imputing the predicted measurement values in the records of the validation dataset;
(h) validating the disease model comprising analysing the validation dataset using the disease model, wherein the records of the validation dataset comprise disease data associated with patient data, and determining a validation error comprising a probability of correctly predicting a patient disease state in the records of the validation set;
(i) repeating Steps (c) to (h) to minimise the validation error and computing a prediction of a probable disease state for each patient record in the base dataset.
2. A method as claimed in Claim 1, wherein said sparsely populated data source comprises a plurality of medical records comprising measurement data obtained from a medical study procedure.
3. A method as claimed in Claim 2, wherein the medical study procedure comprises an echocardiography procedure.
4. A method as claimed in claim any one of the preceding claims, wherein during a measurement procedure, unpopulated measurement data is predicted using the measurement prediction protocols on the basis of data collected by a procedure operator.
5. A method as claimed in Claim 4, wherein during a measurement procedure, a patient disease state is predicted by the disease model on the basis of data collected by a procedure operator and/or on measurement data predicted using the measurement prediction protocols.
6. A method as claimed in Claim 4 or Claim 5, wherein the unpopulated measurement data and the patient disease state prediction is computed in real-time during the measurement procedure.
7. A method as claimed in any one of the preceding claims, wherein: during a measurement procedure, measurements obtained by a procedure operator are incorporated into the training data set to form an updated data set and analysing the updated training data set using the neural network to compute updated measurement prediction protocols and/or an updated disease model; and
analysing the measurement obtained during the measurement procedure using the updated measurement prediction protocols and/or updated disease model to predict a probable disease state for a patient undergoing the measurement procedure.
8. A method according to Claim 7, wherein as a result of the probable disease state prediction, directing the measurement operator to record relevant measurement data to increase the confidence of the probable disease state prediction.
9. A method as claimed in any one of the preceding claims wherein the disease condition comprises aortic stenosis.
10. Apparatus for conducting a measurement procedure comprising:
measurement tools relevant to the measurement procedure;
means for recording measurement data;
means for transmitting measurement data to an analysis means, said analysis means comprising: measurement prediction protocols for predicting measurement data for unpopulated measurement fields; and/or
a disease model for predicting a probable disease state for a patient undergoing a measurement procedure; and
means for alerting the measurement operator of the predicted measurement data and probable disease state and for providing direction to the measurement operator for relevant measurement data to be collected on the basis of the predicted probable disease state.
11. Apparatus as claimed in Claim 10, comprising a display surface adapted to display a notification to the measurement operator comprising the predicted measurement data or a probable disease state.
12. A computer program product having a computer readable medium having a computer program recorded therein for processing a sparsely populated data source, said computer program product comprising:
computer program code means for retrieving data from a sparsely populated data source to form a base dataset, the data source comprising a plurality of records, each record comprising at least one unpopulated data field corresponding to a medical measurement;
computer program code means for dividing the retrieved data into two portions: a first portion comprising a training dataset being a defined percentage X% of the base dataset; and a second portion comprising a validation dataset being a defined percentage (100%-C%) of the base dataset;
computer program code means for analysing the training data set using a non-linear function approximation algorithm applied iteratively over the records of the training data set to obtain a trained data set and measurement prediction protocols for populating unpopulated field in the training data set;
computer program code means for using the measurement prediction protocols, computing prediction values for measurement data for the unpopulated data fields;
computer program code means for imputing the predicted measurement values in the records of the training dataset; computer program code means for analysing the training dataset on the basis of predefined disease conditions in known patient records of the base data set to form a disease model adapted to predict a probability of a disease condition in patient records of the trained data set;
computer program code means for imputing the predicted measurement values in the records of the validation dataset;
computer program code means for validating the disease model comprising analysing the validation dataset using the disease model, wherein the records of the validation dataset comprise disease data associated with patient data, and determining a validation error comprising a probability of correctly predicting a patient disease state in the records of the validation set; and
computer program code means for computing a prediction of a probable disease state for each patient record in the base dataset.
PCT/AU2019/050086 2018-02-06 2019-02-06 Systems and methods for ai-assisted echocardiography WO2019153039A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
AU2018900364A AU2018900364A0 (en) 2018-02-06 AI-Assisted Echocardiography
AU2018900364 2018-02-06

Publications (1)

Publication Number Publication Date
WO2019153039A1 true WO2019153039A1 (en) 2019-08-15

Family

ID=67547801

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/AU2019/050086 WO2019153039A1 (en) 2018-02-06 2019-02-06 Systems and methods for ai-assisted echocardiography

Country Status (1)

Country Link
WO (1) WO2019153039A1 (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111739617A (en) * 2020-06-15 2020-10-02 中国医学科学院医学信息研究所 Medical image artificial intelligence quality control marking method, device, equipment and storage medium
WO2020257592A1 (en) * 2019-06-21 2020-12-24 The Regents Of The University Of California Application of artificial intelligence on detecting canine left atrial enlargement on thoracic radiographs
WO2021097460A1 (en) * 2019-11-15 2021-05-20 Geisinger Clinic Systems and methods for a deep neural network to enhance prediction of patient endpoints using videos of the heart
WO2021122345A1 (en) * 2019-12-17 2021-06-24 Koninklijke Philips N.V. Aortic stenosis classification
CN115151182A (en) * 2020-10-10 2022-10-04 豪夫迈·罗氏有限公司 Method and system for diagnostic analysis
WO2022266651A1 (en) * 2021-06-17 2022-12-22 Cedars-Sinai Medical Center Precision phenotyping of left ventricular hypertrophy with echocardiographic deep learning
US11704803B2 (en) 2020-03-30 2023-07-18 The Board Of Trustees Of The Leland Stanford Junior University Methods and systems using video-based machine learning for beat-to-beat assessment of cardiac function
US11709910B1 (en) * 2019-03-18 2023-07-25 Cigna Intellectual Property, Inc. Systems and methods for imputing missing values in data sets
WO2024000041A1 (en) * 2022-07-01 2024-01-04 ECHOIQ Limited Systems and methods for ai-assisted echocardiography

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5769074A (en) * 1994-10-13 1998-06-23 Horus Therapeutics, Inc. Computer assisted methods for diagnosing diseases
US20110202486A1 (en) * 2009-07-21 2011-08-18 Glenn Fung Healthcare Information Technology System for Predicting Development of Cardiovascular Conditions
US20150170055A1 (en) * 2013-12-18 2015-06-18 International Business Machines Corporation Machine learning with incomplete data sets
US20160253469A1 (en) * 2008-07-25 2016-09-01 Fundação D. Anna Sommer Champalimaud e Dr. Carlos Montez Champalimaud Systems and methods for predicting favorable-risk disease for patients enrolled in active surveillance
WO2017181288A1 (en) * 2016-04-21 2017-10-26 The University Of British Columbia Echocardiographic image analysis

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5769074A (en) * 1994-10-13 1998-06-23 Horus Therapeutics, Inc. Computer assisted methods for diagnosing diseases
US20160253469A1 (en) * 2008-07-25 2016-09-01 Fundação D. Anna Sommer Champalimaud e Dr. Carlos Montez Champalimaud Systems and methods for predicting favorable-risk disease for patients enrolled in active surveillance
US20110202486A1 (en) * 2009-07-21 2011-08-18 Glenn Fung Healthcare Information Technology System for Predicting Development of Cardiovascular Conditions
US20150170055A1 (en) * 2013-12-18 2015-06-18 International Business Machines Corporation Machine learning with incomplete data sets
WO2017181288A1 (en) * 2016-04-21 2017-10-26 The University Of British Columbia Echocardiographic image analysis

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
LYON, A. ET AL.: "Computational techniques for ECG analysis and interpretation in light of their contribution to medical advances", J. R. SOC. INTERFACE, vol. 15, no. 138, 10 January 2018 (2018-01-10), pages 1 - 18, XP055629231 *

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11709910B1 (en) * 2019-03-18 2023-07-25 Cigna Intellectual Property, Inc. Systems and methods for imputing missing values in data sets
WO2020257592A1 (en) * 2019-06-21 2020-12-24 The Regents Of The University Of California Application of artificial intelligence on detecting canine left atrial enlargement on thoracic radiographs
WO2021097460A1 (en) * 2019-11-15 2021-05-20 Geisinger Clinic Systems and methods for a deep neural network to enhance prediction of patient endpoints using videos of the heart
US11864944B2 (en) 2019-11-15 2024-01-09 Geisinger Clinic Systems and methods for a deep neural network to enhance prediction of patient endpoints using videos of the heart
US11957507B2 (en) 2019-11-15 2024-04-16 Geisinger Clinic Systems and methods for a deep neural network to enhance prediction of patient endpoints using videos of the heart
WO2021122345A1 (en) * 2019-12-17 2021-06-24 Koninklijke Philips N.V. Aortic stenosis classification
US11704803B2 (en) 2020-03-30 2023-07-18 The Board Of Trustees Of The Leland Stanford Junior University Methods and systems using video-based machine learning for beat-to-beat assessment of cardiac function
CN111739617A (en) * 2020-06-15 2020-10-02 中国医学科学院医学信息研究所 Medical image artificial intelligence quality control marking method, device, equipment and storage medium
CN115151182A (en) * 2020-10-10 2022-10-04 豪夫迈·罗氏有限公司 Method and system for diagnostic analysis
CN115151182B (en) * 2020-10-10 2023-11-14 豪夫迈·罗氏有限公司 Method and system for diagnostic analysis
WO2022266651A1 (en) * 2021-06-17 2022-12-22 Cedars-Sinai Medical Center Precision phenotyping of left ventricular hypertrophy with echocardiographic deep learning
WO2024000041A1 (en) * 2022-07-01 2024-01-04 ECHOIQ Limited Systems and methods for ai-assisted echocardiography

Similar Documents

Publication Publication Date Title
WO2019153039A1 (en) Systems and methods for ai-assisted echocardiography
US20200334416A1 (en) Computer-implemented natural language understanding of medical reports
Dorado-Díaz et al. Applications of artificial intelligence in cardiology. The future is already here
Li et al. Consistency of variety of machine learning and statistical models in predicting clinical risks of individual patients: longitudinal cohort study using cardiovascular disease as exemplar
US11521716B2 (en) Computer-implemented detection and statistical analysis of errors by healthcare providers
Zheng et al. Resolving the bias in electronic medical records
US20150095069A1 (en) Algorithms to Identify Patients with Hepatocellular Carcinoma
US11587679B2 (en) Generating computer models from implicitly relevant feature sets
Abdar et al. The need for quantification of uncertainty in artificial intelligence for clinical data analysis: increasing the level of trust in the decision-making process
Skitsan et al. Evaluation of the Informative Features of Cardiac Studies Diagnostic Data using the Kullback Method.
Chadaga et al. Artificial intelligence for diagnosis of mild–moderate COVID-19 using haematological markers
Kwon et al. RMExplorer: A visual analytics approach to explore the performance and the fairness of disease risk models on population subgroups
Donnat et al. A Bayesian hierarchical network for combining heterogeneous data sources in medical diagnoses
Pellikka et al. Automated analysis of limited echocardiograms: Feasibility and relationship to outcomes in COVID-19
WO2024000041A1 (en) Systems and methods for ai-assisted echocardiography
Meng et al. Hierarchical continuous-time inhomogeneous hidden Markov model for cancer screening with extensive followup data
Tabarisaadi et al. An optimized uncertainty-aware training framework for neural networks
Cheng et al. Revealing Unforeseen Diagnostic Image Features With Deep Learning by Detecting Cardiovascular Diseases From Apical 4‐Chamber Ultrasounds
Semakova et al. Data-driven identification of hypertensive patient profiles for patient population simulation
EP4076153A1 (en) Aortic stenosis classification
JP2021507392A (en) Learning and applying contextual similarities between entities
Vallon et al. Patient-Level Clinical Expertise Enhances Prostate Cancer Recurrence Predictions with Machine Learning
Shamshuzzoha et al. Early Prediction Model of Macrosomia Using Machine Learning for Clinical Decision Support
Silva et al. Predicting Diabetes Mellitus Using Machine Learning and Optical Character Recognition
Bleich et al. Using random forests with asymmetric costs to predict hospital readmissions

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19750737

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 19/01/2021)

122 Ep: pct application non-entry in european phase

Ref document number: 19750737

Country of ref document: EP

Kind code of ref document: A1