US11423538B2 - Computer-implemented machine learning for detection and statistical analysis of errors by healthcare providers - Google Patents
Computer-implemented machine learning for detection and statistical analysis of errors by healthcare providers Download PDFInfo
- Publication number
- US11423538B2 US11423538B2 US16/849,442 US202016849442A US11423538B2 US 11423538 B2 US11423538 B2 US 11423538B2 US 202016849442 A US202016849442 A US 202016849442A US 11423538 B2 US11423538 B2 US 11423538B2
- Authority
- US
- United States
- Prior art keywords
- training
- encoder
- diagnostic
- network
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/20—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/251—Fusion techniques of input or preprocessed data
-
- G06K9/6256—
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
- G06N3/0442—Recurrent networks, e.g. Hopfield networks characterised by memory or gating, e.g. long short-term memory [LSTM] or gated recurrent units [GRU]
-
- G06N3/0445—
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
- G06N3/0455—Auto-encoder networks; Encoder-decoder networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0475—Generative networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/09—Supervised learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N7/00—Computing arrangements based on specific mathematical models
- G06N7/01—Probabilistic graphical models, e.g. probabilistic networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/80—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
- G06V10/803—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of input or preprocessed data
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/98—Detection or correction of errors, e.g. by rescanning the pattern or by human intervention; Evaluation of the quality of the acquired patterns
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/98—Detection or correction of errors, e.g. by rescanning the pattern or by human intervention; Evaluation of the quality of the acquired patterns
- G06V10/993—Evaluation of the quality of the acquired pattern
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
- G06V30/12—Detection or correction of errors, e.g. by rescanning the pattern
- G06V30/133—Evaluation of quality of the acquired characters
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
- G06V30/19—Recognition using electronic means
- G06V30/191—Design or setup of recognition systems or techniques; Extraction of features in feature space; Clustering techniques; Blind source separation
- G06V30/19147—Obtaining sets of training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
- G06V30/19—Recognition using electronic means
- G06V30/191—Design or setup of recognition systems or techniques; Extraction of features in feature space; Clustering techniques; Blind source separation
- G06V30/1916—Validation; Performance evaluation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
- G06V30/19—Recognition using electronic means
- G06V30/191—Design or setup of recognition systems or techniques; Extraction of features in feature space; Clustering techniques; Blind source separation
- G06V30/19173—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
- G06V30/32—Digital ink
- G06V30/333—Preprocessing; Feature extraction
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/40—Document-oriented image-based pattern recognition
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H30/00—ICT specially adapted for the handling or processing of medical images
- G16H30/40—ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- G06K9/6215—
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/03—Recognition of patterns in medical or anatomical images
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H30/00—ICT specially adapted for the handling or processing of medical images
- G16H30/20—ICT specially adapted for the handling or processing of medical images for handling medical images, e.g. DICOM, HL7 or PACS
Definitions
- the present disclosure relates generally to computer-implemented machine learning systems and methods that are programmed to classify digital image data alone or in combination with unstructured text data, and more specifically pertains to machine learning systems and methods for diagnostic error detection.
- Radiology is one example of a field in which images of patient anatomy, and dictated records of assessment by radiologists, often serve as core records reflecting a diagnosis.
- the interpretation of digital images is often complex, requiring significant medical and anatomical knowledge as well as an ability to detect subtle or complicated patterns of information in the correct context, and therefore the radiology field has a non-zero error rate, in which patients have had their diagnostic image data interpreted incorrectly, leading to the wrong diagnosis.
- the result can have a significant impact on patient comfort, care patterns, treatment outcomes and costs. For example, an erroneous diagnosis could lead to preparation for or performance of a surgical procedure that is unnecessary.
- diagnostic errors result from deficiencies in a radiologist's skill in interpreting image data, other diagnostic errors result from differences in the communication of diagnostic information in written or dictated diagnostic reports. It is commonplace for different radiology practitioners to express a diagnosis in multiple different ways in writing, or with arcane or incorrect terms; some of these variations will correctly express a patient's diagnosis and many will convey an erroneous or misleading diagnosis.
- diagnostic errors and quality issues occur with varying prevalence rates in patient exams.
- categories of diagnostic errors include: (1) false positive reporting of a diagnostic finding, (2) false negative reporting of a diagnostic finding, (3) errors in which a finding is “overcalled” or graded as being overly severe, or (4) errors in which a finding is “undercalled” or graded as being too minor.
- Other quality issues, related to communication issues in the report can include the following categories: (1) findings that are reported in an overly equivocal manner, (2) findings that are reported in an overly vague manner, (3) findings that are reported with inappropriate emphasis, (4) inappropriate or lack of comparisons with prior diagnostic studies, (5) inappropriate or lack of inclusion of relevant standard measures (e.g.
- diagnostic radiology exams can also suffer from technical errors and quality issues that can include: (1) poor image quality (e.g. low signal-to-noise ratio), (2) images degraded or obscured by patient motion or other artifacts, (3) poorly configured exam protocols (e.g. an MRI exam conducted without collecting images that have a necessary image contrast setting or images collected with resolution that is too low), or (4) poor anatomical coverage of the images.
- image quality e.g. low signal-to-noise ratio
- images degraded or obscured by patient motion or other artifacts e.g. an MRI exam conducted without collecting images that have a necessary image contrast setting or images collected with resolution that is too low
- poorly configured exam protocols e.g. an MRI exam conducted without collecting images that have a necessary image contrast setting or images collected with resolution that is too low
- diagnostic accuracy and quality assessment system Since a multitude of diagnostic errors and related quality issues are possible in the context of most diagnostic imaging exams, it can be valuable to prioritize the specific types of diagnostic findings and diagnostic errors that a diagnostic accuracy and quality assessment system will target for evaluation.
- One approach to prioritization is to identify general aspects of diagnoses that are clinically meaningful for patients' care patterns and/or outcomes and achieve high degrees of agreement between radiologist. Since perfect agreement between radiologists is not likely in any category of diagnostic finding or diagnostic error, and the levels of agreement exhibit a wide variability across categories of diagnostic findings and errors, is can be valuable for a diagnostic accuracy and quality assessment system to be able to appropriately quantify the amount of agreement that radiologists exhibit in each category of diagnostic finding and error under evaluation.
- diagnostic accuracy and quality assessment system When using a diagnostic accuracy and quality assessment system to evaluate multiple distinct providers, it is critical to account for the fact that different providers often care for very different patient populations. It may be inappropriate to use unadjusted estimates of diagnostic accuracy rates or error rates as standardized and generalizable measures of radiology care quality. A quality assessment system that can be used across a diverse population of providers will usually need to include some adjustment for differences between the relevant patient populations.
- Computer-implemented image interpretation and medical report interpretation technologies have not been developed, expanded, or adapted for use as part of a diagnostic accuracy and quality assessment system. The technical performance and design requirements for these technologies are different in this distinct application domain.
- a computer-implemented image interpretation system will need to achieve high sensitivity, high specificity, and an ability to target a wide range of diagnostic finding types.
- the method comprises, for each given training data pair of a plurality of training data pairs, where each given training data pair comprises at least a training text derived from a radiological report and a training image derived from a radiological exam image associated with the radiological report, training a diagnostic quality assessment machine learning network by: determining, using a first encoder network, word embeddings for the training text; generating, using a concept generator coupled to one or more layers of the first encoder network, a generated concept based on the operation of the one or more layers in determining the word embeddings; regularizing the first encoder network by calculating a first loss between the generated concept and a labeled concept for the training text; determining, using a second encoder network, features for the training image; generating, using a heatmap generator coupled to one or more layers of the second encoder network, a generated heatmap based on the operation of the one or more layers in determining the
- the training text is a section of text obtained from a radiological report, wherein the section of text corresponds to an identified anatomical region or pathological feature discussed in the radiological report.
- the training image is a section obtained from a sequence of one or more radiological exam images from which the radiological report was prepared.
- the training text and the training image are associated with the same anatomical region or pathological feature.
- the same anatomical region or pathological feature is a motion segment of the lumbar spine.
- one or more of the plurality of training data pairs are obtained from a database of structured checklists corresponding to medical diagnostic data, the medical diagnostic data including radiological reports and radiological exam images.
- the first encoder network is configured as a recurrent neural network, an ordered neuron LSTM (Long short-term memory), or a Transformer based model trained specifically on a corpus of radiology report text.
- LSTM Long short-term memory
- Transformer based model trained specifically on a corpus of radiology report text.
- the labeled concept for a given training text includes an indication of one or more of: an identified pathology, a location of the identified pathology, and a severity of the identified pathology, as contained within the given training text.
- the second encoder network is a densely connected convolutional neural network (DenseNet) or a residual neural network (ResNet) adapted to the anisotropy and intensity distribution of radiology exam images.
- DenseNet densely connected convolutional neural network
- ResNet residual neural network
- the generated heatmap is an attention heatmap determined from the one or more layers of the second encoder network while the second encoder network generates features for the training image; and the labeled heatmap is an annotation corresponding to one or more anatomical features or pathological features as located within the training image.
- the heatmap generator comprises a decoder for performing a specific segmentation of the training image; and the labeled heatmap is an annotated segmentation corresponding to one or more anatomical features or pathological features as located within the training image.
- the determined diagnostic quality category is selected from a set of diagnostic quality categories including ‘Agree’, ‘Overcall’, ‘Undercall’, and ‘Missed’.
- training the diagnostic quality assessment machine learning network on the given training data pair further comprises: regularizing the first encoder network by minimizing a first BCE (binary cross entropy) loss between a labeled pathology for the training text and a generated pathology for the training text, the generated text pathology output by an NLP (natural language processing) pathology classifier over the word embeddings of the first encoder network; regularizing the second encoder network by minimizing a second BCE loss between a labeled pathology for the training image and a generated pathology for the training image, the generated image pathology output by an image pathology classifier over the features of the second encoder network; and the total loss function further comprises the first BCE loss and the second BCE loss.
- first BCE binary cross entropy
- the labeled pathology for the training text is ground-truth pathology information contained within the training text, independent from its specific textual expression; and the labeled pathology for the training image is ground-truth pathology information present in the training image, wherein the ground-truth pathology information for a given training image is determined as a consensus obtained from one or more expert reviews of the given training image.
- the labeled pathology for the training image is generated automatically based on accessing one or more structured checklists generated in response to receiving a user input representing of the one or more expert reviews of the given training image.
- training the diagnostic quality assessment machine learning network on the given training data pair further comprises: providing, to a Siamese function, an input comprising the word embeddings determined for the training text by the first encoder network and the image features determined for the training image by the second encoder network; calculating, using the Siamese function, a Siamese distance between the word embeddings and the image features; calculating, using a Siamese error encoder, a Siamese loss between the Siamese distance and a Siamese label, the Siamese label indicating an extent to which the training text and training image of the given training data pair agree or disagree; and minimizing the Siamese loss to increase a distance between training text and training images that disagree and to decrease a distance between training text and training images that agree.
- the Siamese loss is a multi-task loss; the error encoder classifies the given training data pair into the determined diagnostic quality category based at least in part on the Siamese distance output by the Siamese function; and the total loss function for the given training data pair further includes the Siamese loss.
- the Siamese error encoder acts as a controller to the error encoder by causing the error encoder to regress to an estimated diagnostic error on the basis of the Siamese distance between the word embeddings and the image features.
- the method further comprises providing at least the determined diagnostic error from the error encoder, the word embeddings from the first encoder network, and the image features from the second encoder network, to a clinical significance encoder; and regressing, using the clinical significance encoder, to an estimated clinical significance of the determined diagnostic error, wherein the clinical significance encoder is configured as a regressor network having a sigmoid activation function.
- the method further comprises providing one or more clinical references to a clinical controller of the diagnostic quality assessment machine learning network, the clinical references including one or more of patient age, patient weight, and patient history of previous related pathologies; and generating, from the one or more clinical references and via the clinical controller, a feature vector to control the second encoder network.
- FIG. 1 illustrates an example of functional elements and data flows in a distributed computer system that may be used to implement one embodiment of provider assessment processing
- FIG. 2 illustrates further details of the statistical modeling logic of FIG. 1 ;
- FIG. 3 illustrates an example data assessment process that may be used in an embodiment
- FIGS. 4A-B illustrate an example flowchart of a pre-processing pipeline for input radiological images and/or input radiological reports
- FIG. 5 illustrates an example architecture diagram for a multi-regularizer machine learning network to detect diagnostic errors in radiological examinations
- FIG. 6A illustrates an example architecture diagram for a Siamese-like machine learning network to detect diagnostic errors in radiological examinations
- FIG. 6B illustrates an example architecture diagram for an additional Siamese-like machine learning network to detect diagnostic errors in radiological examinations
- FIG. 7 illustrates an example architecture diagram for a Siamese-like machine learning network that is extended to regress to an estimated clinical significance of error in addition to an estimation of diagnostic error;
- FIG. 8 illustrates an example computer system, with non-transitory computer-readable storage media, that may be used to implement all or part of one or more aspects of the present disclosure
- FIG. 9 illustrates a plate notation for a Bayesian approach to radiology quality scoring with AI and/or human QA data.
- Machine learning is capable of analyzing tremendously large data sets at a scale that continues to increase; however, the ability to build and otherwise curate appropriately large training data sets has lagged and continues to be a major bottleneck in implementing flexible or real-time machine learning systems.
- Section 7 provides a general overview of an example machine learning network for diagnostic quality assessment.
- Section 8 provides architecture and training details of the example machine learning network for diagnostic quality assessment.
- a system for quantifying diagnostic radiology errors uses structured and standardized exam reviews that are performed by independent radiologists to create a repository of clinically meaningful attributes of radiology images and radiology reports. Digital analysis of the attributes yields an objective truth source for any diagnosis that can be associated with digital images of anatomy or other physical features of the subject as well as an objective truth source for any diagnostic error or quality issue associated with the manner in which diagnoses were described or omitted from the radiology report.
- a modified embodiment may supplement the attributes, or categories of attributes, with reliable measures of confidence or probability of correctness. These reliable measures of confidence or probability of correctness may be generated by statistical analysis of the variances across the attributes in reports that were generated by the radiologists performing structured and standardized radiology exam reviews. In some cases, the radiologists performing structured and standardized radiology exam reviews will independently review the same underlying radiology exam and generate reports that will contribute to the analysis of variance.
- the techniques herein are most suitable for assessing diagnostic accuracy, errors, and/or quality related to pathology or disease that is subject to generally good agreement among experts with respect to physical features that are present, location, size and so forth.
- the system for quantifying diagnostic radiology errors will be optimized to generate accurate quantitative measures of diagnostic error rates and quality issues related to specific radiology providers that are selected for assessment and their associated performance with respect to specific pathologies and diseases.
- These quantitative measures of diagnostic error rates may be aggregated to varying levels of anatomical detail, for example: (1) a combined measure representing the rate of any error that a radiology provider makes in the context of diagnostic knee MRI exams, or (2) a more narrow-scope measure representing the rate of any error that a radiology provider makes pertaining to an accurate diagnosis of meniscal tears within knee MRI exams.
- diagnostic error rates may also be aggregated to varying levels of diagnostic error types, for example: (1) a measure representing the rate of any false positive errors that a radiology provider makes in the context of diagnostic imaging exams, or (2) a measure representing the rate of any errors in which a finding is “undercalled”, or mistakenly graded as being too minor, that a radiology provider makes in the context of diagnostic imaging exams.
- these quantitative measures of diagnostic error rates may be aggregated to varying levels of within a radiology provider organization, for example: (1) a measure representing the rate of any diagnostic error that an individual radiologist makes in the context of selected diagnostic imaging exam types, or (2) a combined measure representing the rate of any error that a group of radiologists who practice together at single radiology facility make in the context of selected diagnostic imaging exam types.
- the measures of diagnostic error rates will be entirely based on the empirical diagnostic error data and attributes that are produced by the independent radiologists who perform standardized reviews of the exams performed by the radiology providers under review. In some embodiments, the measures of diagnostic error rates will be based, all or in part, on statistical modeling, including hierarchical Bayesian statistical modeling, of the empirical diagnostic error data and attributes.
- Some embodiments of the system for quantifying diagnostic radiology errors will also be optimized to generate measures of diagnostic quality that are modified versions of radiology provider error rates. These measures of diagnostic quality may be weighted combinations of specific diagnostic errors, such that the weighting may represent the relative likelihood that a specific type of diagnostic error will have an impact on patients' treatment pathways, clinical outcomes, or costs of treatment and subsequent care.
- the method for combining the various diagnostic error rates into the new quality measure may involve weighted averaging, linear or non-linear statistical modeling, or machine learning.
- the assignment of weights that represent the likelihood that specific types of diagnostic errors will have a clinical impact on patients may be accomplished by: (1) capturing additional data elements during the standardized diagnostic exam reviews, (2) stand-alone assessments by radiologist or other medical experts of the likely clinical impact of specific types of diagnostic errors, or (3) analysis of historical medical records of patients in combination with diagnostic error data to estimate the correlation of specific diagnostic errors or providers with specific error rates and impacts to patients' treatment patterns, costs, and outcomes.
- the diagnostic error data and attributes that are generated through standardized review of imaging exams will be supplemented with additional data and attributes about the radiology providers under evaluation.
- additional data and attributes may include: (1) radiologists' educational history, including fellowship training status, (2) radiologists' years of practice, (3) radiologists' historical exam volume and case mix, (4) radiology facilities' imaging equipment, or (5) radiology facilities' imaging exam protocol configurations.
- This supplementary data and attributes may be leveraged by the system to: (1) generate measures of diagnostic error rates or weighted diagnostic error rates with improved accuracy, precision, or narrower confidence intervals; or (2) to generate predicted measures of diagnostic error rates or weighted diagnostic error rates for radiology providers which have not had any of their imaging exams subjected to standardized reviews and for whom only the supplementary data elements and attributes are available.
- the methodologies that can be employed to leverage the supplementary radiology provider data and attributes in this way involves modeling the correlations between these new supplementary data elements and the data elements related to diagnostic errors and quality issues that are generated by the standardized imaging exam reviews; the quantitative methodologies that are used in this context may include Bayesian or log-linear statistical modeling or machine learning techniques.
- the system for quantifying diagnostic radiology errors will also be optimized to generate measures of diagnostic quality that are also adjusted for patient complexity, such that radiology providers may be penalized less for having higher rates of diagnostic errors when caring for a population of more complex patients and vice versa.
- the system may leverage combination of data from: standardized reviews of imaging exams, billing or claims data, patient demographic data, or other data extracted from electronic medical records.
- the system may employ Bayesian or log-linear statistical modeling, linear or non-linear regression, or machine learning methodologies to achieve the patient complexity adjustment of the diagnostic quality measures.
- patient complexity is adjusted for using a two-step process.
- diagnostic error rate estimates for each radiology provider under evaluation are modeled as conditional probabilities, i.e. diagnostic errors rate for each provider are estimated conditional on the presence of specific medical conditions and severities across the patient population observed for the radiology provider.
- conditional probabilities i.e. diagnostic errors rate for each provider are estimated conditional on the presence of specific medical conditions and severities across the patient population observed for the radiology provider.
- the diagnostic error rates estimated for each radiology provider, as conditional probabilities from step 1, can then be evaluated with respect to this distribution, i.e., E[f(Y′
- f(P* p*)] can be calculated for different providers, and these results can be directly compared to evaluate relative provider performance with respect to the same reference patient population.
- the attributes generated by the standardized exam reviews are used to train computer-implemented machine learning algorithms, for example recurrent neural networks or deep learning algorithms, such that the computer-implemented algorithms can then independently analyze digital radiology images and radiology reports and automatically apply the attributes that are included in the standardized exam reviews. Examples of such machine learning networks for automated diagnostic quality assessment are discussed in greater depth below, in Sections 7 and 8.
- These computer-implemented machine learning networks and algorithms can be trained to analyze radiology images to identify the presence or absence and severity of the specific pathologies that are assessed by the radiologists when they perform the standardized exam reviews.
- the algorithms may also be trained to generate attributes that describe the technical quality of the images, for example: (1) poor image quality (e.g. low signal-to-noise ratio), (2) images degraded or obscured by patient motion or other artifacts, (3) poorly configured exam protocols (e.g. an MRI exam conducted without collecting images that have a necessary image contrast setting or images collected with resolution that is too low), or (4) poor anatomical coverage of the images.
- the computer-implemented machine learning networks and algorithms can also be trained to analyze radiology reports to identify the presence or absence of specific diagnostic findings in the reports as well as the severity of the pathologies that are reported.
- the algorithms may also be trained to generate additional attributes related to the quality of the report, for example: (1) findings that are reported in an overly equivocal manner, (2) findings that are reported in an overly vague manner, (3) findings that are reported with inappropriate emphasis, (4) inappropriate or lack of comparisons with prior diagnostic studies, (5) inappropriate or lack of inclusion of relevant standard measures (e.g. not using the Breast Imaging Reporting and Data System or BI-RADS scoring system for mammogram reports), or (6) inappropriate or lack of follow-up recommendations.
- the algorithm performs its assessment on the images and report associated with a specific patient exam, it will compare its assessment of the pathologies in the images with its assessment of the diagnostic findings present in the radiology report to create attributes that represent the accuracy of the radiology report and any diagnostic errors that exist.
- the computer-implemented algorithm will produce measures of uncertainty for each attribute it generates related to the radiology images, radiology reports, and diagnostic errors. These measures of uncertainty will be based on quantitative assessments of the computer-implemented algorithm's performance in training and validation datasets. The measures of uncertainty may also incorporate measures of the underlying variability in accuracy of the training and validation datasets themselves. As discussed in greater depth below, these measures or other outputs of uncertainty from one or more components of the presently disclosed machine learning network(s) can be expressed as a feature vector, which can then be used as an input feature for the disclosed Bayesian approach to estimating physician's accuracies in diagnosing a pathology.
- the same statistical modeling methodologies described above may be applied to the diagnostic error attributes generated by the computer-implemented algorithms, in order to calculate estimates of radiology provider diagnostic error rates and weighted measures of diagnostic error rates and diagnostic accuracy.
- some embodiments may supplement the diagnostic error attributes with additional attributes related to radiology provider characteristics in order to generate measures of diagnostic error rates or weighted diagnostic error rates with improved accuracy, precision, or narrower confidence intervals
- the analytic approaches of embodiments may execute as overnight or background processes at any time after physicians or practitioners generate new radiology images or submit new radiology reports.
- the processes described for FIG. 1 , FIG. 3 may be executed in real-time immediately after a physician submits a report to provide immediate feedback to the healthcare provider in the form of a quality review or quality report.
- data indicating errors can be communicated to an administrator, third-party reviewer, or other system or program without direct notification to the primary physician who submitted a report.
- errors may be scored and ranked according to seriousness or severity, and only errors above a threshold severity value may be communicated to the primary physician.
- Location a region of the human body admitting specific distinct, though perhaps related, pathologies.
- Pathology a well-defined malady, for example, “central canal stenosis of the L2-3 segment in the lumbar spine”.
- Item a checklist question engineered to elicit a pathology-specific diagnosis.
- Diagnosis a selected value for an item, such as None, Small, Medium, Large.
- Checklist a collection of items capturing a specific diagnosis for a particular medical discipline or specialty.
- Reading provider a physician or practitioner who is the one providing diagnoses for evaluation.
- Reviewing provider a physician or practitioner who is evaluating the diagnoses of a reading provider after the fact, for accuracy.
- Practice a group of providers that is defined by business or geographic attributes.
- Provider a broad term for a physician, other healthcare practitioner, practice, group or other aggregation.
- FIG. 1 illustrates an example of functional elements and data flows in a distributed computer system that may be used to implement one embodiment of provider assessment processing.
- computer-implemented processes may be programmed to support assessment of the quality level of radiology providers and practices. Other embodiments may be applied to other medical disciplines.
- a provider data assessment computer system 10 comprises sampling logic 106 which receives unstructured medical data 102 as input, clinical data ingestion logic 108 and structured assessment logic 110 which may receive provider feature data and patient feature data for use in executing statistical modeling operations as further described herein. These functional elements cooperate, under program control as further described functionally herein, to generate structured provider quality data 118 , which may be provided as input to a grading algorithm 122 for calculation of output provider quality scores 126 . The resulting scores may be provided to or used as part of a designation process 130 and/or communication process 132 .
- a digital database 107 may be programmed to store the unstructured medical data 102 after input as well as the structured provider quality data 118 , output provider quality scores 126 , feature data 140 , 142 , and other data such as pathology prevalence data and error data for different fields of specialty.
- Computer system 10 may be implemented using one or more distributed or networked computers, services, processes or other software elements hosted using desktop computers, on-premises server computers or cloud computing instances of virtual computing centers. Each of the functional elements of computer system 10 may execute as a separate asynchronous thread, service or method. In some embodiments, multiple instances of functional elements may be provided.
- structured assessment logic 110 may execute as a plurality of independent instances in a virtualized computer to enable parallel processing of multiple datasets or parts of a single dataset.
- aspects of structured assessment logic 110 may be programmed as a SaaS application hosted on a web server to communicate with a browser executed at a user computer 14 that is coupled to computer system 10 directly or indirectly via one or more computer networks 12 or internetworks.
- sampling logic 106 is programmed to identify which types of exams and how many clinical exams to sample from radiology providers. Exams may be represented in digital images 104 , typically associated with reports 105 consisting of digitally stored text, as part of unstructured medical data 102 . For example, a particular report among the reports 105 may represent a set of comments or notes on pathological structures that are visible or believed to be visible in one or more associated digital images 104 . Thus, reports 105 typically represent physicians' diagnostic findings with respect to corresponding specific digital images 104 , and there may be thousands or millions of sets of images and reports for different patients, exams and diagnoses. In some embodiments, sampling logic 106 is programmed to calculate a sample of exams based upon an estimated or measured prevalence of key pathologies and diagnostic errors, combined with specific criteria relating to a particular kind of designation of the provider.
- the sampling logic 106 may apply a programmed rule to select a relatively high percentage, for example 50%, of all the exams for further analysis.
- a different set of scans with higher pathology prevalence and/or a higher known percentage of diagnostic error might trigger a programmed rule of the sampling logic 106 to select a lower percentage, for example 10%, of all exams in the set for analysis.
- the resulting percentage or number of exams that are selected by the sampling logic 106 may be weighted or biased by other attributes and data elements in database 107 related to the provider that provided the unstructured medical data 102 , for example: pre-existing quality designations or error rate estimates, the provider's patient volumes or cases mixes, or fellowship training status of providers.
- clinical data ingestion logic 108 is programmed to capture raw clinical data.
- raw clinical data may comprise medical images, which could be in the form of DICOM files, and diagnostic reports, as represented by digital images 104 and reports 105 .
- digital images 104 may comprise any form of graphical images that are captured in a radiology practice including X-ray, MRI or CT images, digital film or other diagnostic data. Images 104 may be associated with corresponding reports 105 , which consist of text in any digitally stored form.
- embodiments are not limited to radiology and other disciplines may interoperate with the processes herein based on raw clinical data of other types.
- raw clinical data may comprise electronic medical record (EMR) records or files, free-text notes, PDF files scanned from notes or generated from text files such as dictations, non-digital data such as the contents of a paper chart that has been scanned into image form or processed using optical character recognition (OCR), image-based diagnostic tests other than radiology imagery, claims data, billing data, employer-specific work data, audio files such as recordings of consultations or office visits with physicians or transcripts of the audio files, video recordings of surgeries or other interventions or procedures, or data from wearable devices.
- EMR electronic medical record
- OCR optical character recognition
- audio files such as recordings of consultations or office visits with physicians or transcripts of the audio files, video recordings of surgeries or other interventions or procedures, or data from wearable devices.
- raw clinical data may be partly structured; for example, data files may include metadata such as provider credentials, equipment attributes, length of exam, demographic or diagnostic features of patients.
- Quality attributes may relate to the technical performance of a diagnostic exam, such as poor-quality images or images that do not sufficiently cover the necessary anatomy.
- elements of FIG. 1 are programmed to transform the unstructured raw clinical data described above into at least partly structured data, and structured review procedures and machine-executed statistical analysis are performed to analyze the available data to derive error data and quality score values. Consequently, useful and meaningful values are extracted from previously non-usable data.
- clinical data ingestion logic 108 is programmed to use OCR and natural language processing (NLP) techniques, which may be implemented in external code libraries or web services, to convert unstructured diagnostic report text to structured, machine-readable data.
- clinical data ingestion logic 108 is programmed to use image processing libraries or functions to convert medical image data into structured, machine-readable data.
- clinical data ingestion logic 108 may be programmed to perform image feature identification in digital images 104 and generate output data comprising a graph, tree or list of features that have been identified.
- Other functional elements of computer system 10 are programmed to determine what diagnostic errors were made.
- errors could arise from low-quality images, motion artifacts from movement of the patient at the time of capturing an image, poor positioning of anatomy in relation to a camera or scanner, and so forth.
- trained primary physicians initially prepare the raw clinical data and images, and secondary reviewers use structured processes to assess features for quality.
- structured assessment logic 110 is programmed with parameterization logic 112 to execute clinical data assessment parameterization.
- the parameterization logic 112 executes in the context of a set of one or more digital images, from among the digital images 104 , that have been reviewed by a primary physician or practitioner and interpreted in a corresponding report from among the reports 105 .
- a particular report 105 comprises a written interpretation of a set of associated images, completed by a primary physician.
- the parameterization logic 112 may be programmed to:
- parameterization logic 112 may be programmed to present a list of available images in a graphical user interface with GUI widgets that are programmed to indicate selection of particular images.
- a structured checklist from among a plurality of structured checklists that are stored in database 107 , that applies to the digital image, a medical field that is associated with the selected digital image, or that is specified in configuration data.
- Each checklist may be digitally stored in the database 107 as a row of a database table in which columns represent diagnostic dimensions or parameters, and then rendered in a graphical user interface in the form of a checklist under program control; thus, literal storage as a document is not required and digital data structures may be used to represent checklists in storage.
- the secondary physician reviewer follows the checklist to detect and measure the prevalence of diagnostic errors and to control the generation of training data for artificial intelligence logic such as a neural network or classifier.
- the checklist addresses key diagnostic dimensions or parameters in interpretation of the digital images 104 for radiology or other specialties, customized to specific anatomical areas. Checklists may be created and stored in advance for any medical discipline and the key dimensions or parameters of quality of a checklist will reflect that discipline.
- a checklist may prompt for input from user computer 14 to indicate (a) whether disc herniation is present in the L4-5 lumbar spine and (b) if present, whether it is small, moderate or large.
- Input from user computer 14 may be stored in database 107 in association with identifiers of a dataset, a particular digital image among the digital images 104 , a checklist and a user account.
- the use of a checklist with digital image data will not be required and checklists may be assessed based on written reports or text data, as next described.
- the secondary reviewer physician compares their interpretation of the digital images with the original physician's diagnostic report as abstracted by the checklist.
- the reviewer uses the checklist and uses GUI widgets generated and displayed by the clinical data interpretation logic 114 to parameterize the level of agreement or disagreement between the reviewer's interpretation and the original interpretation, producing data that describes diagnostic errors.
- clinical data interpretation logic 114 may be programmed to presume that the reviewer is correct, but some embodiments may model, under program control, variability of interpretation among reviewers, as further described.
- each stored checklist later may be used as a portion of training data for training the statistical modeling logic 116 when implemented as a neural network or classifier.
- the statistical modeling logic 116 may execute to receive the digital images 104 , receive the reports 105 , interpret the images according to one or more checklists, interpret the original physician's diagnostic report according to the checklist, compare the machine-generated interpretation of the images to the original physician's diagnostic report, utilizing the checklist to parameterize levels of agreement or disagreement, and generate output data identifying diagnostic errors with associated confidence level values.
- the statistical modeling logic 116 may receive provider feature data 140 and patient feature data as input to adjust the classification of images and reports, and output error data, based on variable features of providers and patients, as further described in other sections. Broadly, statistical modeling logic 116 executes as a trained classifier to detect errors in unstructured medical diagnostic data after training on similar medical diagnostic data in which errors have been explicitly identified.
- provider error date data 120 may form one component of stored, structured provider quality data 118 .
- structured provider quality data 118 may be used in several different ways.
- the quality data 118 may be provided as input to the grading algorithm 122 , which is programmed to use weighting logic 124 and patient complexity adjustment 126 to transform the error data.
- weighting logic 124 applies weight values to quality scores based on a combination of expert clinical input and data-drive insights about outcomes. These factors may be used to calculate weight values to assign to specific diagnostic errors, representing a weight of that error relative to its impact on later clinical care or treatment. Thus, a particular error may have a high weight value if its impact on clinical care or treatment, such as the complexity of a later treatment, patient discomfort or cost is high. Thus, a particular quality score 128 may be adjusted upward or downward based on the weight value associated with the error(s) represented in error rate data 120 that led to the score.
- Patient complexity adjustment 126 is programmed to obtain data from database 107 for patient complexity including but not limited to demographic data such as age and sex, and clinical interpretation data such as number and severity of the pathologies identified in exams. Therefore, particular healthcare providers are not inappropriately credited or penalized, as part of determining quality scores 128 , based on patient population dynamics.
- grading algorithm 122 may be programmed to output provider quality scores 128 , representing an overall quality score for a particular healthcare provider based on its error rate, the complexity of patients seen, and various features of the provider.
- the quality scores 128 may be used in a designation process 130 to designate a particular healthcare provider using a particular label or designation from among a plurality of different labels or designations, using an ordered scale, hierarchical arrangement or other association of labels.
- the quality scores 128 also may be provided to healthcare providers according to a structured communication process 132 .
- FIG. 2 illustrates further details of the statistical modeling logic of FIG. 1 .
- FIG. 3 illustrates an example data assessment process that may be used in an embodiment.
- the statistical modeling logic 116 is programmed to execute a hierarchical Bayesian statistical model 200 . All elements of statistical modeling logic 116 are implemented using one or more computer programs, methods, web services, microservices and/or other software elements.
- foundation methodology for the statistical model 200 is to reduce outliers, narrow confidence intervals and improve the accuracy of estimates of true diagnostic error rates based on observed samples, especially for rarer types of diagnostic errors.
- statistical model 200 uses a population-wide priors model 202 , inter-feature correlation model 204 and inter-reviewer variability model 206 .
- the inter-reviewer variability model 206 is programmed to assess the reliability and consistency regarding the detection and measurement of specific types of diagnostic errors by reviewers. Its output may be used to assign confidence interval values and probability values to the provider error rate data 120 ( FIG. 1 ).
- Statistical model 200 may store and use a contingency table 208 and distribution data 210 comprising one or more statistical distributions that are calculated as interim steps, as further described in this section.
- inter-feature correlation model 204 is programmed to use statistical techniques to characterize the correlation between groups of features. For example, groups of diagnostic error rates may be correlated; examples might be errors related to all lumbar spine pathologies, or the relationship between all diagnostic error rates of the type “overcall” to all diagnostic error rates of the type “undercall”.
- the inter-reviewer variability model 206 is programmed to execute the seven-step process described above for parameterization logic 112 , for a subset of exams consisting of associated digital images 104 and reports 105 , for a plurality of different reviewers and to assess the level of agreement or disagreement of different reviewers, yielding an inter-reviewer variability score value.
- the inter-reviewer variability score value may be used as a factor in the statistical modeling logic 116 .
- integration of provider feature data 140 and patient feature data 142 can further improve the estimate of true diagnostic error rates and can allow for estimates of diagnostic error rates for which the database 107 stores limited to no observed error rates.
- examples of features that can be represented in provider feature data 140 comprise educational history, size of practice and type of imaging equipment.
- Examples of features that can be represented in patient feature data 142 are age, sex, other demographic values and diagnosis.
- Statistical model 200 also may receive provider hierarchy metadata 210 , from database 107 for example.
- the provider hierarchy metadata 210 enables statistical model 200 to factor in the hierarchical structure of a healthcare provider.
- provider hierarchy metadata 210 may specify that a particular provider is a practice, facility, individual physician or radiologist, or reflect other hierarchical levels or categories.
- features of each entity represented in provider hierarchy metadata 210 include practice data such as size and academic affiliation; facility data such as type of imaging equipment and imaging protocols that are used; physician data such as years in practice and training attributes; and reviewer data such as years in practice and training attributes.
- Provider hierarchy metadata 210 may be created and stored for all the providers that are assessed using the computer system 10 .
- the use of provider hierarchy metadata 210 enables statistical model 200 to differentiate and cross-relate features at the appropriate hierarchical level for each entity, thereby allowing for the most accurate estimate of true diagnostic error rates achieved by various practitioners.
- statistical model 200 is programmed to execute the following capabilities:
- Modeling of the agreement between reading provider and reviewer provider for a diagnosis at the item level including: estimation of item-level diagnostic accuracy; calibration of the uncertainty of the “gold” standard diagnoses from reviewing providers using variability and inter-reviewer agreement measurements that are calculated from the data generated when multiple reviewing providers assess the same radiology exams and examples of the same pathologies and diagnostic errors.
- G Information sharing and data pooling capabilities, including segmentation of provider populations, hierarchically informed estimation of population, and parsimonious inferential specifications.
- statistical model 200 is programmed to execute, using the computer system 10 , functions that may be expressed for convenience in the following mathematical notation.
- contingency table 208 yields a log-linear contingency table represented in FIG. 2 as contingency table 208 .
- the function provides a co-occurrence distribution of reviewing provider diagnoses R 1l , . . . R pl for p items at location l with risk adjustment for features X (R) .
- Observable variation in —R n represents a categorical distribution capturing the observable variation in R n .
- Observable variation in —R n is identified directly through repeated measures of multiple reviewing providers within specific checklists, as well as parametrically estimated across the population of all relevant checklists.
- an expert informed and healthcare impact driven score value may be derived by calculating: g k (R 1l ,D 1l , . . . , R pl ,D pl
- the function g k is defined on the basis of both expert opinion elicitation (E k ) and empirical evidence (Y k ) and aggregates accuracy portfolios into scores characterizing performance with respect to specific (k-th) financial and care outcomes.
- ⁇ i is a feature-driven, hierarchically informed parameter that is specific to D 1l
- the structure and degree of dependence between ⁇ i (i 1, . . . p), e.g., ( ⁇ 1 , . . . ⁇ p ) approximates f( ⁇ , ⁇ ⁇ ) explicitly models and drives accuracy dependency across item panels; the specification of this form addresses appropriateness and validation of the model.
- X (D) may denote a provider or features characterizing providers, which allows for non-specific provider aggregations.
- Particular ⁇ i specifications reflect X (D) and capture associations attributable to X (D) while informing estimation across I via dependency structure in ⁇ i .
- Predictive extrapolation is available through standard X (D) ⁇ i linear form inference.
- Mixture model or post-hoc subpopulation segmentation provides aggregation driven estimation. Structure and dependency across ⁇ i provides hierarchical information pooling and sharing. Parsimonious feature engineering in log-linear model and multi-class classification contexts addresses infeasible saturated model approaches.
- Block 302 may comprise executing the functions of sampling logic 106 ( FIG. 1 ) that have been previously described, including all alternatives and variations.
- Block 304 clinical data ingestion is performed.
- Block 304 may comprise executing the functions of clinical data ingestion logic 108 that have been previously described, including all alternatives and variations.
- Block 306 clinical data assessment parameterization is performed.
- Block 306 may comprise executing the operations of structured assessment logic 110 as previously described, including all alternatives and variations.
- Block 308 clinical data interpretation is performed.
- Block 308 may involve executing the operations of clinical data interpretation logic 114 as previously described, including all alternatives and variations.
- Block 310 statistical modeling of diagnostic error rates based in part on provider features and patient features is performed.
- Block 310 may comprise executing the operations of statistical modeling logic 116 as previously described, including all alternatives and variations.
- Block 320 quality scoring of providers with clinical impact weighting and patient complexity adjustment may be performed.
- Block 320 may comprise using structured provider quality data 118 , including provider error rate data 120 , with grading algorithm 122 and the weighting and patient complexity adjustment that have been described, to yield output provider quality scores 128 , as previously described, including all alternatives and variations.
- the quality scores 128 may be provided as an element of feedback to block 310 to improve training and refinement of the statistical modeling logic 116 .
- designation process 130 may be programmed, or used manually, to create and store designations of healthcare providers based on thresholds, a hierarchy or a ranking or labeling system.
- radiology providers may be designated as high quality providers or Centers of Excellence based on the output provider quality scores 128 that are generated for the providers.
- Designations may be generated based on absolute values of the quality scores 128 or based on the scores in relation to later or downstream outcomes that are observed in patient populations.
- data for outcomes for this purpose may be obtained from medical insurance claims records.
- the designation process 130 may determine designations based on criteria such as comparison of quality scores 128 to thresholds derived from national benchmark data or regional benchmark data.
- the benchmark data may be stored in database 107 and may be determined over time by the computer system 10 , by computing quality scores 128 for a plurality of providers and storing the score values in the database in association with provider identifying data that specifies geographic location. Thereafter, the score values may be sorted and grouped by region or nation to derive mean, median or other statistically significant values for providers in a particular group, region or nation.
- a new quality score 128 generated for a particular provider can be compared to the benchmark for a region or nation in which that particular provider is located; if the new quality score passes a threshold value corresponding to the benchmark value, then a particular designation may be created and stored, or awarded.
- the processes described in section (2) and section (3) of this document may be repeated on an ongoing basis to monitor the performance of providers over time, recalculate provider error rate data 120 and regenerate output provider quality scores 128 for the same providers. Ongoing repetition and recalculation in this manner is expected to further increase confidence levels associated with scores and designations.
- communication process 132 may be programmed using presentation layer logic of computer system 10 to generate performance reports or dashboards that contain applications of the information generated via section (2) and section (3).
- the communication of provider error rate data 120 , output provider quality scores 128 , designations and/or data distilled from these values is expected to induce providers to elevate the standard of care that they provide.
- Embodiments have been described that provide data-driven, objective assessment of healthcare provider diagnoses with the benefit of generating error data and quality scores that have not been available previously.
- radiology or other healthcare quality measures are based on easily accessible proxy measures of medical care quality that focus on: process or workflow (e.g. average time between stroke patient arrival at provider facility and start of stroke treatment), structure (e.g. percentage of CT exam images and reports that providers make available to unaffiliated providers for the purposes of prior study comparisons), patient safety or outcomes (e.g. death rate of patients undergoing carotid artery stenting procedures), or subjective patient satisfaction surveys (e.g. patient feedback on wait times or physician bedside manner).
- process or workflow e.g. average time between stroke patient arrival at provider facility and start of stroke treatment
- structure e.g. percentage of CT exam images and reports that providers make available to unaffiliated providers for the purposes of prior study comparisons
- patient safety or outcomes e.g. death rate of patients undergoing carotid artery stenting procedures
- subjective patient satisfaction surveys e.g. patient feedback on wait times or physician bedside manner.
- Radiology provider organizations for example the American College of Radiology (ACR) has proposed the “RadPeer” program in which radiologists review a sample of radiology exams performed by other radiologists in their organizations and assign a subjective summary quality score of 1a, 2a, 2b, 3a, or 3b, to indicate if the overall quality of the diagnostic imaging exam under review achieved satisfactory or unsatisfactory quality and whether any diagnostic errors that are present are likely to have a clinically significant impact on the patient.
- ACR American College of Radiology
- RadPeer Radiologists review a sample of radiology exams performed by other radiologists in their organizations and assign a subjective summary quality score of 1a, 2a, 2b, 3a, or 3b, to indicate if the overall quality of the diagnostic imaging exam under review achieved satisfactory or unsatisfactory quality and whether any diagnostic errors that are present are likely to have a clinically significant impact on the patient.
- This approach to quality measurement suffers from deficiencies that include: quality scores that do generalize across provider organizations, low levels of reproducibility, and
- radiology quality measures that: (1) are not proxy measures of clinical care quality and instead focus directly on the quality of diagnostic imaging care (i.e. diagnostic accuracy and rates of diagnostic errors), (2) do not require a secondary diagnostic test like a biopsy to be available to serve as a “gold standard comparison”, and (3) are not based on subjective summary assessments from peers within the same provider organization and instead captures quality assessment data in a structured, granular and systematic manner that allows robust and reliable quantification of diagnostic error rates and associated confidence intervals.
- the framework described here in which structured data attributes related to diagnoses and diagnostic errors are generated from each exam quality assessment review, enables: (1) the method to be scaled and supplemented using machine-implemented algorithms that are trained using the reviewer-generated attributes, and (2) for correlations between the structured data attributes and additional provider attributes to be characterized, which allows measures of diagnostic error rates or weighted diagnostic error rates to be generate with improved accuracy and precision and generated for radiology providers which have not had any of their imaging exams subjected to standardized reviews (for whom only the supplementary data elements and attributes are available).
- the techniques herein provide opportunities for peer improvement by exposing objective and detailed factors that affect quality, rather than leaving medical disciplines to operate in an environment in which practices do not know why a particular practitioner has a high or low error rate, or may be associated with patients who experience better or worse healthcare outcomes. Instead, data features exposed in the present techniques provide reliable and robust measurements of error rates. This evidence can provide reasons to improve a practice's equipment, procedures, types of exam routing or other issues.
- one or more machine learning algorithms can be trained to provide an automated assessment of the quality of a diagnostic made from a radiological exam—similar to the diagnostic assessment described above with respect to FIGS. 1-3 .
- These machine learning algorithms can include, but are not limited to, neural networks, recurrent neural networks, convolutional neural networks, or one or more other machine learning algorithms more commonly referred to as deep learning algorithms.
- a machine learning network trained according to the present disclosure receives as inputs the underlying radiological report and radiological images associated with a given diagnostic or patient, and then automatically regresses to an estimate of the error (if any) contained within the given diagnostic. Notably, the trained machine learning network performs this error regression calculation without requiring additional inputs or external guidance.
- motion segments are physiological units of the spine, each consisting of two adjacent vertebrae, the intervertebral disc and the adjoining ligaments between. Motion segments provide a nomenclature to identify and refer to various locations along the spine, and hence are particular to the example scenario of a lumbar spine exam. It is noted that this example is for illustrative purposes only and is not intended to be limiting as to the scope of the present application.
- lumbar spine exams is provided to illustrate one specific application of the disclosed machine learning networks for automated diagnostic quality assessment—machine learning networks which, it is appreciated, can be applied to various types of different radiological exams, reports, and/or images without departing from the scope of the present disclosure.
- FIGS. 4A-B depict a flowchart of a pre-processing pipeline 400 that receives as input raw radiological images 404 and radiological reports 405 .
- the radiological images 404 may be the same as the digital images 104 that are stored in the database described with respect to FIG. 1 .
- the radiological reports 405 may be the same as the physician diagnostic reports 105 that are also stored in the database described with respect to FIG. 1 .
- FIGS. 4A-B depict a flowchart of a pre-processing pipeline 400 that receives as input raw radiological images 404 and radiological reports 405 .
- the radiological images 404 may be the same as the digital images 104 that are stored in the database described with respect to FIG. 1 .
- the radiological reports 405 may be the same as the physician diagnostic reports 105 that are also stored in the database described with respect to FIG. 1 .
- Pre-processing pipeline 400 consists of a feature identification step 410 and a feature extraction step 420 , which operate to clean and standardize the raw input radiological images 404 and radiological reports 405 into a format that is better and more effectively utilized by the down-stream machine learning networks depicted in FIGS. 5-8 (each of which will be discussed in turn below).
- Pre-processing pipeline 400 is designed to extract information to match anatomical sections from the radiological images and reports 404 , 405 . On the basis of this matching, the identified pairs or groups of anatomical sections identified from the radiological images and reports 404 , 405 can thereafter be processed jointly.
- the anatomical sections upon which pre-processing pipeline 400 operates are motion segments (i.e., specific locations/vertebrae pairs along the spine).
- this pre-processing step is to generalize the isolation of specific anatomical regions, as described in radiological reports, and to extract the corresponding regions in the medical images (e.g., MR/CT/Ultrasound/Digital Pathology, etc.) to match the assessment from both ends. Therefore, aspects of the present disclosure are applicable to any type of radiological and/or pathological exam, and the example application to spinal MRI images described below is not to be construed as limiting.
- pre-processing pipeline 400 receives as input one or more sets of radiological images and reports 404 , 405 that correspond to the same underlying patient/specific diagnostic.
- these inputs might be received in substantially real-time, i.e. after the radiological report 404 is submitted by the examining radiologist, or after the radiological images 405 are retrieved from the scanner where they were captured.
- one or more of the input radiological images and reports 404 , 405 might be from a database or other storage system at some time after the original generation of the radiological image and/or report.
- the input radiological images and reports 404 , 405 are initially processed by independent pipelines.
- a first pre-processing pipeline is trained to perform feature identification 410 and feature extraction 420 with respect to input radiological reports 405
- a second pre-processing pipeline is trained to perform the same with respect to input radiological images 404 .
- specific landmarks of interest that might be extracted as features include the paragraphs or sentences within the report where the radiologist referred to or identified a particular motion segment. Text in the report referring to specific motion segments are isolated to be treated independently. For example, a sentence reading “L1-2: Mild broad-based disc bulge with moderate bilateral face hypertrophy which causes mild AP canal narrowing as well as mild bilateral neural foraminal narrowing” would be extracted as a motion segment feature for the L1-2 motion segment of the spine.
- radiological images 404 specific landmarks of interest that might be extracted as features include the image section or pixel area occupied by a particular motion segment. Because a single radiological exam might produce several different sequences of radiological images 404 (e.g. an MRI exam might produce a first sequence of T1w images, a second sequence of T2w images, etc.), the same given motion segment can be extracted multiple times, i.e. at least once for each constituent image sequence contained within the input radiological images 404 . These multiple corresponding motion segments can then be treated independently, similar to the separate treatment of motion segments referred to in multiple places within the radiological report text.
- MRI exam might produce a first sequence of T1w images, a second sequence of T2w images, etc.
- pre-processing pipeline 400 to input data consisting of radiological images and reports provides structured output data pertaining to specific motion segments, i.e., in the form of corresponding image data 422 and text data 424 extracted from the radiological images and reports 404 , 405 , respectively.
- an example output of one pair/grouping of corresponding extracted data consists of: text section 422 (comprising a sentence reading “L1-2: Mild broad-based disc bulge with moderate bilateral face hypertrophy which causes mild AP canal narrowing as well as mild bilateral neural foraminal narrowing”) and an image motion segment 424 (comprising the pixel area occupied by the L1-2 motion structure).
- image motion segment 424 comprising the pixel area occupied by the L1-2 motion structure.
- pre-processing pipeline 400 identifies all of the motion segments that are present in the input radiological images 404 and extracts one or more image sections corresponding to each motion segment. Similarly, in some embodiments pre-processing pipeline 400 identifies all of the motion segments that are referred to or described in the input radiological reports 405 and extracts one or more text sections corresponding to each motion segment.
- FIG. 5 depicts an architecture diagram for a multi-regularizer machine learning network 500 to detect diagnostic errors in radiological examinations.
- One or more portions, components, and/or layers of the machine learning network 500 can be provided as recurrent networks, non-recurrent networks, or some combination of the two, as will be described in greater depth below.
- Recurrent models can include, but are not limited to, recurrent neural networks (RNNs), gated recurrent units (GRUs), and long short-term memory (LSTMs).
- RNNs recurrent neural networks
- GRUs gated recurrent units
- LSTMs long short-term memory
- one or more portions or components of the machine learning networks disclosed herein can be configured as fully-connected networks, convolutional neural networks (CNNs), or some combination of the two.
- CNNs convolutional neural networks
- the trained ML network 500 receives as input a text section 422 (extracted from a full radiological report) and an image section 424 (extracted from a full radiological image) that both correspond to the same motion segment, pathology or anatomical location.
- ML network 500 can receive the input text section 422 from the output of pre-processing pipeline 400 as applied to the full radiological report and can receive the input of image section 424 from the output of pre-processing pipeline 400 as applied to the full radiological image.
- the trained ML network 500 analyzes the text section 422 and the image section 424 against one another and generates an output indicating the quality of the diagnosis contained within text section 422 .
- an output 550 a denotes “Agree,” or that the finding contained within text section 422 is generally in agreement or otherwise consistent with the pathologies contained within image section 424 .
- An output 550 b denotes “Overcall,” or that the finding contained within text section 422 is more severe than the pathologies contained within image section 424 indicate.
- An output 550 c denotes “Undercall,” or that the finding contained within text section 422 is less severe than the pathologies contained within image section 424 indicate.
- an output 550 d denotes “Missed,” or that the finding contained within text section 422 is inconsistent with the pathologies contained within image section 424 .
- the “Missed” output 550 d can be further divided into false positives, in which the finding contained within text section 422 is absent from the pathologies contained within image section 424 , and false negatives, in which a pathology contained within image section 424 is absent from the findings contained within text section 422 . It is noted that the outputs 550 a - d are provided for purposes of illustration, and that ML network 500 could be trained to provide a different set of outputs, providing more or less granularity as desired, without departing from the scope of the disclosure.
- the degree of Overcall and/or Undercall can also be included or otherwise represented in the regression.
- the regression could introduce Overcall degree 1, 2, or 3 and Undercall degree 1, 2, or 3—although of course it is appreciated that various other granularity scales can be utilized without departing from the scope of the present disclosure.
- this is because the different outputs of ML network 500 , such as the illustrated outputs 550 a - d , are configured as the different categories or classes upon which an output classifier of ML network 500 is trained.
- ML network 500 can additionally contain a second output classifier to regress to a clinical significance of the diagnostic error(s) 550 b - c that are identified by the first output classifier described above.
- the second output classifier could output a clinical significance score of 0, 1 or 2, where a score of 0 indicates no clinical significance (or no error), a score of 1 indicates a moderate clinical significance, and a score of 2 indicates a high clinical significance.
- the exact outputs of a clinical significance classifier can be determined, modified or otherwise adjusted as desired during the training process of ML network 500 .
- the clinical significance scores can be a range of discrete numbers, as in the present example, or can be continuous between a minimum and maximum value.
- the possible range of clinical significance scores might be determined by the definition of clinical significance provided by the overall quality assessment process in which the trained ML network 500 is utilized.
- the trained ML network 500 does not require any additional inputs beyond the text sections 422 and the image sections 424 —both of which are already collected and stored in the course of a conventional radiological exam.
- a large portion of the training data needed to train ML network 500 and its constituent components can be obtained by leveraging the already existing data stored in, for example, database 107 of FIG. 1 , which significantly reduces the burdensome need of actively collecting, collating and annotating training data from scratch.
- additional or supplemental annotations can be generated and applied to the existing data obtained from databases such as database 107 of FIG. 1 .
- these additional/supplemental annotations could be utilized to specifically target pathologies within the images, or to replace missing annotations that otherwise would have already been associated with the existing data in database 107 .
- database 107 contains various forms of structured data generated from digital images 104 and reports 105 —images and reports that are similar or even identical in form to the radiological images and reports 404 , 405 upon which the trained ML network 1500 will be applied.
- the structured data collected and stored in database 107 includes a plurality of checklists generated by parameterization logic 112 , wherein a secondary physician reviewer (or one or more selected expert reviewers) views a radiological image and provides input indicating the presence, location, and extent of any pathologies present in the radiological image.
- the secondary physician reviewer/expert can furthermore view the initial report accompanying the same radiological image and provide input to the checklist of parameterization logic 112 indicative of any diagnostic errors contained within the initial report.
- a secondary physician reviewer or one or more selected expert reviewers
- database 107 contains a multiple thousands of these checklists and other structured data that parameterize the level of agreement or disagreement between the original physician/radiologist who produced the original report and one or more secondary physicians/selected experts who performed a review. Because the input images and reports used to generate the structured data and checklist reviews stored in database 107 are similar or identical to the radiological images 404 and radiological reports 405 that will be provided as inputs to the trained ML network 500 , these checklist reviews can be utilized or transformed into annotated training data.
- ML network 500 consists of three encoder networks, F NLP , F IM and E NLP_IM , and at least one regularizer per encoder network. These regularizers contribute to the overall loss function that is used to train ML network 500 , and more particularly, do so by defining specifically tailored losses to refine the encoder network to which the regularizer is attached.
- the training of ML network 500 is driven by categorical cross entropy loss, as will be explained in greater depth below.
- the first encoder network 512 also referred to herein as F NLP , is trained to generate embeddings for specific pathologies within the input section of report text 422 .
- the input sections of report text 422 are provided to first encoder network 512 after being extracted from the overall radiological report 405 (i.e., using pre-processing pipeline 400 of FIG. 4 ).
- the output of the radiological report pre-processing pipeline can be coupled to the input of first encoder network 512 .
- the outputs from the radiological report pre-processing pipeline can be extracted in advance, then stored in a database and retrieved as needed by ML network 500 and first encoder 512 .
- first encoder 512 is trained to generate embeddings that represent pathologies in a consistent and more computationally advantageous manner.
- a word embedding is a real-valued vector that represents a single word based on the context in which it appears. By doing so, embeddings translate an input of many dimensions (e.g. the words within report text 422 ) into an output with a much smaller number of dimensions.
- semantic information of the input report text 422 is in theory captured by the expectation that embeddings for semantically or syntactically related words will be closer to each other in the vector space than to unrelated words in the vector space.
- the degree to which the embeddings actually embody this relatedness is dependent on the text corpus or training data from which the first encoder network 512 , F NLP , learns to derive these embeddings.
- the first encoder network 512 can be of recurrent form.
- F NLP might be provided as an Ordered Neuron Long Short-Term Memory (ON-LSTM) network, which have information (memory) retention characteristics that are particularly well suited for processing long input sequences such as report text 422 .
- First encoder network 512 can also be a Transformer-based network, which is a deep learning model that is also designed to handle ordered sequences of data—such as report text 422 —but without requiring that the input sequence be processed in order. In other words, a Transformer-based implementation of first encoder network 512 does not need to process the beginning of report text 422 before processing the middle or end of the text.
- Examples of Transformer-based machine learning networks include, but are not limited to, BERT (Bidirectional Encoder Representations from Transformers) and ClinicalBERT (a BERT model that has been pre-trained on a specialty corpus of clinical text and medical concepts).
- Domain-specific training can be provided to first encoder network 512 to better refine F NLP for use in the radiological context in which both it and the overall ML network 500 are to be utilized.
- a plurality of radiology reports can be assembled into a radiology-specific corpus of text, and first encoder network 512 can be obtained by training a dedicated Transformer model on the radiology-specific corpus.
- first encoder network 512 can be pre-trained on a broader corpus, e.g., general English language, medical texts, clinical texts, etc., and then further trained on the radiology-specific corpus.
- the radiology-specific corpus does not require annotation or manual labeling, as first encoder network 512 is able to learn word embeddings directly from the corpus in an unsupervised learning process. Accordingly, the radiology-specific corpus can be assembled from one or more of the radiology reports 105 that are stored in the database 107 , as described with respect to FIG. 1 , although it is also possible that the radiology-specific corpus be externally derived or obtained.
- the word embeddings can be word 2 vec embeddings, although it is appreciated that various other types of embeddings can be utilized without departing from the scope of the present disclosure.
- each encoder network within ML network 500 is associated with at least one regularizer.
- F NLP the architecture diagram of FIG. 5 depicts two associated regularizers: a concept generator 522 (labeled as G C ) and an NLP pathology classifier 532 (C NLP ).
- G C concept generator 522
- C NLP NLP pathology classifier 532
- the first regularization network consists of concept generator 522 , G C , which trains and refines the manner in which the first encoder 512 , F NLP , syntactically parses and analyzes the report text 422 .
- Report text 422 contains diagnosis information that reflects the reviewing physician or radiologist's interpretation of the medical image data 424 .
- This diagnosis information typically consists of a location, a pathology, and a severity—although other constituent components can be used to form the diagnosis information without departing from the scope of the present disclosure.
- concept generator 522 helps standardize the handling and treatment of non-standardized natural language textual inputs such as report text 422 .
- original diagnosis information can be represented by a ‘concept’ data structure C ori , which is a triplet given by (L, P, S), where L is the location of the identified pathology, P is the identified pathology, and S is the severity of the identified pathology.
- Concept generator 522 helps regularize first encoder 512 by applying a training process in which the encoding features of F NLP are used to generate synthetic data (new concept triplets) that are matched against known information of a corresponding type or form (the original concept triplet C ori ). On this basis, concept generator 522 can be used to drive a loss L NLPGen that minimizes the difference between the generated new concept triplet and the original concept triplet C ori .
- Concept generator 522 can be trained to output new concept triplets for inputs of actual report text 422 .
- the requisite annotated training data can consist of labeled pairs of report text and the corresponding original concept triplet C ori for that report text.
- the pre-existing radiological reports and structured data stored within database 107 of FIG. 1 can be leveraged to automatically generate the requisite training data in the form of data pairs comprising ⁇ radiological report text, corresponding original concept triplet C ori ⁇ .
- concept generator 522 can instead, or additionally, be trained to output relevant sections of text that relate to the actual report text input.
- concept generator 522 can be trained to identify relevant regions or sets of words within an input report text 422 for each of the three diagnostic attributes of the concept triplets, i.e. location, pathology, severity.
- concept generator 522 refines first encoder 512 by applying category saliency to highlight the area/regions of report text that are discriminative for the three different diagnostic attributes.
- concept generator 522 constitutes an additional component used to fine-tune the training of first encoder 512 and the remainder of machine learning network 500 . Based on the loss function L NLPGen , the loss of concept generator 522 is back propagated to refine the various layers and parameters of first encoder 512 , F NLP .
- the second regularization network that is applied to first encoder 512 , F NLP consists of an NLP pathology classifier 532 (C NLP ).
- NLP pathology classifier 532 trains and refines first encoder 512 with respect to the independent pathology classification for input report text 422 . Any given segment of input report text 422 has an associated ground truth, which in this case can be thought of as the diagnosis as the reviewing physician/radiologist intended to read the radiological images.
- the second regularization network i.e., NLP pathology classifier 532
- the automated diagnostic quality evaluation performed by machine learning network 500 depends upon a correct interpretation of the radiological report that is the subject of the evaluation.
- NLP pathology classifier 532 consists of classification layers added off of embeddings from the first encoder network 512 , F NLP .
- these classification layers are driven by a binary cross entropy (BCE) loss L NLP_Path .
- BCE loss is utilized here because the output pathology classification for a segment of input report text 422 is either correct (i.e., the same as the ground truth pathology for report text 422 ) or incorrect (i.e., not the same as the ground truth pathology for report text 422 ).
- BCE loss L NLP_Path the first encoder network 512 is regularized and refined with respect to its ability to detect pathologies from input report text 422 relative to the ground truth.
- the requisite training data used in conjunction with NLP pathology classifier 532 can be obtained in much the same way as was described previously with respect to the training data for concept generator 522 —by leveraging pre-existing radiological reports and structured clinical interpretation data stored, for example, in database 107 of FIG. 1 . Because these radiological reports 105 have already been interpreted by, e.g., clinical data interpretation logic 114 , training data for use with NLP pathology classifier 532 can be generated by annotating a given report 105 with the one or more pathologies determined by clinical data interpretation logic 114 , as these pathologies are the ground truth for the given report 105 .
- second encoder network 514 also referred to herein as F IM .
- F IM performs for segments of input radiological image regions 424 .
- Second encoder network 514 is trained to generate features (or embeddings) from the set of imaging sequences available for specific anatomical regions.
- the input radiological image regions 424 are provided to second encoder network 514 after being extracted from the overall radiological images 404 (i.e., using pre-processing pipeline 400 of FIG. 4 ).
- the output of the radiological image pre-processing pipeline can be coupled to the input of second encoder network 514 , although it is also possible that the outputs from the radiological image pre-processing pipeline can be extracted in advance, then stored in a database and retrieved as needed by ML network 500 and second encoder network 514 .
- second encoder network 514 can be based on ResNet (a type of residual neural network) or DenseNet (a dense convolutional network), with a proper adaptation to medical images that handles anisotropy and the diverse intensity distribution that are associated with and typical in many of the radiological images that are provided as input to second encoder network 514 .
- ResNet a type of residual neural network
- DenseNet a dense convolutional network
- first encoder network 512 is regularized by a generator network ( 522 ) and a pathology classifier ( 532 ), so too is second encoder network 514 .
- second encoder network 514 is regularized by a heatmap generator network 524 and an image pathology classifier 534 .
- Heatmap generator 524 , G H refines the manner in which second encoder 514 , F IM analyzes the input images 424 .
- Heatmap generator 524 is trained such that second encoder 514 is fine-tuned to focus on certain image locations or anatomical regions that have been observed or are otherwise known to be relevant to pathological structures.
- second encoder 514 is trained to, in effect, give greater weight to relevant portions of input images 424 (e.g. portions that include anatomical and/or pathological structures) and lesser weight to non-relevant portions of the input images (e.g. the empty space surrounding the anatomical/pathological structures, such as the black space on the left and right sides of the example input image 424 shown in FIG. 5 ).
- Second encoder 514 is therefore trained away from outputting features based on portions of radiological images that a reviewing physician/radiologist would not consider when performing their review. For example, assuming that input image 424 contains one or more pathologies of interest, then these pathologies will usually be located in specific portions of the input image, e.g., a bulged disc will be located between or near two vertebrae
- heatmap generator 524 can be configured to generate attention heatmaps from specific layers of second encoder 514 . These attention heatmaps, or activation maps, represent the discriminative image regions used by second encoder 514 in identifying a specific feature in an input image 424 .
- knowledge of the relevant portions of a radiological image for making a diagnosis or identifying pathological structure(s) can be used to create annotated heatmaps, which serve as training data for heatmap generator 524 .
- one or more annotated heatmaps can be automatically generated by tracking gaze information of a radiologist as he or she reviews radiological images, with heatmap intensity reflecting the amount of time that the radiologist focused on a given location of the radiological image.
- Heatmap generator 524 can be trained through a process that provides training data pairs comprising ⁇ sample input image, annotated heatmap for the sample input image ⁇ .
- the sample input image is fed through second encoder 514 , and heatmap generator 524 generates one or more heatmaps corresponding to layers of the second encoder as they processed the sample input image.
- second encoder 514 is refined such that its discriminative regions become better aligned with the known relevant regions of radiological images.
- second encoder 514 can be regularized via a decoder that performs specific segmentation of anatomical structures and/or pathological structures from an input image. Similar to the description above regarding the generated heatmaps vs. annotated heatmaps, the segmentation decoder can be trained on annotated segmentation data, such that loss L IMGen between the decoder's segmentation of a training data input image and the annotated segmentation of the same training data input image is minimized. In this manner, the second encoder 514 , F IM is refined to optimize its output features such that the segmentation of input images 424 is optimized as well.
- Image pathology classifier 534 provides a further layer of regularization to second encoder 514 (much in the same manner to how NLP pathology classifier 532 regularizes the first encoder 512 ).
- image pathology classifier 534 trains and refines second encoder 514 , F IM with respect to the independent pathology classification for input images.
- the independent pathology classification for input images 424 can be thought of as the diagnosis/pathology identification as is actually contained within the input images 424 (i.e. the ground-truth pathology, independent of what the original reviewing physician or radiologist reported that he saw in the same input image 424 ).
- Image pathology classifier 534 consists of classification layers added off of features from the second encoder network 514 , F IM .
- these classification layers can be driven by a binary cross entropy (BCE) loss L IM_Path , which for a given input image 424 , captures the difference between the ground truth pathology in the input image and the pathology in the features generated by second encoder 514 .
- BCE loss is utilized because the second encoder 514 is either correct or incorrect with respect to the ground truth pathology of any given input image, although in some embodiments a non-binary loss could be used to capture pathology classification errors with greater granularity. Regardless of whether a binary loss is utilized or not, by training ML network 500 while also minimizing L IM_Path , second encoder 514 is regularized and refined with respect to its ability to detect pathologies from input images 424 .
- the requisite training data that can be used in conjunction with image pathology classifier 534 can be obtained as annotated radiological images (or annotated sections of radiological images), where the annotations reflect one or more expert opinions (and/or an expert consensus) as to the pathologies that are present in a given radiological image.
- this annotated radiological image pathology training data can be obtained from the expert review previously described with respect to FIGS. 1 and 2 , wherein an expert or secondary reviewing physician analyzes a given radiological image and provides user input corresponding to structured checklist items that pertain to various pathologies.
- these structured checklists can be stored in database 107 and associated with the radiological image from which the structured checklist was generated.
- the pairs of structured checklists and corresponding radiological images can be processed and use to generate training data in response to the structured checklist and corresponding radiological image initially being written to or stored in database 107 . It is also possible that a plurality of structured checklists and their corresponding radiological images be retrieved from database 107 at a later time and then processed into one or more training data sets (and/or validation sets, test sets, etc.).
- first encoder 512 F NLP receives as input radiological report text 422 and outputs one or more embeddings for pathologies and/or diagnosis information within the report text.
- Second encoder 514 F IMI receives as input radiological image regions 424 (corresponding to report text 422 ) and outputs one or more features for pathologies within the image.
- the embeddings from first encoder 512 and the features from second encoder 514 are concatenated to an error encoder 540 , E NLP_IM which is trained to regress to an estimation of diagnostic error by classifying a ⁇ word embedding, image feature ⁇ pair across the output categories 550 a - d (Agree, Overcall, Undercall, Missed).
- ML network 500 can include an additional encoder network (not shown) that is used as a second task or classifier to regress to an estimation of the clinical significance of a diagnostic error classified by error encoder 540 .
- an additional encoder network (not shown) that is used as a second task or classifier to regress to an estimation of the clinical significance of a diagnostic error classified by error encoder 540 .
- E Clin_Sig 750 is depicted in FIG. 7 and is discussed in greater depth with respect to FIG. 7 ).
- the output categories 550 b - d (Overcall, Undercall, Missed) all indicate that a diagnostic error of some sort is present.
- the clinical significance encoder could output a clinical significance score of 0, 1 or 2, where a score of 0 indicates no clinical significance (or no error), a score of 1 indicates a moderate clinical significance, and a score of 2 indicates a high clinical significance.
- the exact outputs of a clinical significance classifier can be determined, modified or otherwise adjusted as desired during the training process of ML network 500 .
- the clinical significance scores can be a range of discrete numbers or can be continuous between a minimum and maximum value.
- the possible range of clinical significance scores might be determined by the definition of clinical significance provided by the overall quality assessment process in which the trained ML network 500 is utilized.
- Error encoder 540 is trained on the basis of a loss L Error_Path , which is used to minimize the difference between a ground truth diagnostic error, Y, and the output diagnostic error generated by error encoder 540 . As depicted in FIG.
- the output diagnostic error generated by error encoder 540 is given by the function E NLP_IM (F NLP (X rep ), F IM (X IM )), where X rep represents the input report text 422 and X IM represents the input image 424 —F NLP (X rep ) represents the word embedding(s) output by first encoder 512 for input report text 422 and F IM (X IM ) represents the image feature(s) output by second encoder 514 for input image 424 .
- error encoder 540 is trained, driven by the loss L Error_Path , to regress to an estimation of diagnostic error, where the estimation is given by one of the output categories 550 a - d .
- Training data for error encoder 540 is based upon annotated input training data pairs comprising radiological report text (such as report text 422 ) and radiological images (such as images 424 ), where the annotation of relevance uses the output categories 550 a - d to label any diagnostic error(s) present in the training data pair.
- the ground truth presence of diagnostic error can be determined by presenting one or more checklist items to the reviewing physician(s) and/or expert(s) that use the checklists to review radiological reports and images.
- checklist items might directly receive user input indicating a diagnostic error falling within one of the output categories 550 a - d .
- the checklist items might also receive user input pertaining to the pathologies present in a radiological image and corresponding report, in which case annotations or labels for the training data can be automatically generated by determining the appropriate output category 550 a - d based on a structured checklist for the radiological image and a checklist for the corresponding report.
- a structured checklist item can be used to obtain user input indicating a degree of clinical significance for a given diagnostic error in a radiological image/report pair.
- a clinical significance checklist item can be presented in line with the aforementioned checklist items, or can be presented at a later time, e.g. in a separate fashion where secondary reviewing physicians/experts are asked only to indicate a clinical significance for already identified diagnostic errors.
- the checklist items and/or user input can include comments about image quality, i.e., a checklist item is a request for comments about the image quality of the particular radiological image with which the checklist is associated. Based on these comments, ML network 500 can assess the quality of the radiological image (e.g., presence of artifacts such as motion and/or blur, noise, bad acquisition protocol, etc.) to determine whether or not the image is acceptable enough for further assessment by downstream portions of ML network 500 .
- the quality of the radiological image e.g., presence of artifacts such as motion and/or blur, noise, bad acquisition protocol, etc.
- the user input of comments reflecting image quality can be provided as a single checklist item allowing for relatively free form entry and identification of artifacts, or the user input can be provided as a comprehensive set of checklist items, e.g., one checklist item for each type of artifact that may or may not be present in the radiological image being reviewed with the checklist.
- a corresponding feature vector of this network can be added to ML network 500 , wherein the features can be reviewed by one or more controllers contained within ML network 500 .
- the overall ML network 500 is trained end-to-end, not to classify the presence or absence of pathologies, but rather to regress to an estimation of the diagnostic errors made in the assessment of radiological images by the reviewing physician.
- the k i weighting factors can be set empirically, can be grid searched for optimization, or some combination of the two can be applied.
- the application of the aggregate loss function L tot_error simultaneously trains ML network 500 to regress to an estimation of diagnostic error, while also regularizing and refining the various individual components such as F NLP (first encoder 512 ), F IM (second encoder 514 ), and E NLP_IM (error encoder 540 ).
- training data generation can leverage already existing radiological images 104 and radiological reports 105 that are stored within database 107 of FIG. 1 .
- the training data generation can leverage various structured data and structured checklists that contain user input provided by secondary reviewers/experts, to thereby generate and apply annotations and labels to raw training data pairs comprising radiological images 104 and their corresponding radiological reports 105 .
- ML network 500 and its associated training can be provided to be backwards compatible with pre-existing radiological practices and databases, providing automated diagnostic quality assessments in a powerful and integrated fashion.
- a limited amount of training data i.e. radiological images and their corresponding radiological reports
- the automated diagnostic quality assessment of the present disclosure can utilize a Siamese-like network, which are functional even when a limited number of training data are available.
- FIG. 6A depicts an architecture diagram for Siamese-based machine learning network 600 to detect diagnostic errors in radiological examinations.
- the Siamese ML network 600 shares architectural similarities with the multi-regularizer ML network 500 of FIG. 5 .
- Siamese ML network 600 includes a first encoder network 612 (F NLP ) and a second encoder network 614 (F IM )—in some embodiments, the first and second encoder networks 612 , 614 can be substantially similar or identical to the first and second encoder networks 512 , 514 of multi-regularizer ML network 500 .
- Siamese ML network 600 includes a concept generator 622 (G C ) and a heatmap generator 624 (G H ), which in some embodiments can be substantially similar or identical to the concept generator 522 and heatmap generator 524 of multi-regularizer ML network 500 .
- Siamese ML network 600 replaces pathology classifiers (i.e. NLP pathology classifier 532 and image pathology classifier 534 ) with a Siamese network 630 .
- pathology classifiers i.e. NLP pathology classifier 532 and image pathology classifier 534
- a Siamese network optimizes the differences between input objects X 1 , X 2 that are either of a similar class or different classes, using an ⁇ h(X 1 ) ⁇ h(X 2 )
- the input objects are the radiological report text embeddings output by first encoder F NLP 612 and the radiological image features output by second encoder F IM 614 .
- Siamese network 630 operates over the ⁇ text embedding, image feature ⁇ pairs generated for each motion segment within an overall diagnostic examination of a patient.
- the particular Siamese function of Siamese network 630 takes the form
- a Siamese Error Encoder E Siamese 636 drives a Siamese loss L Siamese for input training data consisting of ⁇ text embedding, image feature ⁇ pairs.
- the training data is structured such that each training data pair is either of a similar class (i.e., same pathology present in both) or is of a different class (i.e., same pathology is not present in both).
- the degree to which a training data pair agrees or disagrees is indicated by an annotation/label ‘Z’, which can be a binary or continuous variable depending on the manner or extent in which discrepancies are to be encoded.
- the Siamese loss L Siamese minimizes the difference between the label Z (representing the ground truth) and the calculated Siamese difference between the text embedding and the image feature (which is output by Siamese function 630 ).
- the Siamese loss L Siamese forces the ⁇ text embedding, image feature ⁇ pair to be similar, or to have a very small distance between each other.
- the Siamese loss L Siamese tries to separate the two as much as possible, or to increase their separation distance. In this manner, the outputs of first encoder F NLP 612 and second encoder FIM 614 are refined—their embeddings and features are fine-tuned such that they may be better analyzed and classified for the automated assessment of diagnostic quality and error disclosed herein.
- a second encoder labeled here as Error Encoder E NPL_IM 640 regresses to the actual error value for the input pair of report text 422 and radiological image 424 , in the same or similar manner as described above with respect to error encoder 540 of FIG. 5 , and provides the output indicating diagnostic quality (e.g., Agree, Overcall, Undercall, Missed).
- diagnostic quality e.g., Agree, Overcall, Undercall, Missed.
- the k i weighting factors can be set empirically, can be grid searched for optimization, or some combination of the two can be applied. In this manner, the application of the aggregate loss function L tot_error trains Siamese ML network 600 to regress to an estimation of diagnostic error.
- the regression to diagnostic error is computed off of the Siamese function 630 .
- regression to the estimation of diagnostic error can be computed off of F NLP (first encoder 612 ) and F IM (second encoder 614 ), for example as is shown in FIG. 6B .
- the Siamese function 630 is still present in the architecture of FIG. 6B , but no longer couples to error encoder 640 b . Instead, error encoder 640 b couples to F NLP and F IM and receives their respective outputs of text embeddings and image features. During training, the Siamese function 630 and the Siamese loss L Simese nevertheless still influence F NLP and F IM at back propagation, to refine the two encoders based on the Siamese network principles discussed above. Additionally, Siamese Error Encoder E Simese 636 b is configured as a controller to error encoder 640 b .
- error encoder 640 b regresses to the actual diagnostic error output (Agree, Overcall, Undercall, Missed) and in some embodiments, the clinical significance of error (0, 1, 2) as well.
- FIG. 7 depicts a Siamese ML network 700 that has been extended to regress to an estimated clinical significance of error in addition to the estimation of diagnostic error discussed above.
- Siamese ML network 700 adds a clinical significance encoder E Clin_Sig 750 to the multitask network architecture of FIG. 6B , although it is appreciated that the clinical significance encoder 750 could be combined with or added to any of the previously discussed machine learning architectures for automated diagnostic quality assessment without departing from the scope of the present disclosure.
- clinical significance encoder E Clin_Sig 750 is provided as a final regressor with a sigmoid activation function for the clinical significance score.
- the clinical activation score(s) can be obtained from database 107 , and in particular, obtained as generated by the clinical impact weighting logic 320 discussed previously with respect to FIG. 3 .
- the clinical significance score may alternatively or additionally be obtained as a user input to one or more structured checklists, as described above with respect to ML network 500 of FIG. 5 .
- Clinical significance encoder E Clin_Sig 750 takes into account the Siamese input from Siamese error encoder E Siamese 736 as well as the embeddings/features input obtained from first encoder F NLP 712 and second encoder F IM 714 .
- the Siamese input from E Siamese can be utilized as a regularizer for the weight of clinical significance encoder E Clin_Sig 750 and/or can be added to the features generated by E Clin_Sig .
- the features from error encoder E NLP_IM 740 can be aggregated to clinical significance encoder E Clin_Sig 750 in a similar fashion.
- a clinical significance loss L Clin_Sig is added to the total loss function L tot_error .
- L Clin_Sig is used to minimize the difference between a ground truth clinical significance for the diagnostic error present in a training data pair and the computer clinical significance output by clinical significance encoder E Clin_Sig 750 for that same training data pair.
- one or more clinical references such as patient age, weight, history (e.g., of previous related pathologies) can be added to any of the machine learning networks and architectures discussed above, e.g., added as a feature vector to be used in the automated diagnostic quality assessment or provided as a feature matrix.
- a feature vector can be utilized at the image-based assessment section of the machine learning pipeline, to ensure that the clinical information is appropriately utilized.
- the feature vector can be passed as a controller (e.g., a Clinical Controller) and concatenated with the features from F IM , which is the imaging encoder network (represented variously as encoder 514 , 614 , 714 in the discussion above).
- first encoder F NLP ( 512 ), second encoder F IM ( 514 ), and/or error encoder E NLP_IM ( 540 ) of ML network 500 could output an uncertainty along with their respective prediction outputs.
- Methods for determining these uncertainties can include, but are not limited to, evidential deep learning and stochastic weight averaged gaussian approaches.
- the uncertainty associated with the one or more model parameters is assessed and provided as an additional model output, rather than performing a separate or subsequent calculation in order to obtain the uncertainties.
- the output form can express the uncertainty in a raw number, such as a percentage, or as a feature vector, for example.
- Feature vectors generated by the three models i.e., F NLP , F IM , and E NLP_IM
- F NLP , F IM , and E NLP_IM can be utilized by additional downstream components, systems, or networks associated with or otherwise coupled to ML network 500 .
- a threshold or set of thresholds
- Such confidence levels can be used, for example, to define one or more specific workflows, as will be described in greater depth below.
- Examples of such workflows can include, but are not limited to, rerouting uncertain cases for further or expert assessment, selecting specific examples for model fine-tuning (e.g., the generation of augmented training data or other parameter adjustments), and to improve the assessment of physicians' accuracies in delivering their diagnoses.
- Physicians' diagnostic accuracies are quantified based on review data, which may be produced by one or more human experts (i.e., as described previously with respect to FIGS. 1-3 ) and/or which may be produced by one or more machine learning networks or AI models, as is described below.
- review data can be obtained from a computer vision machine learning model, such as second encoder F IM ( 514 ) of FIG. 5 and/or can be obtained from a natural language processing model, such as first encoder F NLP ( 512 ) of FIG. 5 .
- review data capture deterministically or probabilistically the accuracy of the diagnosis that the original reviewing physician made, e.g., in other words, was the diagnosis correct or incorrect, and if incorrect, what type (and/or degree) of error was made?
- the following example and discussion can utilize one or more of the following form fields as predictors for modeling physicians' diagnostic accuracies:
- Each physician's diagnostic accuracy is estimated for each combination of study body part, field name, and pathology that is present in the set of review data.
- the diagnostic accuracies of the physicians belonging to the same practice are correlated.
- the patient age group affects the diagnostic accuracy of each combination of study body part, field name, and pathology. It is appreciated that one or more (or all) of the above-mentioned form fields can be automatically generated from or by ML network 500 and its constituent encoders F NLP , F IM , E NLP_IM and/or its other constituent systems and components, as previously described above with respect to FIGS. 5-7 .
- N reviews the number of reviews.
- each piece of review data (i.e., produced by one or more of ML networks 500 - 70 for pairs of radiological images and the corresponding report text written by the physician reviewing the radiological images) has several associations.
- Each ML review i is associated with:
- a physician (physician: ⁇ 1, 2, . . . , N reviews ⁇ 1, 2, . . . , N physician ⁇ )
- an age group of the patient (age: ⁇ 1, 2, . . . , N reviews ⁇ 1, 2, . . . , N age groups ⁇ )
- a field name (field names: ⁇ 1, 2, . . . , N reviews ⁇ 1, 2, . . . , N field names ⁇ )
- pathology (path: ⁇ 1, 2, . . . , N reviews ⁇ 1, 2, . . . , N pathologies ⁇ )
- each unique concept is represented as a triplet.
- the number of these triplets is denoted as N body parts,field names,pathologies .
- each ML review i is further associated with one of these triplets:
- deterministic review data produced by an ML network/AI, or by expert human reviewer(s) are of the form:
- a generative hierarchical model is formulated for the deterministic review data as follows:
- FIG. 9 depicts a plate notation for the generative hierarchical model described above.
- White (unshaded) circles represent latent variables, while gray (shaded) circles represent observed variables.
- the circular black points represent fixed parameters provided as inputs.
- the directed edges between the fixed parameters and the variables represent dependencies.
- the plates represent variable repetitions.
- the model described above and depicted in FIG. 9 is conditioned on observed data.
- the posterior distribution of the latent variables e.g., one or more of ⁇ l bp_fn_path , ⁇ m,l age,bp_fn_path , ⁇ j,l practice,bp_fn_path , ⁇ 2 , ⁇ k,l physician,bp_fn_path , ⁇ i , p i
- the posterior distribution of the latent variables e.g., one or more of ⁇ l bp_fn_path , ⁇ m,l age,bp_fn_path , ⁇ j,l practice,bp_fn_path , ⁇ 2 , ⁇ k,l physician,bp_fn_path , ⁇ i , p i
- the posterior distribution of the latent variables e.g., one or more of ⁇ l bp_fn_path ,
- analysis of ⁇ m,l age,bp_fn_path allows for the quantification of the effect of patient age on diagnostic accuracy across body part, field name, and pathology combinations.
- ⁇ k,l practice,bp_fn_path the effect of a physician's practice or practice group on diagnostic accuracy can be quantified across body part, field name, and pathology combinations.
- an analysis of ⁇ k,l physician,bp_fn_path can quantify diagnostic accuracies of individual physicians across body part, field name, and pathology combinations.
- probabilistic review data produced by an ML network such as networks 500 - 700 , an AI, and/or one or more expert human reviewers
- an ML network such as networks 500 - 700 , an AI, and/or one or more expert human reviewers
- the vectors y i are treated as parameters defining Dirichlet distributions over probabilities of agree, missed finding, overcall, undercall, and false positive.
- a hierarchical model is formulated for the probabilistic review data as follows:
- the distributions of the latent variables are estimated in order to quantify the certainty about the variables. For example, analysis of ⁇ m,l age,bp_fn_path allows for the quantification of the effect of patient age on diagnostic accuracy across body part, field name, and pathology combinations. Similarly, by analyzing ⁇ k,l practice,bp_fn_path , the effect of a physician's practice or practice group on diagnostic accuracy can be quantified across body part, field name, and pathology combinations. Likewise, an analysis of ⁇ k,l physician,bp_fn_path can quantify diagnostic accuracies of individual physicians across body part, field name, and pathology combinations.
- the techniques described herein are implemented by at least one computing device.
- the techniques may be implemented in whole or in part using a combination of at least one server computer and/or other computing devices that are coupled using a network, such as a packet data network.
- the computing devices may be hard-wired to perform the techniques, or may include digital electronic devices such as at least one application-specific integrated circuit (ASIC) or field programmable gate array (FPGA) that is persistently programmed to perform the techniques, or may include at least one general purpose hardware processor programmed to perform the techniques pursuant to program instructions in firmware, memory, other storage, or a combination.
- ASIC application-specific integrated circuit
- FPGA field programmable gate array
- Such computing devices may also combine custom hard-wired logic, ASICs, or FPGAs with custom programming to accomplish the described techniques.
- the computing devices may be server computers, workstations, personal computers, portable computer systems, handheld devices, mobile computing devices, wearable devices, body mounted or implantable devices, smartphones, smart appliances, internetworking devices, autonomous or semi-autonomous devices such as robots or unmanned ground or aerial vehicles, any other electronic device that incorporates hard-wired and/or program logic to implement the described techniques, one or more virtual computing machines or instances in a data center, and/or a network of server computers and/or personal computers.
- FIG. 8 is a block diagram that illustrates an example computer system with which an embodiment may be implemented.
- a computer system 800 and instructions for implementing the disclosed technologies in hardware, software, or a combination of hardware and software are represented schematically, for example as boxes and circles, at the same level of detail that is commonly used by persons of ordinary skill in the art to which this disclosure pertains for communicating about computer architecture and computer systems implementations.
- Computer system 800 includes an input/output (I/O) subsystem 802 which may include a bus and/or other communication mechanism(s) for communicating information and/or instructions between the components of the computer system 800 over electronic signal paths.
- the I/O subsystem 802 may include an I/O controller, a memory controller and at least one I/O port.
- the electronic signal paths are represented schematically in the drawings, for example as lines, unidirectional arrows, or bidirectional arrows.
- At least one hardware processor 804 is coupled to I/O subsystem 802 for processing information and instructions.
- Hardware processor 804 may include, for example, a general-purpose microprocessor or microcontroller and/or a special-purpose microprocessor such as an embedded system or a graphics processing unit (GPU) or a digital signal processor or ARM processor.
- Processor 804 may comprise an integrated arithmetic logic unit (ALU) or may be coupled to a separate ALU.
- ALU arithmetic logic unit
- Computer system 800 includes one or more units of memory 806 , such as a main memory, which is coupled to I/O subsystem 802 for electronically digitally storing data and instructions to be executed by processor 804 .
- Memory 806 may include volatile memory such as various forms of random-access memory (RAM) or other dynamic storage device.
- RAM random-access memory
- Memory 806 also may be used for storing temporary variables or other intermediate information during execution of instructions to be executed by processor 804 .
- Such instructions when stored in non-transitory computer-readable storage media accessible to processor 804 , can render computer system 800 into a special-purpose machine that is customized to perform the operations specified in the instructions.
- Computer system 800 further includes non-volatile memory such as read only memory (ROM) 808 or other static storage device coupled to I/O subsystem 802 for storing information and instructions for processor 804 .
- the ROM 808 may include various forms of programmable ROM (PROM) such as erasable PROM (EPROM) or electrically erasable PROM (EEPROM).
- a unit of persistent storage 810 may include various forms of non-volatile RAM (NVRAM), such as FLASH memory, or solid-state storage, magnetic disk or optical disk such as CD-ROM or DVD-ROM and may be coupled to I/O subsystem 802 for storing information and instructions.
- Storage 810 is an example of a non-transitory computer-readable medium that may be used to store instructions and data which when executed by the processor 804 cause performing computer-implemented methods to execute the techniques herein.
- the instructions in memory 806 , ROM 808 or storage 810 may comprise one or more sets of instructions that are organized as modules, methods, objects, functions, routines, or calls.
- the instructions may be organized as one or more computer programs, operating system services, or application programs including mobile apps.
- the instructions may comprise an operating system and/or system software; one or more libraries to support multimedia, programming or other functions; data protocol instructions or stacks to implement TCP/IP, HTTP or other communication protocols; file format processing instructions to parse or render files coded using HTML, XML, JPEG, MPEG or PNG; user interface instructions to render or interpret commands for a graphical user interface (GUI), command-line interface or text user interface; application software such as an office suite, internet access applications, design and manufacturing applications, graphics applications, audio applications, software engineering applications, educational applications, games or miscellaneous applications.
- the instructions may implement a web server, web application server or web client.
- the instructions may be organized as a presentation layer, application layer and data storage layer such as a relational database system using structured query language (SQL) or no SQL, an object store, a graph database, a flat file system or other data storage.
- SQL structured query language
- Computer system 800 may be coupled via I/O subsystem 802 to at least one output device 812 .
- output device 812 is a digital computer display. Examples of a display that may be used in various embodiments include a touch screen display or a light-emitting diode (LED) display or a liquid crystal display (LCD) or an e-paper display.
- Computer system 800 may include other type(s) of output devices 812 , alternatively or in addition to a display device. Examples of other output devices 812 include printers, ticket printers, plotters, projectors, sound cards or video cards, speakers, buzzers or piezoelectric devices or other audible devices, lamps or LED or LCD indicators, haptic devices, actuators or servos.
- At least one input device 814 is coupled to I/O subsystem 802 for communicating signals, data, command selections or gestures to processor 804 .
- input devices 814 include touch screens, microphones, still and video digital cameras, alphanumeric and other keys, keypads, keyboards, graphics tablets, image scanners, joysticks, clocks, switches, buttons, dials, slides, and/or various types of sensors such as force sensors, motion sensors, heat sensors, accelerometers, gyroscopes, and inertial measurement unit (IMU) sensors and/or various types of transceivers such as wireless, such as cellular or Wi-Fi, radio frequency (RF) or infrared (IR) transceivers and Global Positioning System (GPS) transceivers.
- RF radio frequency
- IR infrared
- GPS Global Positioning System
- control device 816 may perform cursor control or other automated control functions such as navigation in a graphical interface on a display screen, alternatively or in addition to input functions.
- Control device 816 may be a touchpad, a mouse, a trackball, or cursor direction keys for communicating direction information and command selections to processor 804 and for controlling cursor movement on display 812 .
- the input device may have at least two degrees of freedom in two axes, a first axis (e.g., x) and a second axis (e.g., y), that allows the device to specify positions in a plane.
- An input device 814 may include a combination of multiple different input devices, such as a video camera and a depth sensor.
- computer system 800 may comprise an internet of things (IoT) device in which one or more of the output device 812 , input device 814 , and control device 816 are omitted.
- the input device 814 may comprise one or more cameras, motion detectors, thermometers, microphones, seismic detectors, other sensors or detectors, measurement devices or encoders and the output device 812 may comprise a special-purpose display such as a single-line LED or LCD display, one or more indicators, a display panel, a meter, a valve, a solenoid, an actuator or a servo.
- IoT internet of things
- input device 814 may comprise a global positioning system (GPS) receiver coupled to a GPS module that is capable of triangulating to a plurality of GPS satellites, determining and generating geo-location or position data such as latitude-longitude values for a geophysical location of the computer system 800 .
- Output device 812 may include hardware, software, firmware and interfaces for generating position reporting packets, notifications, pulse or heartbeat signals, or other recurring data transmissions that specify a position of the computer system 800 , alone or in combination with other application-specific data, directed toward host 824 or server 830 .
- Computer system 800 may implement the techniques described herein using customized hard-wired logic, at least one ASIC or FPGA, firmware and/or program instructions or logic which when loaded and used or executed in combination with the computer system causes or programs the computer system to operate as a special-purpose machine. According to one embodiment, the techniques herein are performed by computer system 800 in response to processor 804 executing at least one sequence of at least one instruction contained in main memory 806 . Such instructions may be read into main memory 806 from another storage medium, such as storage 810 . Execution of the sequences of instructions contained in main memory 806 causes processor 804 to perform the process steps described herein. In alternative embodiments, hard-wired circuitry may be used in place of or in combination with software instructions.
- Non-volatile media includes, for example, optical or magnetic disks, such as storage 810 .
- Volatile media includes dynamic memory, such as memory 806 .
- Common forms of storage media include, for example, a hard disk, solid state drive, flash drive, magnetic data storage medium, any optical or physical data storage medium, memory chip, or the like.
- Storage media is distinct from but may be used in conjunction with transmission media.
- Transmission media participates in transferring information between storage media.
- transmission media includes coaxial cables, copper wire and fiber optics, including the wires that comprise a bus of I/O subsystem 802 .
- Transmission media can also take the form of acoustic or light waves, such as those generated during radio-wave and infra-red data communications.
- Various forms of media may be involved in carrying at least one sequence of at least one instruction to processor 804 for execution.
- the instructions may initially be carried on a magnetic disk or solid-state drive of a remote computer.
- the remote computer can load the instructions into its dynamic memory and send the instructions over a communication link such as a fiber optic or coaxial cable or telephone line using a modem.
- a modem or router local to computer system 800 can receive the data on the communication link and convert the data to a format that can be read by computer system 800 .
- a receiver such as a radio frequency antenna or an infrared detector can receive the data carried in a wireless or optical signal and appropriate circuitry can provide the data to I/O subsystem 802 such as place the data on a bus.
- I/O subsystem 802 carries the data to memory 806 , from which processor 804 retrieves and executes the instructions.
- the instructions received by memory 806 may optionally be stored on storage 810 either before or after execution by processor 804 .
- Computer system 800 also includes a communication interface 818 coupled to bus 802 .
- Communication interface 818 provides a two-way data communication coupling to network link(s) 820 that are directly or indirectly connected to at least one communication networks, such as a network 822 or a public or private cloud on the Internet.
- network link(s) 820 may be directly or indirectly connected to at least one communication networks, such as a network 822 or a public or private cloud on the Internet.
- communication interface 818 may be an Ethernet networking interface, integrated-services digital network (ISDN) card, cable modem, satellite modem, or a modem to provide a data communication connection to a corresponding type of communications line, for example an Ethernet cable or a metal cable of any kind or a fiber-optic line or a telephone line.
- Network 822 broadly represents a local area network (LAN), wide-area network (WAN), campus network, internetwork or any combination thereof.
- Communication interface 818 may comprise a LAN card to provide a data communication connection to a compatible LAN, or a cellular radiotelephone interface that is wired to send or receive cellular data according to cellular radiotelephone wireless networking standards, or a satellite radio interface that is wired to send or receive digital data according to satellite wireless networking standards.
- communication interface 818 sends and receives electrical, electromagnetic or optical signals over signal paths that carry digital data streams representing various types of information.
- Network link 820 typically provides electrical, electromagnetic, or optical data communication directly or through at least one network to other data devices, using, for example, satellite, cellular, Wi-Fi, or BLUETOOTH technology.
- network link 820 may provide a connection through a network 822 to a host computer 824 .
- network link 820 may provide a connection through network 822 or to other computing devices via internetworking devices and/or computers that are operated by an Internet Service Provider (ISP) 826 .
- ISP 826 provides data communication services through a world-wide packet data communication network represented as internet 828 .
- a server computer 830 may be coupled to internet 828 .
- Server 830 broadly represents any computer, data center, virtual machine or virtual computing instance with or without a hypervisor, or computer executing a containerized program system such as DOCKER or KUBERNETES.
- Server 830 may represent an electronic digital service that is implemented using more than one computer or instance and that is accessed and used by transmitting web services requests, uniform resource locator (URL) strings with parameters in HTTP payloads, API calls, app services calls, or other service calls.
- URL uniform resource locator
- Computer system 800 and server 830 may form elements of a distributed computing system that includes other computers, a processing cluster, server farm or other organization of computers that cooperate to perform tasks or execute applications or services.
- Server 830 may comprise one or more sets of instructions that are organized as modules, methods, objects, functions, routines, or calls. The instructions may be organized as one or more computer programs, operating system services, or application programs including mobile apps.
- the instructions may comprise an operating system and/or system software; one or more libraries to support multimedia, programming or other functions; data protocol instructions or stacks to implement TCP/IP, HTTP or other communication protocols; file format processing instructions to parse or render files coded using HTML, XML, JPEG, MPEG or PNG; user interface instructions to render or interpret commands for a graphical user interface (GUI), command-line interface or text user interface; application software such as an office suite, internet access applications, design and manufacturing applications, graphics applications, audio applications, software engineering applications, educational applications, games or miscellaneous applications.
- Server 830 may comprise a web application server that hosts a presentation layer, application layer and data storage layer such as a relational database system using structured query language (SQL) or no SQL, an object store, a graph database, a flat file system or other data storage.
- SQL structured query language
- Computer system 800 can send messages and receive data and instructions, including program code, through the network(s), network link 820 and communication interface 818 .
- a server 830 might transmit a requested code for an application program through Internet 828 , ISP 826 , local network 822 and communication interface 818 .
- the received code may be executed by processor 804 as it is received, and/or stored in storage 810 , or other non-volatile storage for later execution.
- the execution of instructions as described in this section may implement a process in the form of an instance of a computer program that is being executed, and consisting of program code and its current activity.
- a process may be made up of multiple threads of execution that execute instructions concurrently.
- a computer program is a passive collection of instructions, while a process may be the actual execution of those instructions.
- Several processes may be associated with the same program; for example, opening up several instances of the same program often means more than one process is being executed. Multitasking may be implemented to allow multiple processes to share processor 804 .
- computer system 800 may be programmed to implement multitasking to allow each processor to switch between tasks that are being executed without having to wait for each task to finish.
- switches may be performed when tasks perform input/output operations, when a task indicates that it can be switched, or on hardware interrupts.
- Time-sharing may be implemented to allow fast response for interactive user applications by rapidly performing context switches to provide the appearance of concurrent execution of multiple processes simultaneously.
- an operating system may prevent direct communication between independent processes, providing strictly mediated and controlled inter-process communication functionality.
- cloud computing is generally used herein to describe a computing model which enables on-demand access to a shared pool of computing resources, such as computer networks, servers, software applications, and services, and which allows for rapid provisioning and release of resources with minimal management effort or service provider interaction.
- a cloud computing environment (sometimes referred to as a cloud environment, or a cloud) can be implemented in a variety of different ways to best suit different requirements.
- a cloud environment in a public cloud environment, the underlying computing infrastructure is owned by an organization that makes its cloud services available to other organizations or to the general public.
- a private cloud environment is generally intended solely for use by, or within, a single organization.
- a community cloud is intended to be shared by several organizations within a community; while a hybrid cloud comprises two or more types of cloud (e.g., private, community, or public) that are bound together by data and application portability.
- a cloud computing model enables some of those responsibilities which previously may have been provided by an organization's own information technology department, to instead be delivered as service layers within a cloud environment, for use by consumers (either within or external to the organization, according to the cloud's public/private nature).
- the precise definition of components or features provided by or within each cloud service layer can vary, but common examples include: Software as a Service (SaaS), in which consumers use software applications that are running upon a cloud infrastructure, while a SaaS provider manages or controls the underlying cloud infrastructure and applications.
- SaaS Software as a Service
- PaaS Platform as a Service
- PaaS Platform as a Service
- PaaS provider manages or controls other aspects of the cloud environment (i.e., everything below the run-time execution environment).
- Infrastructure as a Service IaaS
- IaaS Infrastructure as a Service
- IaaS in which consumers can deploy and run arbitrary software applications, and/or provision processing, storage, networks, and other fundamental computing resources, while an IaaS provider manages or controls the underlying physical cloud infrastructure (i.e., everything below the operating system layer).
- Database as a Service in which consumers use a database server or Database Management System that is running upon a cloud infrastructure, while a DbaaS provider manages or controls the underlying cloud infrastructure, applications, and servers, including one or more database servers.
- DBaaS Database as a Service
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Biomedical Technology (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Medical Informatics (AREA)
- Mathematical Physics (AREA)
- Computational Linguistics (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Quality & Reliability (AREA)
- Databases & Information Systems (AREA)
- Public Health (AREA)
- Radiology & Medical Imaging (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Primary Health Care (AREA)
- Epidemiology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Biology (AREA)
- Pathology (AREA)
- Pure & Applied Mathematics (AREA)
- Mathematical Optimization (AREA)
- Mathematical Analysis (AREA)
- Computational Mathematics (AREA)
- Algebra (AREA)
- Probability & Statistics with Applications (AREA)
- Medical Treatment And Welfare Office Work (AREA)
- Measuring And Recording Apparatus For Diagnosis (AREA)
Abstract
Description
f(R1l, . . . , Rpl|X(R))
f(Dil|{tilde over (R)}il,X(D),θi)f({tilde over (R)}il|Ril,X({tilde over (R)}))
f((Dil|{tilde over (R)}il,X(D),θi)
f({tilde over (R)}il|Ril,X({tilde over (R)}))
gk(R1l,D1l, . . . , Rpl,Dpl|Ek,Yk)
-
- practice: {1, 2, . . . , Nphysicians}→{1, 2, . . . , Npractices}
-
- bp_fn_path: {1, 2, . . . , Nreviews}→{1, 2, . . . , Nbody parts,field names,pathologies
Bayesian Approach using Deterministic Reviews
- bp_fn_path: {1, 2, . . . , Nreviews}→{1, 2, . . . , Nbody parts,field names,pathologies
-
- yi∈{agree, missed finding, overcall, undercall, false positive}
where i=1, 2, . . . , Nreviews. In other words, each review represented in the deterministic review data is classified according to one of the labels/categories above.
- yi∈{agree, missed finding, overcall, undercall, false positive}
-
- βl bp_fn_path˜N(0, 22I)
- where l=1, 2, . . . , Nbody parts,field names,pathologies
- βm,l age,bp_fn_path˜N(0,I)
- where m=1, 2, . . . , Nage groups and l=1, 2, . . . , Nbody parts,field names,pathologies
- βj,l practice,bp_fn_path˜N(0,I)
- where j=1, 2, . . . , Npractice and l=1, 2, . . . ,Nbodyparts,field names,pathologies
- σ2˜Γ−1(3,1)
- βk,l physician,bp_fn_path˜N(βpractice(j),l practice,bp_fn_path, σ2I)
- where k=1, 2, . . . , Nphysicians and l=1, 2, . . . , Nbody parts,field names,pathologies
- γi=βbp_fn_path(i) bp_fn_path+βage(i),bp_fn_path(i) age,bp_fn_path+βphysician(i),bp_fn_path(i) physician,bp_fn_path
- where i=1, 2, . . . , Nreviews
- pi=Softmax((γi T, 0)T
- where i=1, 2, . . . , Nreviews
- γi˜Categorical(pi)
- where i=1, 2 . . . , Nreviews
and where:
- where i=1, 2 . . . , Nreviews
- βl bp_fn_path˜N(0, 22I)
- βl bp_fn_path,βm,l age,bp_fn_path,βj,l practice,bp_fn_path,βk,l practice,bp_fn_path,βk,l physician,bp_fn_path,γi∈ 4, σ2 ∈ >0, and pi ∈Δ5.
-
- βl bp_fn_path˜N(0, 22I)
- where l=1, 2, . . . , Nbody parts,field names,pathologies
- βm,l age,bp_fn_path˜N(0,I)
- where m=1, 2, . . . , Nage groups and l=1, 2, . . . , Nbody parts,field names,pathologies
- βj,l practice,bp_fn_path˜N(0,I)
- where j=1, 2, . . . , Npractice and l=1, 2, . . . ,Nbodyparts,field names,pathologies
- σ2˜Γ−1(3,1)
- βk,l physician,bp_fn_path˜N(βpractice(j),l practice,bp_fn_path, σ2I)
- where k=1, 2, . . . , Nphysicians and l=1, 2, . . . , Nbody parts,field names,pathologies
- γi=βbp_fn_path(i) bp_fn_path+βage(i),bp_fn_path(i) age,bp_fn_path+βphysician(i),bp_fn_path(i) physician,bp_fn_path
- where i=1, 2, . . . , Nreviews
- pi=Softmax((γi T, 0)T where i=1, 2, . . . , Nreviews
- pi˜Dirichlet(yi), where i=1, 2 . . . , Nreviews
and where:
- βl bp_fn_path˜N(0, 22I)
- βl bp_fn_path,βm,l age,bp_fn_path,βj,l practice,bp_fn_path,βk,l practice,bp_fn_path,βk,l physician,bp_fn_path,γi∈ 4, σ2 ∈ >0, and pi ∈Δ5.
Claims (21)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US16/849,442 US11423538B2 (en) | 2019-04-16 | 2020-04-15 | Computer-implemented machine learning for detection and statistical analysis of errors by healthcare providers |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US16/386,006 US11521716B2 (en) | 2019-04-16 | 2019-04-16 | Computer-implemented detection and statistical analysis of errors by healthcare providers |
| US16/849,442 US11423538B2 (en) | 2019-04-16 | 2020-04-15 | Computer-implemented machine learning for detection and statistical analysis of errors by healthcare providers |
Related Parent Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US16/386,006 Continuation-In-Part US11521716B2 (en) | 2019-04-16 | 2019-04-16 | Computer-implemented detection and statistical analysis of errors by healthcare providers |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20200334809A1 US20200334809A1 (en) | 2020-10-22 |
| US11423538B2 true US11423538B2 (en) | 2022-08-23 |
Family
ID=72832661
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US16/849,442 Active 2040-03-04 US11423538B2 (en) | 2019-04-16 | 2020-04-15 | Computer-implemented machine learning for detection and statistical analysis of errors by healthcare providers |
Country Status (1)
| Country | Link |
|---|---|
| US (1) | US11423538B2 (en) |
Families Citing this family (50)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US11640403B2 (en) * | 2019-07-03 | 2023-05-02 | Kpn Innovations, Llc. | Methods and systems for automated analysis of behavior modification data |
| EP3767544A1 (en) * | 2019-07-19 | 2021-01-20 | Robert Bosch GmbH | Training a generative model and a discriminative model |
| US20210019296A1 (en) * | 2019-07-19 | 2021-01-21 | Surescripts, Llc | System and method for data de-duplication and augmentation |
| KR102222011B1 (en) * | 2019-07-29 | 2021-03-04 | 주식회사 코어라인소프트 | Medical image analyzing apparatus and method based on medical use artificial neural network evaluating analysis result thereof |
| US20220013232A1 (en) * | 2020-07-08 | 2022-01-13 | Welch Allyn, Inc. | Artificial intelligence assisted physician skill accreditation |
| CN116134529A (en) * | 2020-07-16 | 2023-05-16 | 皇家飞利浦有限公司 | Automated Deterministic Evaluator for Radiology Reports |
| US11438211B1 (en) * | 2020-07-23 | 2022-09-06 | Marvell Asia Pte Ltd. | Adaptive diagnostics for communication systems |
| WO2022101109A1 (en) * | 2020-11-10 | 2022-05-19 | Koninklijke Philips N.V. | System and method to detect and mitigate commonly missed radiology findings in an emergency department |
| GB202017851D0 (en) * | 2020-11-12 | 2020-12-30 | Volpara Health Tech Limited | Method and system to embed imaging and coomunication meta data in a deep learning network |
| CN112509690B (en) * | 2020-11-30 | 2023-08-04 | 北京百度网讯科技有限公司 | Method, device, device and storage medium for quality control |
| CN112487146B (en) * | 2020-12-02 | 2022-05-31 | 重庆邮电大学 | A method, device and computer equipment for obtaining the focus of disputes in a legal case |
| US12217191B2 (en) * | 2020-12-02 | 2025-02-04 | International Business Machines Corporation | Learning unpaired multimodal feature matching for semi-supervised learning |
| CN112581431B (en) * | 2020-12-07 | 2022-08-23 | 重庆邮电大学 | Method for generating ultrasonic image from ultrasonic diagnosis report based on content maintenance and attention mechanism |
| US11848097B2 (en) * | 2020-12-17 | 2023-12-19 | Evicore Healthcare MSI, LLC | Machine learning models for automated request processing |
| EP4272221B1 (en) * | 2021-01-01 | 2024-04-03 | Koninklijke Philips N.V. | Method and system for facilitating reading of medical images |
| JP7776455B2 (en) * | 2021-01-26 | 2025-11-26 | 富士フイルム株式会社 | Learning device, learning method, trained model, and program |
| CN113192625A (en) * | 2021-03-08 | 2021-07-30 | 北京航空航天大学 | Lung disease auxiliary diagnosis cloud platform based on deep learning |
| CN112966774B (en) * | 2021-03-24 | 2024-05-31 | 黑龙江机智通智能科技有限公司 | A histopathological image classification method based on image Bert |
| CN117157648A (en) * | 2021-03-30 | 2023-12-01 | 西门子工业软件有限公司 | Method and system for detecting false errors on components of an AOI machine inspected board |
| US11782957B2 (en) | 2021-04-08 | 2023-10-10 | Grail, Llc | Systems and methods for automated classification of a document |
| CN113241139B (en) * | 2021-04-21 | 2024-07-19 | 辽宁万象联合医疗科技有限公司 | Artificial intelligence system and method based on child radiological image quality control |
| WO2022226949A1 (en) * | 2021-04-29 | 2022-11-03 | 深圳硅基智控科技有限公司 | Artificial neural network-based identification method and system for tissue lesion identification |
| US12353995B2 (en) * | 2021-05-27 | 2025-07-08 | Sap Se | Determining causality for cloud computing environment controller |
| CN113470777B (en) * | 2021-06-04 | 2024-04-09 | 江苏大学 | Tumor auxiliary diagnosis report generation method and device, electronic equipment and storage medium |
| CN113468804B (en) * | 2021-06-10 | 2023-09-19 | 电子科技大学 | An underground pipeline identification method based on matrix bundles and deep neural network |
| EP4134977A1 (en) * | 2021-08-13 | 2023-02-15 | Siemens Healthcare GmbH | Predicting radiologic findings with machine-learning |
| CN113539408B (en) * | 2021-08-31 | 2022-02-25 | 北京字节跳动网络技术有限公司 | Medical report generation method, training device and training equipment of model |
| CN114241271B (en) * | 2021-11-16 | 2023-04-25 | 电子科技大学 | A method for intelligent diagnosis of mechanical faults based on multi-twin transfer learning and multi-information fusion |
| US20230153663A1 (en) * | 2021-11-18 | 2023-05-18 | Unitedhealth Group Incorporated | Transfer learning techniques for using predictive diagnosis machine learning models to generate consultation recommendation scores |
| US11914635B2 (en) * | 2021-11-19 | 2024-02-27 | Adobe Inc. | Performing image search based on user input using neural networks |
| US20230186052A1 (en) * | 2021-12-15 | 2023-06-15 | Microsoft Technology Licensing, Llc | Siamese neural network model for linking incident management tickets |
| US12183463B2 (en) * | 2022-02-14 | 2024-12-31 | Siemens Healthineers Ag | Generation of findings in radiology reports by machine learning based on impressions |
| CN114782972A (en) * | 2022-04-22 | 2022-07-22 | 广东泓胜科技股份有限公司 | An OCR-based photo classification method, system and related equipment |
| CN117132878A (en) * | 2022-05-16 | 2023-11-28 | 百度在线网络技术(北京)有限公司 | Task processing methods, neural networks and their training methods, equipment and media |
| EP4540735A1 (en) * | 2022-06-14 | 2025-04-23 | Koninklijke Philips N.V. | Reading error reduction by machine learning assisted alternate finding suggestion |
| CN115187130B (en) * | 2022-07-29 | 2023-11-21 | 青岛美迪康数字工程有限公司 | Method and device for judging working efficiency based on mouse movement track |
| CN115495773A (en) * | 2022-08-17 | 2022-12-20 | 翼健(上海)信息科技有限公司 | Method and device for constructing special disease database based on privacy and security computing platform |
| US12014220B2 (en) * | 2022-09-12 | 2024-06-18 | International Business Machines Corporation | Learning-based automatic selection of AI applications |
| CN115206512B (en) * | 2022-09-15 | 2022-11-15 | 武汉大学人民医院(湖北省人民医院) | Hospital information management method and device based on internet of things |
| CN116503872B (en) * | 2023-06-26 | 2023-09-05 | 四川集鲜数智供应链科技有限公司 | Trusted client mining method based on machine learning |
| CN116738323B (en) * | 2023-08-08 | 2023-10-27 | 北京全路通信信号研究设计院集团有限公司 | Fault diagnosis method, device, equipment and medium for railway signal equipment |
| CN117316011A (en) * | 2023-09-21 | 2023-12-29 | 南京宁域智能科技有限公司 | Intelligent virtual simulation training system based on traditional Chinese medicine clinical diagnosis |
| CN117174240B (en) * | 2023-10-26 | 2024-02-09 | 中国科学技术大学 | A medical imaging report generation method based on large model domain migration |
| CN117827898B (en) * | 2023-11-27 | 2025-04-15 | 中国标准化研究院 | A data stream processing method for dynamically updating knowledge graphs |
| WO2025171167A1 (en) * | 2024-02-09 | 2025-08-14 | Board Of Regents Of The University Of Nebraska | Methods for improved artificial intelligence prediction of a diagnosis |
| CN117831757B (en) * | 2024-03-05 | 2024-05-28 | 之江实验室 | Lung cancer diagnosis method and system based on multimodal prior knowledge guidance of pathological CT |
| CN119361166B (en) * | 2024-09-12 | 2025-11-07 | 海南大学 | Senile chronic disease typing method integrating semi-supervised learning and multi-branch neural network |
| CN119226857B (en) * | 2024-09-30 | 2025-04-08 | 南京航空航天大学 | A method and system for analyzing and online monitoring the causes of low-probability aviation unsafe events based on flight data |
| CN119807893B (en) * | 2024-12-12 | 2026-01-16 | 天津大学 | Selective learning method and device for media samples, electronic equipment and storage medium |
| CN120689333B (en) * | 2025-06-23 | 2025-11-28 | 河北医科大学 | Orthopaedics disease auxiliary detection method, device, equipment and medium |
Citations (8)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20090204436A1 (en) * | 2008-02-08 | 2009-08-13 | Premerus, Llc | Method and system for managing medical professionals |
| US20140278448A1 (en) * | 2013-03-12 | 2014-09-18 | Nuance Communications, Inc. | Systems and methods for identifying errors and/or critical results in medical reports |
| WO2017151757A1 (en) | 2016-03-01 | 2017-09-08 | The United States Of America, As Represented By The Secretary, Department Of Health And Human Services | Recurrent neural feedback model for automated image annotation |
| WO2017152121A1 (en) | 2016-03-03 | 2017-09-08 | Geisinger Health System | System and method for automated analysis in medical imaging applications |
| WO2018011432A1 (en) | 2016-07-15 | 2018-01-18 | Koninklijke Philips N.V. | Apparatus for assessing medical device quality |
| US20180137244A1 (en) * | 2016-11-17 | 2018-05-17 | Terarecon, Inc. | Medical image identification and interpretation |
| US20190197358A1 (en) * | 2017-12-21 | 2019-06-27 | International Business Machines Corporation | Generative Adversarial Network Medical Image Generation for Training of a Classifier |
| US20190313963A1 (en) * | 2018-04-17 | 2019-10-17 | VideaHealth, Inc. | Dental Image Feature Detection |
-
2020
- 2020-04-15 US US16/849,442 patent/US11423538B2/en active Active
Patent Citations (8)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20090204436A1 (en) * | 2008-02-08 | 2009-08-13 | Premerus, Llc | Method and system for managing medical professionals |
| US20140278448A1 (en) * | 2013-03-12 | 2014-09-18 | Nuance Communications, Inc. | Systems and methods for identifying errors and/or critical results in medical reports |
| WO2017151757A1 (en) | 2016-03-01 | 2017-09-08 | The United States Of America, As Represented By The Secretary, Department Of Health And Human Services | Recurrent neural feedback model for automated image annotation |
| WO2017152121A1 (en) | 2016-03-03 | 2017-09-08 | Geisinger Health System | System and method for automated analysis in medical imaging applications |
| WO2018011432A1 (en) | 2016-07-15 | 2018-01-18 | Koninklijke Philips N.V. | Apparatus for assessing medical device quality |
| US20180137244A1 (en) * | 2016-11-17 | 2018-05-17 | Terarecon, Inc. | Medical image identification and interpretation |
| US20190197358A1 (en) * | 2017-12-21 | 2019-06-27 | International Business Machines Corporation | Generative Adversarial Network Medical Image Generation for Training of a Classifier |
| US20190313963A1 (en) * | 2018-04-17 | 2019-10-17 | VideaHealth, Inc. | Dental Image Feature Detection |
Non-Patent Citations (4)
| Title |
|---|
| "Predicting Diagnostic Error in Radiology Via Eye-Tracking and Image Analytics: Preliminary Investigation in Mammography", Medical Physicals, vol. 40, No. 10, Sep. 11, 2013. |
| International Preliminary Report on Patentability in related international application No. PCT/2020/028279 dated Jul. 20, 2021. |
| International Search Report and Written Opinion in corresponding international application No. PCT/2020/028279 dated Jun. 24, 2020. |
| Written Opinion of the International Preliminary Examining Authority in corresponding international application No. PCT/2020/028279 dated Oct. 28, 2020. |
Also Published As
| Publication number | Publication date |
|---|---|
| US20200334809A1 (en) | 2020-10-22 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US11423538B2 (en) | Computer-implemented machine learning for detection and statistical analysis of errors by healthcare providers | |
| AU2020260078B2 (en) | Computer-implemented machine learning for detection and statistical analysis of errors by healthcare providers | |
| US11790171B2 (en) | Computer-implemented natural language understanding of medical reports | |
| US11954610B2 (en) | Active surveillance and learning for machine learning model authoring and deployment | |
| US20230120282A1 (en) | Systems and methods for managing autoimmune conditions, disorders and diseases | |
| US11791048B2 (en) | Machine-learning-based healthcare system | |
| US20220059200A1 (en) | Deep-learning systems and methods for medical report generation and anomaly detection | |
| US10984024B2 (en) | Automatic processing of ambiguously labeled data | |
| US12431226B2 (en) | Intelligent generation of personalized CQL artifacts | |
| Ojha et al. | Navigating uncertainty: A user-perspective survey of trustworthiness of ai in healthcare | |
| Dai et al. | Evaluating a natural language processing–driven, ai-assisted international classification of diseases, 10th revision, clinical modification, coding system for diagnosis related groups in a real hospital environment: Algorithm development and validation study | |
| Rehman et al. | Advancement in medical report generation: current practices, challenges, and future directions | |
| Aissaoui Ferhi et al. | Empowering medical diagnosis: a machine learning approach for symptom-based health checker | |
| Nasarian et al. | Designing interpretable ml system to build trust in healthcare: A systematic review of the last decade to proposed responsible clinician-ai-collaboration framework | |
| Madhubala et al. | RETRACTED: Bridging the gap in biomedical information retrieval: Harnessing machine learning for enhanced search results and query semantics | |
| Chand et al. | Building an AI-Driven Symptom Checker Using Python Django for Enhanced Telemedicine Services | |
| Mushtaq et al. | Fetal MoAt Net: a light-weight deep learning model for fetal diagnostic plane classification | |
| Li | MULTIMODAL FUSION FOR EARLY DETECTION OF DEMENTIA USING ELECTRONIC HEALTH RECORDS | |
| BR112021020717B1 (en) | Computer-Implemented Natural Language Understanding of Medical Reports | |
| Hancox | Temporal graph-based convolutional neural networks for electronic health records | |
| Syed | Transforming Medical Imaging with CNN-Based Detection Systems | |
| Farfoura et al. | An Uncertainty-Aware Boosting Ensemble for Parkinson’s Disease Early Detection | |
| Phan et al. | SDCANet: Enhancing Symptoms-Driven Disease Prediction with CNN-Attention Networks | |
| Subedi | MASTER THESIS in Universal Design of ICT |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: APPLICATION DISPATCHED FROM PREEXAM, NOT YET DOCKETED |
|
| AS | Assignment |
Owner name: COVERA HEALTH, NEW YORK Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:VIANU, RON;AIJO, TARMO HENRIK;BROWNING, JAMES ROBERT;AND OTHERS;SIGNING DATES FROM 20200429 TO 20200512;REEL/FRAME:054037/0961 |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
| AS | Assignment |
Owner name: COVERA HEALTH, NEW YORK Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:EAKIN, BRYCE ERON;REEL/FRAME:057439/0992 Effective date: 20210903 |
|
| AS | Assignment |
Owner name: COVERA HEALTH, NEW YORK Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:DUBBIN, GREGORY ALLEN;REEL/FRAME:058055/0781 Effective date: 20200515 |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: AWAITING TC RESP, ISSUE FEE PAYMENT VERIFIED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |