US20170337682A1 - Method and System for Image Registration Using an Intelligent Artificial Agent - Google Patents

Method and System for Image Registration Using an Intelligent Artificial Agent Download PDF

Info

Publication number
US20170337682A1
US20170337682A1 US15/587,094 US201715587094A US2017337682A1 US 20170337682 A1 US20170337682 A1 US 20170337682A1 US 201715587094 A US201715587094 A US 201715587094A US 2017337682 A1 US2017337682 A1 US 2017337682A1
Authority
US
United States
Prior art keywords
image
action
actions
images
transformation parameters
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US15/587,094
Inventor
Rui Liao
Shun Miao
Pierre de Tournemire
Julian Krebs
Li Zhang
Bogdan Georgescu
Sasa Grbic
Florin Cristian Ghesu
Vivek Kumar Singh
Daguang Xu
Tommaso Mansi
Ali Kamen
Dorin Comaniciu
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Siemens Healthcare GmbH
Original Assignee
Siemens Healthcare GmbH
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Siemens Healthcare GmbH filed Critical Siemens Healthcare GmbH
Priority to US15/587,094 priority Critical patent/US20170337682A1/en
Priority to EP17171689.7A priority patent/EP3246875A3/en
Priority to CN201710351955.5A priority patent/CN107403446B/en
Assigned to SIEMENS MEDICAL SOLUTIONS USA, INC. reassignment SIEMENS MEDICAL SOLUTIONS USA, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MIAO, SHUN, DE TOURNEMIRE, Pierre, COMANICIU, DORIN, Ghesu, Florin Cristian, GRBIC, SASA, SINGH, VIVEK KUMAR, KAMEN, ALI, KREBS, JULIAN, MANSI, TOMMASO, ZHANG, LI, GEORGESCU, BOGDAN, LIAO, Rui, XU, Daguang
Assigned to SIEMENS HEALTHCARE GMBH reassignment SIEMENS HEALTHCARE GMBH ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SIEMENS MEDICAL SOLUTIONS USA, INC.
Publication of US20170337682A1 publication Critical patent/US20170337682A1/en
Priority to US16/861,353 priority patent/US11557036B2/en
Priority to US18/064,366 priority patent/US11741605B2/en
Priority to US18/351,548 priority patent/US20230368383A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/33Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
    • G06T7/337Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods involving reference images or patches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/72Signal processing specially adapted for physiological signals or for diagnostic purposes
    • A61B5/7235Details of waveform analysis
    • A61B5/7264Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems
    • A61B5/7267Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems involving training the classification device
    • G06T3/147
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/38Registration of image sequences
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10081Computed x-ray tomography [CT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10088Magnetic resonance imaging [MRI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10104Positron emission tomography [PET]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10132Ultrasound image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Definitions

  • the present invention relates to registration of medical images, and more particularly to computer-based automated medical image registration using an intelligent artificial agent.
  • Medical image registration is an important technology that enables image guided therapy, automated diagnosis imaging, and machine-driven image understanding.
  • the goal of image registration is to recover correspondences between two or more medical images, which can be acquired from different patients/objects, the same patient at different time points, different medical imaging modalities, etc., in order to align or fuse the images.
  • the aligned images can provide important information in various applications.
  • the aligned images can be used to guide minimally invasive therapy by fusing pre-operative volume scans (e.g., MRI) with interventional imaging (e.g., DynaCT or fluoroscopy), or to provide complimentary diagnostic information (e.g., co-registration of cardiac ultrasound and MRI for joint function, flow and substrate analysis) or longitudinal analysis (e.g., radiotherapy monitoring through longitudinal CT scans).
  • pre-operative volume scans e.g., MRI
  • interventional imaging e.g., DynaCT or fluoroscopy
  • diagnostic information e.g., co-registration of cardiac ultrasound and MRI for joint function, flow and substrate analysis
  • longitudinal analysis e.g., radiotherapy monitoring through longitudinal CT scans
  • Image registration problems are typically treated as optimization problems in which a generic matching metric (e.g., Mutual Information, Cross Correlation, etc.) is defined to measure the similarity of image pairs to be registered, and transformation parameters between the two images (e.g., rigid body, affine, deformable) are then estimated by an optimizer (e.g., Simplex, Powell, trust region optimization, etc.) via maximization of the defined matching metric.
  • a generic matching metric e.g., Mutual Information, Cross Correlation, etc.
  • transformation parameters between the two images e.g., rigid body, affine, deformable
  • an optimizer e.g., Simplex, Powell, trust region optimization, etc.
  • maximization of the defined matching metric e.g., Simplex, Powell, trust region optimization, etc.
  • a generic matching metric does not guarantee a good representation of the accuracy of the alignment of the data at hand for all use cases, in all circumstances. That is, the global maximum of the matching metric does not necessarily correspond to the correct alignment of the images, for example when the data is noisy, partially occluded, or with drastic different appearances due to different imaging physics. Therefore, the hand-crafted engineering of the matching metric is often required for the specific registration task, for example, by introducing a task-specific region-of-interest (ROI) or calculating hand-crafted features.
  • ROI region-of-interest
  • a generic matching metric is often non-convex for a given registration task and generic optimizers typically perform poorly on non-convex optimization problems. To avoid being trapped into local minima, prior knowledge is often incorporated to develop optimization schemes for specific registration tasks, for example by prior knowledge-driven initialization/seeding, hierarchical optimization or application specific regularizers.
  • the present invention provides methods and systems for registration of medical images using an intelligent artificial agent.
  • Embodiments of the present invention provide artificial intelligent agents that learn registration tasks from training data, and achieve better and better registration results as more data are used for training.
  • a current state observation of an artificial agent is determined based on a plurality of medical images and current transformation parameters.
  • Action-values are calculated for a plurality of actions available to the artificial agent based on the current state observation using a machine learning based model trained based on a plurality of registered training images, wherein the plurality of actions correspond to predetermined adjustments of the transformation parameters.
  • An action having a highest action-value is selected from the plurality of actions and the transformation parameters are adjusted by the predetermined adjustment corresponding to the selected action.
  • the determining, calculating, and selecting steps are repeated for a plurality of iterations, and the plurality of medical images are registered using final transformation parameters resulting from the plurality of iterations.
  • the machine learning based model used to calculate the action-values based on the current state observation may be a trained deep neural network (DNN).
  • the trained DNN may be trained to predict action-values for the plurality of actions based on a plurality of training image pairs with known ground truth transformation parameters using reinforcement learning in which, for each training image pair, a reward for each action of the plurality of actions at a given state is used to train the DNN to learn an optimal registration policy.
  • the trained DNN may be trained using supervised reinforcement learning, in which the reward for each action of the plurality of actions at a given state is directly calculated based on a decrease in a geodesic distance between the transformation parameters at the given a state and the ground truth transformation parameters for the training image pair resulting from applying the action.
  • the plurality of training image pairs may include training image pairs synthetically generated by artificially de-aligning other training image pairs using randomly generated perturbations to the ground truth transformation parameters of the other training image pairs.
  • the plurality of training image pairs may include training image pairs synthetically generated from other training image pairs to provide dense sampling close to the ground truth transformation parameters by co-deforming the other training image pairs by randomly generated affine transformations within a range of the ground truth transformation parameters.
  • the plurality of training image pairs includes training image pairs synthetically generated by artificially altering image appearance of other training image pairs using synthetic image generators.
  • the transformation may be a rigid-body transformation and the plurality of actions may comprise actions corresponding to predetermined adjustments of translation and rotation rigid-body transformation parameters.
  • the plurality of actions may comprise respective actions corresponding to increasing and decreasing each of the translation and rotation rigid-body parameters by a constant or time-varying step size.
  • the rigid-body transformation may be a 3D rigid-body transformation with rigid-body transformation parameters including three translation parameters and three rotation parameters.
  • the rigid-body transformation may be a 2D rigid-body transformation with rigid-body transformation parameters including two translation parameters and one rotation parameter.
  • the plurality of medical images may include a first medical image and a second medical image and the transformation may be a dense deformation model that models deformations applied to the second medical image.
  • the plurality of actions may comprise actions corresponding to adjustments to parameters of the deformation model.
  • the first and second medical images may be registered by generating a dense deformation field corresponding to final parameters of the deformation model resulting from the plurality of iterations and warping the second medical image to register the second medical image with the first medical image by applying the dense deformation field corresponding to the final parameters of the deformation model.
  • the intelligent artificial agent based image registration may be performed hierarchically using multi-scale image data.
  • the determination of the current state observation, calculation of the action-values, and selection of the action having the highest action value may be repeated using a first image resolution of the plurality of medical images and a first machine learning based model trained using training images at the first image resolution until a first stop condition is reached.
  • the determination of the current state observation, calculation of the action-values, and selection of the action having the highest action value may then be repeated using a second image resolution of the plurality of medical images and a second machine learning based model trained using training images at the second image resolution until a second stop condition is reached.
  • a method for training an intelligent artificial agent to perform image registration may include obtaining training image pairs and generating synthetic training image pairs, defining a state observation input for the artificial agent, defining a set of possible actions available to the artificial agent, defining a reward mechanism for learning a registration policy, and training a deep neural network (DNN) to predict action-values for the set of possible actions based on the state observation input using the training image pairs.
  • DNN deep neural network
  • FIGS. 1A and 1B illustrate examples of medical image registration problems that present challenges to traditional optimization based image registration techniques
  • FIG. 2 illustrates a method for training an intelligent artificial agent to perform image registration according to an embodiment of the present invention
  • FIG. 3 illustrates a method for image registration using an intelligent artificial agent according to an embodiment of the present invention
  • FIG. 4 illustrates a framework for artificial agent image registration using the method of FIG. 3 according to an embodiment of the present invention
  • FIG. 5 illustrates a method for rigid registration of medical images using an intelligent artificial agent according to an embodiment of the present invention
  • FIG. 6 illustrates and exemplary deep neural network (DNN) for rigid 3D/3D medical image registration according to an embodiment of the present invention
  • FIG. 7 illustrates a method of hierarchical image registration using multi-scale image data according to an embodiment of the present invention
  • FIG. 8 illustrates exemplary registration results for spine computed tomography (CT) and cone beam CT (CBCT) registration and cardiac CT and CBCT registration;
  • CT computed tomography
  • CBCT cone beam CT
  • FIG. 9 illustrates a comparison of supervised deep reinforcement learning and unsupervised deep reinforcement learning for 2D/2D spine image registration
  • FIG. 10 illustrates a method for deformable registration of medical images using an intelligent artificial agent according to an embodiment of the present invention
  • FIG. 11 an exemplary idealized deformation trajectory between a moving image and a fixed image
  • FIG. 12 illustrates an exemplary DNN for deformable image registration according to an embodiment of the present invention
  • FIG. 13 illustrates a Gaussian radial basis function (RBF).
  • FIG. 14 illustrates a framework for decoding encoded flow actions output from a trained DNN according to an embodiment of the present invention
  • FIG. 15 illustrates examples of deforming a contour of an organ of interest using Gaussian kernels
  • FIG. 16 illustrates patch extraction from landmark positions of a moving image
  • FIG. 17 illustrates a framework for performing a second stage of a multi-stage deformable registration according to an embodiment of the present invention
  • FIG. 18 illustrates a method of autonomous intelligent artificial agent based cognitive image fusion according to an embodiment of the present invention
  • FIG. 19 illustrates a method of training a DNN for deformable image registration using weakly supervised deep dense correspondence learning according to an embodiment of the present invention
  • FIG. 20 illustrates an exemplary network architecture for a Convolutional Encoder-Decoder Network for dense correspondence estimation
  • FIG. 21 illustrates an exemplary Flow-Net like deep architecture for dense correspondence estimation
  • FIG. 22 illustrates a method for deformable registration of medical images according to an embodiment of the present invention.
  • FIG. 23 is a high-level block diagram of a computer capable of implementing the present invention.
  • the present invention relates to registration of medical images using an intelligent artificial agent.
  • Embodiments of the present invention are described herein to give a visual understanding of the image registration methods.
  • a digital image is often composed of digital representations of one or more objects (or shapes).
  • the digital representation of an object is often described herein in terms of identifying and manipulating the objects.
  • Such manipulations are virtual manipulations accomplished in the memory or other circuitry/hardware of a computer system. Accordingly, is to be understood that embodiments of the present invention may be performed within a computer system using data stored within the computer system or available through a network system.
  • Image registration refers to the alignment of two or more images or transformation of different images into one coordinate system.
  • medical image registration can be used to recover correspondences between two or more images acquired using from different patients, the same patient at different times, different medical imaging modalities (e.g., computed tomography (CT), magnetic resonance imaging (MRI), positron emission tomography (PET), ultrasound, etc.), and/or statistical or geometric models.
  • CT computed tomography
  • MRI magnetic resonance imaging
  • PET positron emission tomography
  • ultrasound etc.
  • the images can be brought into the same coordinate system via various transformation models, such as, but not limited to, rigid-body transformation, affine transformation, parametric splines, and dense motion fields.
  • Medical image registration is utilized in many different applications.
  • the aligned images can provide complimentary information for fusion-based decision making (e.g., cardiac assessment from MR, CT and ultrasound), motion compensation in image reconstruction (e.g., to cope with breathing motion), enabling longitudinal change analysis (e.g., follow-up reading of cancer patients by registering new and past images of the same patient), or guiding minimally invasive therapy (e.g., fusion of preoperative medical image data into the surgical coordinate system for guidance).
  • fusion-based decision making e.g., cardiac assessment from MR, CT and ultrasound
  • motion compensation in image reconstruction e.g., to cope with breathing motion
  • longitudinal change analysis e.g., follow-up reading of cancer patients by registering new and past images of the same patient
  • guiding minimally invasive therapy e.g., fusion of preoperative medical image data into the surgical coordinate system for guidance.
  • FIGS. 1A and 1B illustrate examples of medical image registration problems that present challenges to traditional optimization based image registration techniques.
  • FIG. 1A shows an overlay of a spine CT volume 102 and a Cone-beam CT (CBCT) volume 104 before registration ( 100 ) and after registration ( 105 ), with a large difference in FOV between the CT and CBCT volumes 102 and 104 .
  • FOV field of view
  • FIG. 1B shows a cardiac CT volume 110 with contrast enhanced vessels and a cardiac CBCT volume 112 with soft tissue contrast and streaking artifacts.
  • This use case of registering cardiac CT and cardiac CBCT is very common for cardiac interventions, and is often used in a catheterization laboratory (cath lab) to fuse preoperative information to interventional images for guidance of valve replacement, PCI, and other interventions.
  • a catheterization laboratory cath lab
  • FIG. 1B the differences in the CT and CBCT volumes 110 and 112 due to the streaking artifacts and the difference in appearance of the vessels may result in an inaccurate registration of the CT and CBCT volumes 110 and 112 using traditional optimization based image registration.
  • Embodiments of the present invention reformulate medical image registrations problems by training an intelligent artificial agent to more closely mimic how humans perform image registration as a process of sequential actions of object recognition and manipulation.
  • Embodiments of the present invention train an intelligent artificial agent to register image pairs through a combination of supervised learning and hierarchical image processing.
  • the embodiments can be extended to the registration of groups of images by different strategies, such as extending the number of actions so that there is a set of actions for each image or by running a set of agents together, at the same time and cooperatively.
  • the intelligent artificial agent is implemented on one or more computers or processors by executing computer program instructions (code) loaded into memory.
  • the intelligent artificial agent observes its environment (i.e., the images to be registered) and autonomously acts on that environment to register the images using a registration strategy (herein referred to as a “policy” to be consistent with reinforcement learning terminology) learned using machine learning.
  • the agent can be triggered on demand, or be persistent and always running in the background, ensuring robust image registration even if one of the image moves, for instance due to patient motion in the case of real-time imaging.
  • the input for the intelligent artificial agent (the fixed image and the updated moving image as it is transformed by the current estimate of the transformation parameters) is used to determine its “state”, which corresponds to the current alignment of the raw image pairs.
  • the output of the intelligent artificial agent is an “action” to improve the alignment between the fixed and moving image by changing the parameters that define the transformation between the images.
  • a deep neural network DNN
  • the intelligent artificial agent learns a registration policy via a DNN that maps the current state to the optimal action that best improves the accuracy of the alignment.
  • the trained artificial agent applies the learned registration policy to improve the alignment of the images until it converges to the correct pose.
  • FIG. 2 illustrates a method for training an intelligent artificial agent to perform image registration according to an embodiment of the present invention.
  • one image is designated as the reference image or fixed image and the other image is designated as the moving image or floating image.
  • the reference and moving images to be registered can be denoted as I ref :R N ⁇ R and I mov :R M ⁇ R, respectively, where N and M are the respective dimensionalities of the two images to be registered.
  • N and M are typically equal to 2, 3, or 4 (3D+time).
  • the goal of image registration is to estimate a transformation, denoted as T: R N ⁇ R M , that transforms the moving image I mov to be aligned with the reference image I ref .
  • the transformation in general terms, can represent a curved manifold and can be modeled using K transformation parameters. The goal of the registration then becomes estimating the K transformation parameters such that the similarity between I ref and I mov is maximized.
  • various types of transformation models such as an affine transformation, spline model, or biomechanical model, can be used to model the transformation of the moving image I mov to be aligned with the reference image I ref .
  • a non-parametric deformable registration can be estimated with a dense motion field.
  • motion parameters can be estimated for each pixel/voxel on the grid, and the motion parameters can be the displacement directly, stationary velocities, or kinetic moments, or any subspace representation of those (e.g., PCA modes among others).
  • the method of FIG. 2 trains a deep neural network (DNN) to predict action-values for a set of actions corresponding to the transformation parameters. By parameterizing the action space, the DNN can be trained to perform any registration task using the method of FIG. 2 .
  • DNN deep neural network
  • training images are obtained and/or generated.
  • the DNN can be trained for a particular registration task based on training image pairs (i.e., a reference image I ref and a moving image I mov ) corresponding to the particular registration task and known ground truth transformations between the training image pairs.
  • the training image pairs may be medical images acquired using different imaging modalities (e.g., CT, MRI, ultrasound, PET, etc.), medical images acquired from the same patient (using the same imaging modality) at different times, or medical images acquired from different patients.
  • the training image pairs can be pairs of 3D medical images for 3D/3D registration, pairs of 2D medical images for 2D/2D registration, or a 3D image and a 2D image for 2D/3D registration.
  • the training image pairs can include actual medical images of patients acquired from medical image acquisition devices, such as a CT scanner, MRI scanner, C-arm image acquisition device, ultrasound device, etc. These training image pairs can be obtained by receiving the images from the image acquisition devices used to acquire the images or by loading previously acquired images from a storage or memory of a computer system.
  • the ground truth transformation for these image pairs can be acquired by manual expert annotation or by applying an existing image registration method.
  • the reference image in the training image pair may be replaced with the transformed moving image using the computed transformation resulting from the existing registration method, in order to compensate for inaccuracies in the existing registration method.
  • the training image pairs can also include synthetic images which are generated by augmenting other training images rather than being acquired in a scan of a patient using an image acquisition device. It may be difficult to obtain a large number of labeled training pairs of actual scanned medical images of patients. According to an advantageous embodiment, a large training set can be synthetically generated from a small number of labeled training pairs.
  • each of the aligned training image pairs with a known ground truth transformation can be artificially de-aligned by transforming one of the images with a set of transformations sampled in the space of the transformation parameters. Since the original ground truth transformation is known, each of these transformation results in an additional training image pair for which the ground transformation is also known.
  • each aligned training image pair can be co-deformed by a number of perturbations within specified range of the ground truth transformation parameters to generate additional training image pairs that are close to the ground truth transformation in the transformation parameter space.
  • Additional synthetic training images may also be generated by altering the appearance of one or both images in existing training pairs by adding noise and/or artifacts, or even by simulating other imaging modalities from unpaired training images using synthetic image generators including, but not limited to, physics-based simulators (e.g., MRI simulator, ultrasound simulator) or machine learning based simulators (e.g., training a generative adversarial network (GAN) to simulate an imaging modality from another one, the GAN being trained on paired images).
  • physics-based simulators e.g., MRI simulator, ultrasound simulator
  • machine learning based simulators e.g., training a generative adversarial network (GAN) to simulate an imaging modality from another one, the GAN being trained on paired images.
  • GAN generative adversarial network
  • the state input to the DNN is defined for the particular image registration task for which the DNN is being trained.
  • the state of the current alignment can be computed by comparing the transformed moving image I mov using current transformation parameters (or projections of the transformed moving image in the case of 2D/3D registration) and the reference image I ref using a specified function or operation.
  • the state of the current alignment can be computed by subtracting the transformed moving image I mov using current transformation parameters from the reference image I ref .
  • the reference image I ref and the transformed moving image I mov using current transformation parameters are passed to the neural network which features a multi-channel architecture (for example, in FIG. 12 ). The comparison of the images is then automatically performed and learned by the neural network.
  • a set of possible actions are defined for the registration task.
  • the set of possible actions include a set of actions that alter the transformation parameters.
  • the set of actions can include actions that alter rigid body transformation parameters (e.g., 3 translation parameters and 3 orientation parameters for 3D/3D registration).
  • a deformation model can be used to model the deformation field and the set of actions can include actions that alter parameters of the deformation model used to model the deformation field.
  • the set of possible actions can include actions to change (increase and decrease) individual transformation parameters by a certain step size ⁇ , actions to change multiple (l) transformation parameters simultaneously by certain step sizes ⁇ 1 . . .
  • actions to optimize the step size ⁇ i (i 1 . . . k) for each transformation parameter, an action that determines that the correct registration is achieved and terminates the registration process, and an action that compounds the transformation resulting from the previously selected actions in the curved manifold of registration parameters to ensure that the entire path of corrective actions stay within the manifold. It is to be understood that the present invention is not limited to the above described actions and other actions may be included based on the particular registration task for which the intelligent artificial agent is trained.
  • a reward mechanism is defined.
  • a reward mechanism is defined that assigns a reward for each action, depending on the effect of that action of the resulting alignment of the images. A higher reward is assigned for actions that lead to more accurate alignment, while a lower reward or penalty (negative reward) is assigned for actions that lead to less accurate alignment. Reinforcement learning can then be used to train the agent.
  • the reward mechanism can be defined as the reduction of L2 error in the transformation parameters:
  • m before and m after are the transformation parameters before and after taking the action
  • m gt is the ground truth parameters for the correct registration.
  • the agent training is fully supervised and this type of reward can be thought of as a “GPS”, which directly tells the agent what is the best action to take.
  • GPS global positioning system
  • a DNN is trained to predict action-values for the set of possible actions based on the state input using the training images.
  • a core part of the intelligent agent is a DNN (or possibly multiple DNNs) that takes the state as input and outputs action-values for all possible actions with proper parametrization for the particular image registration task.
  • the action-value for an action can represent the highest possible future reward if an action is taken discounted based on the number of steps to reach the ground truth transformation parameters. With the reward mechanism described above (e.g., reduction of L2 error in the transformation parameters), a higher action-value indicates that the action is driving the registration toward the correct direction (i.e., toward the correct final registration result) and therefore is the preferred action to be taken.
  • the action space can be discrete or continuous with parameterization.
  • the DNN can be trained using a supervised Deep Reinforcement Learning (DRL) technique.
  • Reinforcement Learning is a type of machine learning in which a software based artificial agent uses reward feedback to automatically learn ideal behavior in a specific context and for a specific task.
  • DRL Deep Reinforcement Learning
  • RL is a type of machine learning in which a software based artificial agent uses reward feedback to automatically learn ideal behavior in a specific context and for a specific task.
  • DRL which combines DNNs with RL and has recently been applied to train an artificial agent to play Atari and Go games
  • a policy learning process is formulated as a RL problem and the action-value function (also known as Q function) is estimated following the Bellman equation as an iterative update.
  • the training of the agent is typically unguided and the agent is free to evolve in its environment according to its current policy estimate.
  • the DRL training of the DNN is supervised based on the known ground truth transformations of the training image pairs using a greedy search strategy in which the action-values are approximated analytically.
  • This supervised DRL has an advantage that the Q-network guarantees to converge and the behavior distribution can be arbitrarily smooth without the need to deploy a memory replay strategy, which can have a very high memory footprint.
  • the DNN can be trained in an end-to-end fashion, or can be trained separately for image feature extraction layers and policy learning layers in order to bring more robustness to variations in image content if needed.
  • the trained DNN can be trained to predict the action-values based on the training image pairs using evolution strategies (ES) to learn the parameters (e.g., weights) of the DNN.
  • ES evolution strategies
  • genetic-like algorithms or sampling strategies are used to optimize the parameters of the DNN directly based on the training data.
  • the trained DNN is stored, for example, in memory or storage of a computer system or on a remote “cloud-based” computer system.
  • the trained DNN can then be loaded and used when registering two newly received images to iteratively predict action-values for each of the set of possible actions based on the current state of the images to be registered and execute the selected action having the highest predicted action-value.
  • FIG. 3 illustrates a method for image registration using an intelligent artificial agent according to an embodiment of the present invention.
  • first and second medical images are received.
  • the first and second medical images can be images acquired using different imaging modalities (e.g., CT, MRI, ultrasound, PET, DynaCT, etc.), images of the same patient acquired at different times, or images acquired from different patients.
  • the first and second medical images may both be 3D medical images (volumes) (3D/3D registration), both be 2D medical images (2D/2D registration), or may be one 2D medical image and one 3D medical image (2D/3D registration).
  • the first sand second medical images may also be image sequences (3D+t or 2D+t), which can also be registered using the method of FIG. 3 .
  • the first and second medical images may be received directly from image acquisition devices, such as a CT scanner, MRI scanner, ultrasound device, C-arm image acquisition device, etc., or may be received by loading previously stored medical images from a memory or storage of a computer system or receiving the medical images in an electronic transmission from another computer system.
  • image acquisition devices such as a CT scanner, MRI scanner, ultrasound device, C-arm image acquisition device, etc.
  • One of the medical images is designated as the reference image I ref and the other is designated as the moving image I mov .
  • the current state is determined based on the first and second medical images and the current set of transformation parameters.
  • the current state is input to the trained deep neural network and action-values for the set of possible actions are calculated based on the current state using the trained DNN.
  • the possible actions can correspond to adjustments to the transformation parameters.
  • the action with the highest action-value is selected and performed.
  • the selected action will typically adjust at least one transformation parameter, which changes the transformation of the moving image I mov .
  • a correct registration is achieved when a “stop registration” action is selected as the action with the highest action-score or by comparing an error value between the reference image and the transformed moving image to a predetermined threshold value. If the stop condition is not met, the method returns to step 304 and repeats steps 304 - 310 . Accordingly, steps 304 - 310 are iterated until the stop condition is met. When the stop condition is met, the method proceeds to step 312 . At step 312 , the registration result is output.
  • the moving image is transformed by the final transformation parameters, and the transformed moving image and the reference image can be displayed on a display of a computer system. For example, the transformed moving image can be overlaid on the reference image and be displayed as a fused image.
  • the registered images can also be electronically transmitted to a remote computer system to be displayed on a display of the remote computer system.
  • FIG. 3 only shows outputting the final registration result, it is also possible that the incremental registration results using the current transformation parameters can be output (e.g., displayed on a display of a computer system) as each iteration of steps 304 - 310 is performed. Such incremental results would allow a user to view each adjustment made by the intelligent artificial agent in real-time as the artificial agent is registering the images.
  • FIG. 4 illustrates a framework for artificial agent image registration using the method of FIG. 3 according to an embodiment of the present invention.
  • the state 404 which is an observation of the current alignment of the images 402 , is determined based on the current transformation parameters.
  • the state observation 404 can be an image resulting from subtracting the moving image as transformed by the current transformation parameters from the reference image.
  • the state observation 404 is input to the trained DNN 406 , which acts as the “brain” of the intelligent artificial agent.
  • the trained DNN 406 calculates a respective action-score for each action in a set of possible actions, and the action 408 with the highest action score is selected.
  • the action 408 is performed to adjust at least one of the transformation parameters, and registration is performed with the updated transformation parameters, resulting in an updated state observation 404 .
  • This process is iterated until a stop condition is met. For example, the process can be iterated until a stop registration action is selected or until a predetermined maximum number of iterations are performed.
  • multiple agents can be trained in a coordinated fashion to register multiple objects (e.g., organs, anatomical structures, etc.) in the images simultaneously.
  • each object is aligned independently in each step by a corresponding agent (with a corresponding DNN) trained to register that object, and a master agent is trained to coordinate the actions suggest from the individual object-specific agents and output the optimal overall action given the constraints of the relative positions of the multiple objects to be registered.
  • each agent can be trained to collaborate with other agents and/or unknown teammates.
  • multiple agents can also be trained in a coordinated fashion to register multiple images simultaneously by incorporating the constraints of the relative transformations among the images.
  • a multi-task agent can be trained to perform registration of a variety of images coming from different sources, e.g., varying modalities, imaging protocols, and organs.
  • Transfer and hierarchical reinforcement learning can be applied to train the multi-task agent by transferring the knowledge from agents trained for individual registration tasks (e.g., specific organs) and abstracting the features and policies for registration.
  • New task-specific agents can also be trained more efficiently by transferring the knowledge from previously trained agents on similar registration tasks.
  • an actor-mimic approach can be used to train a single intelligent artificial agent capable of registering multiple different types of image pairs. Let N be the number of registration tasks, between N number of imaging modality pairs. N different artificial agents A i (i ⁇ 1 . .
  • N can be trained as described above, each corresponding to a specific registration task between a specific type of imaging modality pairs.
  • a “mimic” agent A m is then trained whose objective is to be as accurate and robust as all of the artificial agents A i in their respective tasks.
  • a m is trained using N different types of training image pairs and the loss function to minimize is the sum of the least-square differences between the action-values of A m and those of the respective individual agents A i .
  • Other loss functions can be designed to enable domain transfer and task generalization. With such a framework, the mimic agent will learn generic features present in all imaging modalities that will enable the registration task.
  • an intelligent artificial agent can be trained to register the medical images in a coarse-to-fine manner using different transformation models. For example, the intelligent artificial agent may first adjust rigid-body transformation parameters until a first stop condition is reached (e.g., a predetermined number of iterations or a switch transformation model action is selected), then affine transformation parameters until a second stop condition is reached (e.g., a predetermined number of iterations or a switch transformation model action is selected), then polyaffine transformation parameters until a third stop condition is reached (e.g., a predetermined number of iterations or a switch transformation model action is selected), and then dense deformation parameters until a final stop condition is reached (e.g., a predetermined number of iterations or a stop registration action is selected).
  • a first stop condition e.g., a predetermined number of iterations or a switch transformation model action is selected
  • affine transformation parameters until a second stop condition is reached
  • polyaffine transformation parameters until a third stop condition is reached
  • the trained DNN can be trained to calculate the action-values for actions corresponding to adjustments to all of the various transformation models (e.g., rigid, affine, polyaffine, dense) as well as action-values for actions corresponding switching between the various transformation models.
  • the various transformation models e.g., rigid, affine, polyaffine, dense
  • an intelligent artificial agent is used to perform rigid registration of medical images.
  • the rigid registration is described herein for 3D/3D medical image registration. It is to be understood that the methods for rigid registration described herein are not limited to t 3D/3D medical image registration and can be similarly applied to 2D/2D medical image registration as well.
  • the methods for rigid registration described herein register images using a rigid-body transformation including three translation parameters and three rotation parameters. Such methods can be similarly applied to perform registration using other types of parametric transformations, such as a nine parameter affine transformation (three translation parameters, three rotation parameters, and three scale parameters), a 12 parameter affine transformation (three translation parameters, three rotation parameters, three scale parameters, and three shear parameters), or a poly-affine transformation.
  • FIG. 5 illustrates a method for rigid registration of medical images using an intelligent artificial agent according to an embodiment of the present invention.
  • a DNN is trained to predict action-values for a set of actions corresponding to rigid-body transformation parameters.
  • Step 500 is performed in an offline training stage prior to the on-line image registration for newly input/received images performed in steps 502 - 514 .
  • T m [ 1 0 0 t x 0 cos ⁇ ⁇ ⁇ x - sin ⁇ ⁇ ⁇ x t y 0 sin ⁇ ⁇ ⁇ x cos ⁇ ⁇ ⁇ x t z 0 0 0 1 ] ⁇ [ ⁇ cos ⁇ ⁇ ⁇ y 0 sin ⁇ ⁇ ⁇ y 0 0 0 0 0 0 0 - sin ⁇ ⁇ ⁇ y 0 cos ⁇ ⁇ ⁇ y 0 0 0 0 1 ] ⁇ [ ⁇ cos ⁇ ⁇ ⁇ z - sin ⁇ ⁇ ⁇ z 0 0 sin ⁇ ⁇ ⁇ z cos ⁇ ⁇ ⁇ z 0 0 0 1 0 0 0 0 0 1 ] ( 2 )
  • I ref and I mov are anatomically aligned.
  • the problem can be cast as a Markov Decision Process (MDP), and the goal is to train an intelligent artificial agent that learns the policy to register two images by making a series of decisions based on the observation (state) of the current alignment.
  • the current observation, or state s t is the intensity difference between I ref and I mov transformed using the current transformation parameters m t :
  • the agent learns a registration policy using a DNN that maps the current state s t to the optimal action a t,o .
  • the agent applies the learned policy in a sequence of N actions ⁇ a 1,o , . . . , a N,o ⁇ , to improve the alignment until it converges to the correct pose or a maximum number of iterations is reached.
  • the training is supervised by instructing the agent to follow a greedy registration path, mimicking how humans register two objects in a most efficient manner.
  • the optimal action a t,o at time t along the supervised registration path is defined as the action that minimizes the Euclidean distance between m t +a t,i and m g :
  • a t,o min a t,i ⁇ A ⁇ m g ⁇ ( m t +a t,i ) ⁇ 2 . (5)
  • the immediate reward r(s t , a t,i ) for action a t,i is proportional to the decrease of the distance of the transformation parameters to the ground truth pose m g by applying action a t,i ,
  • a DNN is trained to represent the action-value Q-function in Equation (6).
  • the DNN is a deep convolutional neural network CNN).
  • the input to the DNN is the current state s t
  • the output of the DNN has 12 nodes, each corresponding to one of the 12 actions in the action set A
  • the loss function can be defined as:
  • y i (s k ) is the i-th output of the CNN for the k-th sample among M training samples.
  • the DNN e.g., deep CNN
  • the DNN training scheme described herein has various advantages as compared to previous RL and DRL used in gaming applications. For example, the target Q-function described in the present disclosure is given analytically without iterative estimation, such that the network can be trained much more efficiently and with a more stable convergence property.
  • the target Q calculation described herein does not require the exploration history of the agent, meaning that the data can be sampled arbitrarily randomly with little correlation, without the need for a large memory storage required by experience replay.
  • Labeled training image pairs (i.e., training image pairs with known ground truth transformations) of medical image may not be easy to obtain.
  • a limited set of labeled training data can be augmented to generate a large number synthetic training image pairs that can be used as training samples for training the DNN.
  • each of the aligned training image pairs with a known ground truth transformation can be artificially de-aligned by transforming one of the images with a set of transformations sampled in the transformation parameter space. For example, for each aligned training image pair, a predetermined number of rigid-body perturbations can be randomly generated within a predetermined range of each rigid-body transformation parameter.
  • each of these transformation results in an additional training image pair for which the ground transformation is also known.
  • denser sampling of the transformation parameter space close to the ground truth pose m g can be performed in order to train the DNN efficiently and reliably, because the network input-output mapping in this space is more complicated and less smooth as compared to that for the transformation parameter space far away from the ground truth aligned pose.
  • each of the aligned training image pairs can be geometrically co-deformed by affine transformations T A :
  • T A I + [ t 11 t 12 t 13 0 t 21 t 22 t 23 0 t 31 t 32 t 33 0 0 0 0 1 ] ( 10 )
  • FIG. 6 illustrates an exemplary DNN for rigid 3D/3D medical image registration according to an embodiment of the present invention.
  • the DNN 600 is a deep CNN with multiple hidden layers that inputs a 3D volume corresponding to the state of the alignment of the images I r and I f to be registered.
  • I r is the reference image
  • I f is the moving or floating image.
  • the input 3D volume is generated by subtracting I f transformed by the current rigid-body transformation parameters from I r .
  • the DNN 600 has five 3D convolutional layers followed by three fully connected layers, but the present invention is not limited to this particular network structure.
  • the output layer of the DNN 600 has 12 nodes, each corresponding to an action in the action set.
  • the 12 output nodes represent the 12 candidate actions corresponding to adjustments of ⁇ 1 mm for each of the translation parameters [t x , t y , t z ] and ⁇ 1° for the orientation (rotation) parameters [ ⁇ x , ⁇ y , ⁇ z ].
  • the output nodes of the DNN 600 output predicted action-values/rewards for the corresponding actions.
  • the intelligent artificial agent selects the action with the highest predicted action-value and performs the action (i.e., the corresponding adjustment to the transformation) to update the transformation of the moving image I f .
  • the current state is determined based on the current rigid body transformation parameters.
  • the current state is determined by subtracting the moving image transformed by the current rigid-body transformation parameters, I mov (T m t ) from the reference image I ref , i.e., using Equation (3). This results in a 3D state volume s t .
  • the initial state can be calculated by subtracting the moving image I mov from the reference image I ref since no transformation has been applied to the moving image yet.
  • action-values are calculated for each action in the set of possible actions based on the current state using the trained DNN.
  • the set of possible actions includes actions corresponding to adjustments to each of the rigid-body transformation parameters.
  • the set of possible actions can include 12 action actions corresponding to adjustments of ⁇ 1 mm for each of the translation parameters [t x , t y , t z ] and ⁇ 1° for the orientation (rotation) parameters [ ⁇ x , ⁇ y , ⁇ z ].
  • the trained DNN such as the trained DNN 600 shown in FIG. 6 , inputs the current state volume s t and calculates a predicted action-value (Q value) for each action.
  • the action with the highest predicted action value is selected and the rigid-body transformation parameters are updated by performing the selected action.
  • the selected action will adjust one of the translation parameters by ⁇ 1 mm or adjust one of the orientation parameters by ⁇ 1°.
  • the moving image is transformed using the updated rigid-body transformation parameters.
  • FIG. 5 only shows outputting the final registration result
  • the incremental registration results using the current rigid-body transformation parameters can be output (e.g., displayed on a display of a computer system) as each iteration of steps 504 - 512 is performed.
  • Such incremental results would allow a user to view each adjustment made by the intelligent artificial agent to iteratively align the moving image to the fixed image in real-time as the artificial agent is registering the images.
  • the trained first and second DNNs are then used for hierarchical image registration of newly received images.
  • This hierarchical registration can be applied to any type of registration approach (e.g., rigid, affine, deformable) and any dimensionality (e.g., 2D/2D, 3D/3D. etc.).
  • FIG. 7 illustrates a method of hierarchical image registration using multi-scale image data according to an embodiment of the present invention.
  • first and second medical images are received.
  • the first and second medical images can be images acquired using different imaging modalities (e.g., CT, MRI, ultrasound, PET, DynaCT, etc.), images of the same patient acquired at different times, or images acquired from different patients.
  • the first and second medical images may be received directly from image acquisition devices, such as a CT scanner, MRI scanner, ultrasound device, C-arm image acquisition device, etc., or may be received by loading previously stored medical images from a memory or storage of a computer system or receiving the medical images in an electronic transmission from another computer system.
  • One of the medical images is designated as the reference image I ref and the other is designated as the moving image I mov .
  • the method of FIG. 7 is described for 3D/3D registration using a rigid-body transformation, but can be similarly applied for 2D/2D registration and/or for other types of transformation parameterizations.
  • the first and second medical images are down-sampled to generate first and second reduced-resolution images.
  • Various well-known down-sampling techniques can be used to down-sample the first and second medical images.
  • a current reduced resolution state volume is generated using the first and second reduced resolution images and the current transformation parameters.
  • the current reduced resolution state volume can be determined by subtracting the reduced-resolution moving image transformed by the current transformation parameters from the reduced resolution reference image.
  • action-values are calculated for each of the actions in the set of possible actions using the first trained DNN.
  • the set of possible actions can include actions corresponding to adjustments to each of the rigid-body transformation parameters.
  • the set of possible actions can include 12 action actions corresponding to adjustments of ⁇ 1 mm for each of the translation parameters [t x , t y , t z ] and ⁇ 1° for the orientation (rotation) parameters [ ⁇ x , ⁇ y , ⁇ z ]d.
  • the first trained DNN which is trained using down-sampled reduced resolution images, inputs the current reduced-resolution state volume calculated from the first and second reduced resolution images and calculates a predicted action-value (Q value) for each action.
  • the action with the highest action value is selected and the transformation parameters are updated by performing the selected action.
  • step 712 it is determined whether a stop condition is met.
  • the stop condition for registration using the first trained DNN is met when a first predetermined number N 1 of iterations using the first trained DNN have been performed.
  • the stop conditions has not yet been met and the method returns to step 706 and repeats steps 706 - 712 . Accordingly, steps 706 - 712 are repeated until N 1 iterations using the first trained DNN are performed.
  • the stop condition has been met and the method proceeds to step 714 .
  • N 1 200, such that 200 sequential actions are applied based on the reduced resolution images using the first trained DNN in order to roughly align the images.
  • other stop conditions for the registration using the first trained DNN may be used.
  • the stop condition may be met when the registration using first trained DNN achieves a predetermined accuracy threshold (e.g., an error value between the reduced-resolution reference image and the transformed reduced-resolution moving image is less than a threshold), when the registration using the first trained DNN converges (e.g., the transformation parameters converge or an error value between the reduced-resolution reference image and the transformed reduced-resolution moving image converges), or in a case in which the first trained DNN includes a “stop” action, when the “stop” action is the selected as the action having the highest action value.
  • Other stopping criteria such as when loops in actions are detected, may be used as well.
  • a region of interest is defined based on the first trained DNN.
  • the reduced resolution state volume is updated based on the current transformation parameters.
  • a single back-propagation pass is then used to compute the derivative of the sum of the outputs of the layers of the first trained DNN with respect to the input reduced resolution state volume, resulting in a saliency map of the input reduced resolution state image.
  • the resulting saliency map is then thresholded, for example using 95 th percentile thresholding, and the weighted mean is calculated on the thresholded saliency map to localize an ROI to be used for the input to the second trained DNN.
  • a current high-resolution state volume is generated using the first and second medical images and the current transformation parameters and the defined ROI is extracted from the current high-resolution state volume.
  • the current high-resolution state volume can be generate by subtracting the original resolution moving image transformed by the current transformation parameters from the original resolution reference image.
  • the ROI defined step 714 is then extracted from the current high-resolution state volume.
  • action-values are calculated for each of the actions in the set of possible actions using the second trained DNN based on the ROI of the current high-resolution state volume.
  • the set of possible actions can include actions corresponding to adjustments to each of the rigid-body transformation parameters.
  • the set of possible actions can include 12 actions corresponding to adjustments of ⁇ 1 mm for each of the translation parameters [t x , t y , t z ] and ⁇ 1° for the orientation (rotation) parameters [ ⁇ x , ⁇ y , ⁇ z ].
  • the second trained DNN which is trained using high resolution images in a limited FOV, inputs only a portion of the current high-resolution state volume corresponding to the ROI defined in step 714 , and calculates a predicted action-value (Q value) for each action. Since the ROI is defined based on the saliency map of the first trained DNN, the ROI focuses the second trained DNN on the most salient portions of the input volume for image registration.
  • the action with the highest action value is selected and the transformation parameters are updated by performing the selected action.
  • step 722 it is determined whether a stop condition has been met.
  • the stop condition for registration using the second trained DNN is met when a second predetermined number N 2 of iterations have been performed using the second trained DNN.
  • N 2 iterations have not yet been performed using the second trained DNN
  • the stop condition has not been met and the method returns to step 716 and repeats steps 716 - 722 . Accordingly, steps 716 - 722 are repeated until N 2 iterations are performed using the second trained DNN.
  • the stop condition has been met and the method proceeds to step 724 .
  • N 2 100, such that 100 sequential actions are applied using the second trained DNN.
  • the stop condition may be met when the registration using second trained DNN achieves a predetermined accuracy threshold (e.g., an error value between the reference image and the transformed moving image is less than a threshold), when the registration using the second trained DNN converges (e.g., the transformation parameters converge or an error value between the reference image and the transformed moving image converges), or in a case in which the second trained DNN includes a “stop” action, when the “stop” action is the selected as the action having the highest action value.
  • a predetermined accuracy threshold e.g., an error value between the reference image and the transformed moving image is less than a threshold
  • the registration using the second trained DNN converges
  • the second trained DNN includes a “stop” action
  • Other stopping criteria such as when loops in actions are detected, may be used as well.
  • the registration result is output.
  • the first and second medical images are registered by transforming the moving image by the final rigid-body transformation to align the moving image with the reference image.
  • the registered first and second medical images can be output by displaying the registered first and second medical images on a display of a computer system.
  • the registered first and second images can be overlaid and displayed as a fused image.
  • the registered first and second medical images can also be output by electronically transmitting the registered first and second images to a remote computer system to be displayed on a display of the remote computer system.
  • FIG. 7 only shows outputting the final registration result
  • the incremental registration results using the current transformation parameters can be output (e.g., displayed on a display of a computer system) as each iteration of steps 704 - 710 and each iteration of steps 716 - 722 is performed.
  • Such incremental results would allow a user to view each adjustment made by the intelligent artificial agent in real-time as the artificial agent is registering the images
  • the present inventors evaluated the above described embodiments for 3D/3D rigid medical image registration on registration of abdominal spine CT and cone beam CT (CBCT) images and registration of cardiac CT and CBCT images.
  • abdominal spine CT and CBCT registration a main challenge is that CT has a much larger view that CBCT, leading to many local optima in the registration space due to the repetitive nature of the spine.
  • 87 image pairs were used for the abdominal spine CT and CBCT registration, with 82 image pairs used for training and 5 pairs used for testing.
  • cardiac CT and CBCT registration a main challenge is the poor quality of CBCT with severe streaking artifacts and weak soft tissue contrast at the boundary of the object to be registered, i.e., the epicardium.
  • the same network architecture was used for the DNN for both the spine and cardiac applications and both registration layers, as follows.
  • the DNN trained for each application includes five convolutional layers followed by three fully connected layers.
  • the five convolutional layers use 8, 32, 32, 128, and 128 filters, respectively, all with 3 ⁇ 3 ⁇ 3 kernels.
  • the first two convolutional layers are each followed by a 2 ⁇ 2 ⁇ 2 max-pooling layer.
  • the three fully connected have 512, 512, and 64 activation neurons, respectively, and the output has 12 nodes corresponding to the twelve possible actions in A.
  • Each layer is followed by a non-linear rectified layer, and batch normalization is applied to each layer.
  • each training image pair was augmented 64,000 times, leading to >5M training samples for each application.
  • rigid-body perturbation was randomly generated within [ ⁇ 30 mm, ⁇ 30 mm, ⁇ 30 mm, ⁇ 30°, ⁇ 30°, ⁇ 30° ] for the cardiac data and [ ⁇ 30 mm, ⁇ 30 mm, ⁇ 150 mm, ⁇ 30°, ⁇ 30°, ⁇ 30° ] for the spine data to cover the large FOV in the head-foot direction in the spine CT.
  • rigid-body perturbation was randomly generated within [ ⁇ 10 mm, ⁇ 10 mm, ⁇ 10 mm, ⁇ 10°, ⁇ 10°, ⁇ 10° ], in order to more densely sample the transformation parameter space close to the correct alignment.
  • the rigid-body perturbation range was reduced to [ ⁇ 5 mm, ⁇ 5 mm, ⁇ 5 mm, ⁇ 5°, ⁇ 5°, ⁇ 5° ].
  • the DNN both for coarse alignment and for refinement registration
  • the learning rate was 0.00006 with a decay of 0.7 every 1000 mini-batch based back propagations. For each case, the training took approximately 24 hours on a single GeForce Titan X.
  • FIG. 8 illustrates exemplary registration results for spine CT and CBCT registration and cardiac CT and CBCT registration.
  • FIG. 8 shows registration results for three spine CT and CBCT registration examples and two cardiac CT and CBCT registration examples.
  • images 800 and 802 show the difference between the reference and moving images (i.e., the state image) before and after registration, respectively, for a first spine CT and CBCT registration example.
  • Images 810 and 812 show the difference between the reference and moving images before and after registration, respectively, for a second spine CT and CBCT registration example.
  • Images 820 and 822 show the difference between the reference and moving images before and after registration, respectively, for a third spine CT and CBCT registration example.
  • Arrows 805 , 815 , and 825 in the first, second, and third spine CT and CBCT registration examples point to the kidney, the black background outside the imaged volume, and deployed stent graphs, respectively, showing the high robustness of the intelligent agent in dealing with various interfering objects and artifacts during registration.
  • Hierarchical registration using the method of FIG. 7 was applied to each spine example, and boxes 807 , 817 , and 827 show the detected ROI at which the attention was focused in the refinement registration using the second trained DNN for the first, second, and third spine CT and CBCT examples, respectively.
  • images 830 and 832 show the difference between the reference and moving images before and after registration, respectively, for a first cardiac CT and CBCT registration example.
  • Images 834 and 836 show the mesh overlay for the epicardium meshes in the CT and CBCT images before and after registration, respectively, for the first cardiac CT and CBCT registration example.
  • Images 840 and 842 show the difference between the reference and moving images before and after registration, respectively, for a second cardiac CT and CBCT registration example.
  • Images 844 and 846 show the mesh overlay for the epicardium meshes in the CT and CBCT images before and after registration, respectively, for the second cardiac CT and CBCT registration example.
  • the present inventors evaluated the efficiency of the supervised deep RL described herein in comparison with deep RL without supervision on a modified 2D/2D registration problem using the spine data, due to the un-affordable large memory storage required by experience replay in the unsupervised RL for 3D/3D registration.
  • 2D multiplanar reconstruction (MPR) image pairs were extracted from 82 aligned CT and CBCT spine pairs using various MPR cuttings.
  • MPR multiplanar reconstruction
  • SL supervised learning
  • the network architecture was modified from the 3D/3D registration examples to take 128 ⁇ 128 2D images as the input, and the output has six nodes corresponding to 6 possible actions in changing [t x , t y , ⁇ z ].
  • the network architecture and training hyper-parameters were the same for SL and unsupervised RL.
  • the present inventors also compared the training efficiency and accuracy with and without denser sampling of the transformation parameter space close to the ground truth alignments.
  • FIG. 9 illustrates a comparison of supervised deep reinforcement learning (SL) and unsupervised deep reinforcement learning for the 2D/2D spine registration.
  • FIG. 9 illustrates a comparison of supervised deep reinforcement learning (SL) and unsupervised deep reinforcement learning for the 2D/2D spine registration.
  • FIG. 9 shows a plot of success rate versus the number of training steps for the 2D/2D spine registration using supervised learning (SL) 902 , supervised learning without denser sampling 904 , and unsupervised RL 906 .
  • SL supervised learning
  • supervised learning without denser sampling 904 supervised learning without denser sampling
  • unsupervised RL 906 unsupervised RL
  • an intelligent artificial agent is used to perform deformable registration of medical images.
  • a major difficulty in medical image registration is incorporating non-rigid deformations, which frequently occur in medical imaging due to different patient positioning, pathological and physiological changes (e.g. empty vs. filled bladder), and/or cardiac and respiratory motion.
  • Current deformable registration algorithms based on variational approaches try to counter such problems in finding a transformation that maps one image to the other by solving an objective function typically consisting of a similarity measure (e.g., SSD, mutual information) and differently motivated regularizers (e.g., elastic, fluid, curvature-based models).
  • a similarity measure e.g., SSD, mutual information
  • regularizers e.g., elastic, fluid, curvature-based models
  • Embodiments of the present invention can overcome these limitations through learning the intrinsic and complex mapping in the spatial and intensity domain.
  • Embodiments of the present invention utilize artificial deep neural networks to extract features from both objects/images to be registered and compare them to assess the similarity, followed by a prediction of an optimal way to change the features of one image should be changed to get closer to the other image. This pattern is applied in a step-by-step approach to approximate the registration in a coarse-to-fine manner.
  • Embodiments of the present invention combine an efficient encoding/decoding of the deformation field with an intelligent artificial agent based trajectory learning that predicts parametric actions operating on the encoded parameters.
  • FIG. 10 illustrates a method for deformable registration of medical images using an intelligent artificial agent according to an embodiment of the present invention.
  • the method of FIG. 10 can be applied to any deformable (non-rigid) registration task.
  • the method of FIG. 10 can be used for applications including, but not limited to, contour propagation between different medical imaging modalities or atlas-based segmentation (in which segmentation is performed by registering an atlas to a medical image a patient) for therapy planning.
  • the method of FIG. 10 is not engineered for a specific application and can be used to perform deformable registration in problematic cases, such as cases involving large deformations or sliding motion.
  • a DNN is trained to predict action-values for a set of actions corresponding to parameters of a deformation model used to encode a deformation field.
  • Step 1000 is performed in an offline training stage prior to the on-line image registration for newly input/received images performed in steps 1002 - 1014 .
  • the DNN is trained to predict action-values for the actions based on a reward function that evaluates the current state of the system.
  • supervised trajectory learning is used to train the DNN.
  • ground truth flow field image deformation field
  • supervised trajectory learning is used to train the DNN to learn steps on the trajectory of small deformations to reach the final deformation field and reward actions corresponding to the small deformations on the learned trajectory.
  • ⁇ DF small deformation flow field
  • the action selection can be performed in a probabilistic way to enable the intelligent artificial agent to not always select the best action in a specific state and allow circuitous routes to avoid local minima ( ⁇ -greedy selection).
  • the probability of each action being selected is based on the action-value (reward) predicted for that action by the DNN.
  • the composition of all of the ⁇ DFs over all of the iterations results in the final flow field that warps the moving image to register the moving image to the fixed image.
  • FIG. 11 an exemplary idealized deformation trajectory between a moving image M and a fixed image F. As shown in FIG.
  • the artificial intelligent agent selects an action based on the current reward determined by the DNN and applies the ⁇ DF corresponding to the selected action to warp the moving image M.
  • the moving image is registered to the fixed image F over four iterations, and the composition of ⁇ DF 1 , ⁇ DF 2 , ⁇ DF 3 , and ⁇ DF 4 result in the final flow field that registers the moving image M with the fixed image F.
  • FIG. 12 illustrates an exemplary DNN for deformable image registration according to an embodiment of the present invention.
  • the DNN 1200 inputs a fixed image F and a moving image M.
  • the DNN 1200 takes 64 ⁇ 64 2D images as input, but the present invention is not limited thereto.
  • Features of both images are extracted with separate but identical, in points of network architecture, processing streams.
  • the CNN 1200 uses multiple convolutional layers, whose weights are not shared between the fixed/reference image stream and the moving image stream, for feature extraction. That is separate weights are learned for the convolutional layers that extract features from the fixed image F and the convolutional layers that extracted features from the moving image M.
  • FIG. 12 illustrates an exemplary DNN for deformable image registration according to an embodiment of the present invention.
  • the CNN 1200 includes three convolutional layers that use 128, 256, and 512 filters, respectively, in each of the fixed image stream and the moving image stream.
  • the convolutional layers used for feature extraction are followed by fully connected layers, at which the extracted features from the fixed and moving images are compared at a higher level to compute reward values for various possible actions.
  • the fully connected layers act as a learned similarity measure based on the features extracted from the fixed and moving images.
  • the CNN 1200 includes three fully connected layers having 500, 128, and 64 activation neurons, respectively.
  • An action prediction is retrieved/selected based on the action-values (rewards) output by the DNN 1200 , and the action prediction can then be transferred to a low-dimensional deformation encoding.
  • Dense prediction follows from decoding the few parameters of the encoded action to a delta dense deformation field ( ⁇ DF) with help from a chose parametric deformation model.
  • the ⁇ DF is then directly applied on the moving image to warp the moving image.
  • the reward may be computed directly at the level of encoded components.
  • the DNN framework is trained to predict the next best action based on the current state (determined by the feature extraction from the fixed and moving images) from available actions corresponding to +/ ⁇ adjustments on the encoded parameters (i.e., the parameters of the parametric deformation model).
  • Coarse-to-fine support may be given by joining actions that are “similar” in parametric space to enable faster convergence, avoid unrealistic states, and reduce dimensionality of otherwise huge parametric spaces.
  • the loss function can be computed as:
  • s is the current state
  • a is the action
  • s′ is the next state if the agent takes action a from state s
  • r GT is the known ground truth reward
  • R is the learned reward computed by the trained DNN
  • B is the number of landmarks or mesh points on the contour or surface being deformed.
  • the DNN can be trained using a gradient descent algorithm and backpropagation to learn weights for layers of the DNN that minimize the loss function over all of the training samples.
  • a deformation model is used to encode the dense deformation field in order to reduce the dimensionality of the dense deformation field.
  • a parametric deformation model whose parameters are applicable for smooth transitions from one state to its neighbor can be used to encode the dense deformation field.
  • Radial Basis Functions are an example of such a parametric deformation model that can be used to encode dense deformation field.
  • FIG. 13 illustrates a Gaussian RBF. As shown in FIG. 13 , the Gaussian RBF 1302 can be expressed as
  • ⁇ ⁇ ( x ) d * e - ( x - p x ) 2 + ( x - p y ) 2 2 * ⁇ 2 .
  • multiple Gaussian RBFs can be used to approximate the dense deformation field for the moving image.
  • the set of possible actions can correspond to adjustments to the Gaussian parameters, such as location and magnitude, for each of the Gaussian RBFs.
  • a free form deformation (FFD) approach can be used to encode the dense deformation field.
  • the DDN learns the desired action for moving a limited number of points of a uniform hull while preserving topology. Dense deformation inside the whole is approximated, for example using B-splines or thin plate splines (TPS).
  • TPS thin plate splines
  • the agent can learn actions that directly modify the B-splines parameters (e.g., direction, amplitude, and extent of the displacement at the B-spline node).
  • all B-spline parameters are projected onto a subspace representation (e.g. via PCA or any other manifold learning method) and the agent directly modify the transformation parameters in the subspace (e.g. by normalized increment/decrement along a PCA mode).
  • various types of manifolds that reduce the dimensionality of the action space through encoding the dense deformation field can be used as the deformation model.
  • the action space of deformations is learned from a set of training examples using manifold learning (linear or non-linear). Multiple possibilities exist for this purpose, such as principal component analysis (PCA), stacked denoising autoencoder, or deep learning based generative models.
  • PCA principal component analysis
  • the agent actions are performed directly on the subspace parameters (e.g., via PCA or any other manifold learning method).
  • the agent directly modifies the transformation parameters in the subspace (e.g., by normalized increment/decrement along a PCA mode) and the resulting deformation field is reconstructed from the updated subspace parameterization.
  • the ground truth ⁇ DFs can be obtained from existing deformable registration algorithms.
  • the algorithm's output can be used to produce a new warped version of the fixed image (generated by warping the moving image by the deformation field output by the algorithm) which can be used in the training instead of the original fixed image.
  • the new fixed training image may be different that the original fixe training image, but by applying this scheme it is guaranteed that the real ground truth deformation field is used for training, as long as any reasonable deformation field is provided by the existing deformable registration algorithm used.
  • the deformation model is a correspondence based model.
  • the deformation model encodes a sparse deformation field corresponding to deformations of the particular points (e.g., landmarks or control points) or structures in the moving image that having correspondences in the fixed image.
  • point correspondences or segmentations of deforming structures in the images are also used for training. The segmentations can be performed manually, automatically, or semi-automatically using various segmentation algorithms. In the case of precise segmentations, exact point correspondences exist and can be used as the ground truth spare deformation field for training.
  • a ray-shooting algorithm Given a segmentation mask, in order to retrieve distributed points on the surface of a convex structure of interest, a ray-shooting algorithm can be performed that shoots rays from the center of mass to the surface with equal angular distances and selects the points where the rays leave the mask as landmark positions. By applying this procedure, the topology of landmarks is guaranteed. If the landmark positions are extracted in the same way across the whole training data set, the DNN learns the topology in training and may learn to connect rewards to specific landmarks.
  • FIG. 14 illustrates a framework for decoding encoded flow actions output from a trained DNN according to an embodiment of the present invention.
  • a fixed image F and a moving image M are input to the trained DNN 1400 .
  • the DNN 1400 extracts features from the fixed and moving images and predicts best flow action based on the current state of the fixed and moving images.
  • a deformation model is selected for flow encoding, and the trained DNN 1400 outputs encoded flow actions 1402 in the parameter space/component space of the deformation model used for the flow encoding.
  • the encoded flow action 1402 performed to adjust the encoded flow 1404
  • flow decoding is performed to decode the encoded flow 1404 into a decoded flow field 1406 .
  • Path A represents the case in which the deformation model is a dense model.
  • the decoded flow field 1406 is a dense deformation field and is applied to the moving image generate the warped moving image 1408 , which is input to the DNN 1400 as the moving image M for the next iteration.
  • Path B represents the case in which the deformation model is correspondence based model.
  • the decoded flow field is used to transform a mesh ( 1410 ) of a segmented structure in the moving image.
  • a dense deformation field is then computed based on the deformation of the mesh using interpolation and/or extrapolation ( 1412 ), and the dense deformation field is applied to the moving image to generate the warped moving image 1408 , which is input the DNN 1400 as the moving image M for the next iteration.
  • Gaussian kernels can be placed at landmark positions in the moving image, such as at control points on a contour or mesh of a segmented organ of interest, and the agent can be driven to align the contours of the of the organ of interest as in an FFD approach.
  • FIG. 15 illustrates examples of deforming a contour of an organ of interest using Gaussian kernels. As shown in example 1500 of FIG. 15 , contour F represents a contour of a segmented target organ in the fixed image and contour M represents a contour of the segmented target organ in the moving images.
  • Gaussian kernels are placed at each of a plurality of landmarks 1501 , 1502 , 1503 , 1504 , 1505 , 1506 , 1507 , and 1508 and the parameters of the Gaussian kernels are adjusted to iteratively deform contour M to align contour M with contour F. Coarse-to-fine registration can be reached by combining or joining neighboring Gaussian kernels and moving them together, such as in Example 1510 of FIG. 15 . That is when an action is selected to move a Gaussian kernel at one landmark point, the Gaussian kernels at neighboring landmark points may be moved as well.
  • the magnitude of movement can be controlled by a separate Gaussian process such that movements of the neighboring Gaussian kernels decrease with a distance to the chose Gaussian kernel.
  • Neighbor relationships can be fixed by predefining the topology.
  • the Gaussian kernels may be moved without the joint movement of the neighboring Gaussian kernels. As shown in example 1510 of FIG.
  • the Gaussian kernels at landmarks 1512 and 1513 are each moved by 0.72 mm and the Gaussian kernels at landmarks 1514 and 1515 are each moved by 0.26 mm.
  • the same action only causes the Gaussian kernel at landmark 1511 to move 1 mm.
  • a fixed step size can be used for action values on a pixel level (e.g., +/ ⁇ 1 pixel movements) that can be adjusted through the above described joint action method.
  • the step size can be used to control convergence speed in contrast to precision. Bigger step sizes may be used in early deformations to roughly approach the fixed structure and then smaller step sizes can be used combined with smaller impact on the neighbors and joining less actions to reach a finer registration.
  • adjusted step sizes may be applied in relation to the reward either using a fixed set of step sizes or using continuous action spaces.
  • a fixed number e.g., +/ ⁇ 1 can be used to increase or decrease the components action values.
  • a greedy agent exploration scheme may be computationally inefficient and lead to long training times
  • sparse exploration schemes can be used. This can be achieved by applying a hierarchical approach on the parameters (multi-scale parameter representation). Different DNNs are trained with different step sizes. The first DNN with largest step size starts, the second takes over, and so on. Alternatively, an actor-critic architecture can be applied, where the actor (a first trained DNN) decides which action to take and the critic (a second trained DNN) evaluates the value of the action. This makes the training of continuous step sizes possible, as the two deep neural networks are trained iteratively, one to pick the action, the second one to choose the step size.
  • the reward function may be based on human-level recognition or may be a reward that is directly obtained from available ground truth deformation fields.
  • Direct reward functions require the knowledge of the ground truth deformation field during training. As such ground truth data is not typically available for real medical image data, artificially generated deformation fields which are generated under controlled conditions can be used instead.
  • the distance of encoded components ⁇ c gt ⁇ c current ⁇ 2 ⁇ c gt ⁇ c next ⁇ 2 can be directly used as the reward, where c gt is the ground truth encoded components or parameters of the deformation model.
  • Reward functions can also be designed based on human-level recognition, that is, the deformation is considered satisfactory when the image content (anatomic structures and boundaries) are well aligned.
  • 2 can be used as the reward function.
  • landmark/mesh differences: ⁇ P ref ⁇ P mov(t) ⁇ 2 can be used as the reward function, where P ref denotes a set of landmarks or mesh points in the reference image and P mov(t) denotes the set of landmarks or mesh points in the current moving image.
  • I ref ⁇ I mov(t) ) can be used as the reward function, where ⁇ is a weighting factor.
  • a deep learned similarity measure sim(I ref ⁇ I mov(t) can be used as the reward function.
  • a pre-processing stage may be performed before proceeding to step 1004 .
  • one or more anatomical objects of interest e.g., organs
  • a landmark or point of interest such as a center of mass of an organ of interest, can be detected in each of the first and second images, and the first and second images can be initially aligned with respect to the point of interest.
  • action values for possible actions can be calculated based on the current by the trained DNN.
  • the trained DNN inputs the fixed image and the current moving image which is warped based on a current dense deformation field. In the first iteration, the current moving image is not yet warped.
  • the trained DNN extracts features from the fixed and moving images, which provide an observation of the current state, and computes action-values for actions corresponding to adjustments of parameters in a parameter space of a deformation model used to encode the deformation of the moving image.
  • the trained DNN may apply a probabilistic model to determine the action-values for the set of actions.
  • an action is selected based on the action-values calculated by the trained DNN.
  • the action corresponds to an adjustment to a parameter in the parameter space of the deformation model used to encode the deformation of the moving image.
  • the action with the highest predicted action-value may be selected.
  • a delta dense deformation field ⁇ DF is computed based on the selected action and the deformable model used to encode the deformation of the moving image.
  • the ⁇ DF is a small dense deformation field corresponding to the selected action that provides an incremental adjustment to the overall deformation of the moving image.
  • the deformation model used to encode the deformation of the moving image is a dense model that represents an entire deformation field of the moving image
  • the adjustment to the parameter in the parameter space of the deformation model corresponding to the selected action is performed and the change in the parameter space of the deformation model is decoded which maps the change in the parameter space of the deformation model to the ⁇ DF.
  • the deformation model used to encode the deformation of the moving image is a correspondence-based model that models deformation of landmark or control points of a structure of interest in the moving image.
  • the adjustment to the parameter in the parameter space of the deformation model corresponding to the selected action is performed and the change is the parameter space of the deformation model is decoded, which maps the change in the parameter space of the deformation model to movement of the landmarks or control points to deform a contour or mesh of the structure of interest in the moving image.
  • the ⁇ DF for the entire moving image is then calculated from the deformation of the contour or mesh in the moving image, for example using interpolation and/or extrapolation.
  • the delta dense deformation field ⁇ DF is applied to the moving image to warp the moving image.
  • the registered first and second medical images can be output by displaying the registered first and second medical images on a display of a computer system.
  • the registered first and second medical images can be displayed in the same coordinate system.
  • the registered first and second images can be overlaid and displayed as a fused image.
  • the final dense deformation field used to warp the moving image which is a composition of all the ⁇ applied to the moving image, can be visualized and displayed on the display of the computer system.
  • the individual ⁇ DFs applied to the moving image can also be visualized and displayed.
  • the registered first and second medical images, as well as the final dense deformation field and/or the individual ⁇ DFs can also be output by being electronically transmitting to a remote computer system to be displayed on a display of the remote computer system.
  • FIG. 10 only shows outputting the final registration result
  • the incremental registration results resulting from warping the moving image using the current dense deformation field can be output (e.g., displayed on a display of a computer system) as each iteration of steps 1004 - 1012 is performed.
  • Such incremental results would allow a user to view each adjustment made by the intelligent artificial agent to the deformation of the moving image in real-time as the artificial agent iteratively warps the moving image to register the moving image to the fixed image.
  • the point correspondence between the fixed and the moving images can be established using a ray-shooting algorithm that, given a segmentation mask of an object of interest shoots rays from the center of mass of the object to the surface with equal angular distances and selects the points where the rays leave the segmentation mask and landmark positions.
  • This ray-shooting algorithm can be performed in the pre-processing phase prior to performing the registration.
  • these landmarks are not anatomically distinct points and are more utilized as regions for which actions can be selected by the trained DNN, this approach can be extended to shoot rays multiple times for the same landmark region with slightly varying angles from the center. That is, in a possible embodiment, the registration using the method of FIG.
  • the 10 can be performed multiple times using different landmarks acquired by ray-shooting with different angles to drive the registration to provide better representation of the shape of the object of interest in between the original landmark locations.
  • the registration results from the various registrations can then be combined. After running the registration process individually for each set of landmarks, multiples of the number of landmarks can be retrieved as point correspondences, which can result in a more accurate registration.
  • a hierarchical (multi-stage) framework based on higher resolution image patches may be utilized to perform deformable image registration.
  • the first stage of the multi-stage deformable image registration method is performed on the full moving and fixed images using a first trained DNN.
  • the first stage can be performed as describe above in the method of FIG. 10 .
  • a second stage performs refinement then performs refinement of the registration estimated in the first stage based on higher resolution image patches extracted from the fixed and moving images using a second trained DNN.
  • patches centered at the new landmark positions of the moving image are extracted from higher resolution versions of the fixed image and the current moving image.
  • the higher resolution versions of the fixed and moving images can high-resolution images generated by up-sampling the fixed and moving images.
  • the first stage can be performed on reduced resolution fixed and moving images generated by down-sampling the fixed and moving images acquired from the image acquisition devices, and the higher resolution versions are the original resolution fixed and moving images acquired from the image acquisition devices.
  • FIG. 16 illustrates patch extraction from landmark positions of the moving image. As shown in FIG. 16 , patches 1602 , 1604 , 1606 , 1608 , and 1610 are extracted at landmark locations in the moving image. Each of the image patches 1602 , 1604 , 1606 , 1608 , and 1610 is centered at the respective landmark and is oriented to be aligned with a normal of the contour at the respective landmark location. The image patches are extracted from the same locations in the fixed image.
  • FIG. 17 illustrates a framework for performing the second stage of the multi-stage deformable registration according to an embodiment of the present invention.
  • the image patches (F) extracted from the higher resolution fixed image and the image patches (M) extracted from the higher resolution moving image are input to a trained DNN 1700 . All of the patches are input to the DNN 1700 in different channels.
  • the deformation model, reward system, and action definitions can stay the same as in the first stage.
  • the image warping is executed on a higher resolution. As the network is likely slower due to patch extraction, multiple channels, and higher resolution warping, in an advantageous embodiment, the method of FIG. 17 is used as a refinement step after the first stage is used to perform rough registration.
  • the DNN 1700 extracts features from the input higher resolution image patches and calculates rewards for the set of actions.
  • the best action is selected and the higher resolution moving image is warped based on the selected action, which results in movement of one or more of the landmark positions.
  • Updated image patches are obtained from the higher resolution moving image at the locations of the updated landmark locations in the moving image, and updated image patches at the same locations are obtained from the higher resolution fixed image. This process then repeats until a stop condition is reached.
  • the step size for the actions can also be increased to reduce the number of steps at the higher resolution or decreased to reach a higher accuracy.
  • FIG. 18 illustrates a method of autonomous artificial intelligence (AI)-based image registration according to an embodiment of the present invention.
  • the method of FIG. 18 utilizes various intelligent artificial agents to perform various tasks related to image registration.
  • Each intelligent artificial agent is implemented by one or more processors of one or more computer systems executing computer program instructions defining operations the artificial agent.
  • each artificial agent autonomously observes a state of its environment and performs actions learned via artificial intelligence/machine learning techniques.
  • the medical images may be 2D or 3D images acquired using any medical imaging modality.
  • the medical images may be received directly from one or more medical image acquisition devices, received by loading previously stored medical images from a memory or storage of a computer system, or received via one or more electronic transmissions from a remote computer system.
  • an imaging modality is detected for each of the medical images to be fused.
  • a trained intelligent artificial agent automatically detects the imaging modality (e.g., CT, MRI, Ultrasound, DynaCT, PET, etc.) for each image to be fused.
  • the trained artificial agent can utilize one or more trained deep neural networks (DNN) to detect the imaging modalities of the medical images.
  • DNN deep neural networks
  • body parts are detected in each of the medical images to be fused.
  • a trained intelligent artificial agent corresponding to the imaging modality detected for that agent automatically detects a region of the body (e.g., cardiac, abdominal, head, full body, etc.) present in that image.
  • an intelligent artificial agent can be trained to perform the modality and the body region detection in the same step.
  • a trained artificial agent corresponding to the detected body region and imaging modality can be used to detect and/or segment specific anatomical structures, such as organs, vessels, landmarks, etc. in each image.
  • a trained artificial agent may use deep learning based segmentation.
  • rigid-body alignment of the medical images is performed by a trained artificial agent.
  • a trained artificial agent corresponding to the right configuration i.e., imaging modality and anatomy aligns the images by performing rigid-body registration of the images, as described above in the embodiment FIG. 5 .
  • Multiple intelligent artificial agents can be trained using supervised deep reinforcement learning as described above, each trained to perform rigid-body registration for a different configuration of imaging modality and anatomy.
  • the medical images can be down-sampled to generate reduced-resolution images can the rigid registration can be performed based on the reduced-resolution images for fast alignment.
  • an accuracy analysis of the alignment is performed and it is determined whether the accuracy of the alignment is sufficient.
  • a trained artificial agent can evaluate the accuracy of the alignment and assign an accuracy measure or confidence value to the alignment.
  • the trained artificial agent may also be trained to consider whether the alignment would be improved by deformable registration. If the accuracy of the alignment is determined to be sufficient, the method proceeds to step 1818 . If the accuracy of the alignment is determined to be insufficient and considered to be improvable by deformable registration, the method proceeds to step 1812 .
  • a trained artificial agent corresponding to the right configuration (i.e., imaging modality and anatomy) can perform the sparse deformable image registration of the medical images using the method of FIG. 10 with a sparse parametric deformation model.
  • a thin plate spline (TPS) model can be used as the deformation model.
  • the trained artificial agent can be trained using supervised trajectory learning, as described above, with the movement (action) in each step corresponding to movement of one of the TPS control points. The total maximum movement of a TPS control point can be constrained since rigid-body alignment is performed beforehand.
  • an accuracy analysis of the alignment is performed and it is determined whether the accuracy of the alignment is sufficient.
  • a trained artificial agent can evaluate the accuracy of the alignment and assign an accuracy measure or confidence value to the alignment.
  • the trained artificial agent may also be trained to consider whether the alignment would be improved by dense deformable registration. If the accuracy of the alignment is determined to be sufficient, the method proceeds to step 1818 . If the accuracy of the alignment is determined to be insufficient and considered to be improvable by dense deformable registration, the method proceeds to step 1816 .
  • dense field alignment of the medical images is performed by a trained artificial agent.
  • a trained artificial agent corresponding to the right configuration i.e., imaging modality and anatomy
  • a final alignment accuracy evaluation may be performed by a trained artificial agent before proceeding to step 1818 .
  • modality and anatomy specific fusion is performed for the aligned medical images.
  • a trained intelligent artificial agent fuses the information from all of the images using learned proper visualization and quantification methods according to the detected modality, anatomy, and alignment accuracy.
  • the trained artificial agent can tune cinematic rendering based on the fused knowledge for hyper-realistic accuracy.
  • Data enhancement can be achieved by the trained artificial agent selecting, for each voxel, which information (from which of the aligned medical images) to show, such as an ultrasound image with scar tissue highlighted.
  • procedure phase detection and phase-specific fusion are performed.
  • a trained intelligent artificial agent detects the phase of the procedure (e.g., surgery/intervention) based on the medical images and fuses information associated with the current phase in the aligned medical images.
  • longitudinal trend analysis and prediction are performed.
  • a trained intelligent artificial agent learns trends from longitudinal data and predicts a next step to guide the procedure based on the learned trends.
  • a deep neural network for regression can be trained based on training pairs with known ground truth transformations with the loss function being calculated on the transformation parameters.
  • a deep encoder-decoder can be trained based on trained pairs with known ground truth transformations to directly output the deformation field. A method for training such a deep neural network using weakly supervised deep dense correspondence learning is described below.
  • FIG. 19 illustrates a method of training a deep neural network (DNN) for deformable image registration using weakly supervised deep dense correspondence learning according to an embodiment of the present invention.
  • a first pool (A) of medical images with unknown dense correspondence is obtained.
  • the medical images in pool A can be obtained from a database of stored medical images.
  • the medical images in pool A can be obtained by acquiring images using an image acquisition device.
  • the medical images in pool A can be any type of medical images depending on a registration task for which the DNN is being trained.
  • a second pool (B) of medical image pairs with known dense correspondence is obtained.
  • the medical image pairs can be acquired using one or more image acquisition devices or can be obtained from a database of stored medical images.
  • the size of pool B will likely be much smaller than the data pool (A) with unknown dense correspondence.
  • Various techniques can be used to obtain the dense correspondence between the medical image pairs in pool B.
  • the dense correspondence between the medical image pairs can be manually annotated.
  • one medical image can be artificially deformed via a deformation model, such as a spline model, statistical model biomechanical model, etc. The original image and the deformed image form a pair with a known dense correspondence.
  • image pairs can from the same patient can be acquired with a certain motion tracking method that provides the dense correspondence between the images.
  • a certain motion tracking method that provides the dense correspondence between the images.
  • tagged cardiac MRIs, or CT/MR images acquired with respiratory and/or cardiac phase recorded as a surrogate signal of the underlying motion can be used to obtain paired images with known dense correspondence.
  • M image pairs that form a loop are selected from pool B.
  • the dense correspondences for the selected image pairs (I a ,I b ), (I b ,I c ), (I c ,I a ) are denoted as F ab , F bc , F ca , respectively.
  • the image pairs can be picked randomly.
  • steps 1906 - 1910 are repeated multiple times to generate a plurality of training datasets.
  • Each of these training datasets formed by combining the image pairs selected from pool B and the images selected from pool A is treated as a single training example for training the DNN. Accordingly, in an advantageous implementation, steps 1906 - 1910 are repeated a large number of times to generate a large number of training datasets (training examples).
  • a deep neural network is constructed.
  • a DNN is constructed that takes two images an input, and outputs a 3D vector field, which represents the dense correspondence between the two input images.
  • the DNN can be a Convolutional Encoder-Decoder Network.
  • FIG. 20 illustrates an exemplary network architecture for Convolutional Encoder-Decoder Network 2000 for dense correspondence estimation.
  • the DNN can use a FlowNet-like architecture.
  • FIG. 21 illustrates an exemplary FlowNet-like deep architecture 2100 for dense correspondence estimation.
  • Other types of DNN architectures can be used as well.
  • the DNN is denoted as F W (•,•), where W denotes the weights of the DNN that are learned in training.
  • step 1916 for each training dataset, dense correspondences are estimated for each neighboring pair of images in the training dataset.
  • the DNN is applied on every neighboring pair of images in (I a , I 1 , . . . I k , I b , I k+1 , . . . , I j , I c , I j+1 , . . . , I N ), resulting in estimated an estimated dense correspondence being calculated by the DNN for each neighboring image pair in the training dataset, i.e., F W (I a ,I 1 ), F W (I 1 ,I 2 ), . . . , F W (I N ⁇ 1 ,I N ), F W (I N ,I a ).
  • the first and last images in the training dataset I a and I N are considered to be neighboring.
  • distance metrics are computed to compare the known dense correspondences for the image pairs selected from pool B in each training dataset to dense correspondences artificially generated by concatenating the dense correspondences estimated by the DNN.
  • the dense correspondences estimated for the neighboring image pairs by the DNN in step 1916 can be concatenated to artificially generate dense correspondences for the image pairs (I a ,I b ), (I b ,I c ), (I c ,I a ) selected from pool B in the training dataset.
  • the artificially generated dense correspondence for (I a ,I b ) can be calculated by concatenating the estimated dense correspondences for the neighboring image pairs from I a to I b , F W (I a ,I 1 ), F W (I 1 ,I 2 ), . . . , F W (I k ⁇ 1 ,I k ), F W (I k ,I b ).
  • the artificially generated dense correspondence for (I b ,I c ) can be calculated by concatenating the estimated dense correspondences for the neighboring image pairs from I b to I c , F W (I b ,I k+1 ), F W (I k+1 ,I k+2 ), . . .
  • the artificially generated dense correspondence for (I c ,I a ) can be calculated by concatenating the estimated dense correspondences for the neighboring image pairs from I c to I a , F W (I c ,I j+1 ), F W (I j+1 ,I j+2 ), . . . , F W (I N ⁇ 1 ,I N ), F W (I N ,I a ).
  • Distance metrics are calculated to compare to known dense correspondences F ab , F bc , F ca for (I a ,I b ), (I b ,I c ), (I c ,I a ) with the artificially generated dense correspondences for (I a ,I b ), (I b ,I c ), (I c ,I a ):
  • D ab D ( F ab ,( F W ( I a ,I 1 ), F W ( I 1 ,I 2 ), . . . , F W ( I k ⁇ 1 ,I k ), F W ( I k ,I b ))).
  • D bc D ( F bc ,( F W ( I b ,I k+1 ), F W ( I k+1 ,I k+2 ), . . . , F W ( I j ⁇ 1 ,I j ), F W ( I j ,I c ))).
  • D ca D ( F ca ,( F W ( I c ,I j+1 ), F W ( I j+1 ,I j+2 ), . . . , F W ( I N ⁇ 1 ,I N ), F W ( I N ,I a ))).
  • the DNN is trained based on a loss function constructed from the distance metrics. If the dense correspondences produced by the DNN are correct, i.e., they represent the real correspondences between the images, the concatenated dense correspondence between two images should be the same as the real dense correspondence. This property is referred to as Cycle Consistency.
  • the distance metrics D ab , D bc , D ca are used to construct the loss function for training the DNN.
  • the loss function can be a linear combination of D ab , D bc , D ca .
  • the DNN is then trained (e.g., using backpropagation and gradient descent techniques) to learn weights that minimize the loss function over all of the training datasets.
  • FIG. 22 illustrates a method for deformable registration of medical images according to an embodiment of the present invention.
  • first and second medical images are received.
  • a DNN trained using weakly supervised deep dense correspondence learning calculates a dense correspondence between the first and second medical images.
  • the first and second medical images are input to a DNN trained using the method of FIG. 19 , and the trained DNN outputs a 3D vector field representing the dense correspondence (i.e., dense deformation field) between the first and second images.
  • the first and second medical images are registered using the dense correspondence.
  • the first image can be registered to the second image by warping the first medical image using the dense correspondence between the first and second medical images.
  • registration result is output, for example, by displaying the registered first and second images on a display of a computer system.
  • the above described methods for intelligent agent based image registration may be applied to perform fusion of ultrasound and volumetric cardiac images (e.g., MR or CT) for joint function, flow, anatomy, and coronary assessment.
  • Previous registrations techniques require explicitly defining a segmentation of a coronary structure (e.g., full heart, chamber(s), valve) or anatomical landmarks.
  • the above described methods enable the fusion of ultrasound and volumetric cardiac image (e.g., MR or CT) without the need to explicitly define a segmentation or landmarks.
  • the above described methods of intelligent agent based image registration may be applied to register CT or MR images to ConeBeam CT images.
  • Interventional ConeBeam CT is often used to guide surgeries or interventions (e.g., cardiac interventions, spine surgery, etc.).
  • Interventional ConeBeam CT often suffers from image artifacts dues to device, which makes robust image registration difficult using previous registration techniques.
  • the above described methods provides robust registration of CT or MR images to Conebeam CT images even in images with noise or image artifacts, which allows to image data from the pre-operative CT or MR images to be fused with the interventional ConeBeam CT images used for guiding an intervention or surgical procedure.
  • the above described methods of intelligent agent based image registration may be applied to perform automatic change detection through deformable registration.
  • a previously acquired medical image of a patient and a follow-up medical image of the same patient can be registered using the above described methods for intelligent agent based deformable image registration in order to determine a deformation field that shows changes to the anatomy of the patient over time.
  • Such automatic change detection can be used in cancer screening to detect tumor growth or reduction over time, to track changes in stenoses, plaques, or other abnormalities in coronary arteries or other vessels, or to track changes in any other anatomical structure of patient due to disease or treatment.
  • the above described methods for intelligent agent based image registration can be used to perform motion compensation.
  • a pre-operative medical image e.g., CT or MR
  • a trained intelligent artificial agent for image registration can be kept always running in a perpetual state as real-time interventional medical images are acquired and the images are always registered even when movement occurs in the real-time interventional images (e.g., due to patient movement, respiratory motion, or cardiac motion).
  • the intelligent artificial agent automatically re-registers the images as the interventional medical images are received in real-time, such that the registration immediately “catches up” to the movement in the real-time images, thus allowing for real-time perpetual compensation of respiratory motion.
  • Computer 2302 contains a processor 2304 , which controls the overall operation of the computer 2302 by executing computer program instructions which define such operation.
  • the computer program instructions may be stored in a storage device 2312 (e.g., magnetic disk) and loaded into memory 2310 when execution of the computer program instructions is desired.
  • a storage device 2312 e.g., magnetic disk
  • FIGS. 2, 3, 4, 5, 6, 7, 10, 12, 14, 17, 18, 19, and 22 may be defined by the computer program instructions stored in the memory 2310 and/or storage 2312 and controlled by the processor 2304 executing the computer program instructions.
  • One or more image acquisition devices 2320 can be connected to the computer 2302 to input image data to the computer 2302 . It is possible that the computer and one or more of the image acquisition devices 2320 may be implemented as one device. It is also possible that the image acquisition devices 2320 and the computer 2302 communicate wirelessly through a network or wireless communication protocol. In a possible embodiment, the computer 2302 may be located remotely with respect to the image acquisition devices 2320 and may perform some or all of the method steps of FIGS. 2, 3, 4, 5, 6, 7, 10, 12, 14, 17, 18, 19, and 22 as part of a server or cloud based service.
  • the computer 2302 also includes one or more network interfaces 2306 for communicating with other devices via a network.
  • the computer 2302 also includes other input/output devices 908 that enable user interaction with the computer 2302 (e.g., display, keyboard, mouse, speakers, buttons, etc.).
  • input/output devices 908 that enable user interaction with the computer 2302 (e.g., display, keyboard, mouse, speakers, buttons, etc.).
  • FIG. 9 is a high level representation of some of the components of such a computer for illustrative purposes.
  • a server communicates with one or more client computers via a network.
  • a client computer may communicate with the server via a network browser application residing and operating on the client computer, for example.
  • a client computer may store data on the server and access the data via the network.
  • a client computer may transmit requests for data, or requests for online services, to the server via the network.
  • the server may perform requested services and provide data to the client computer(s).
  • the server may also transmit data adapted to cause a client computer to perform a specified function, e.g., to perform a calculation, to display specified data on a screen, etc.
  • Certain steps of the above described methods may be performed by a server or by other computers/processors in the network-based cloud-computing system. Certain steps of the above described methods may be performed locally by a client computer in a network-based cloud computing system.
  • the steps of the above described methods for intelligent artificial agent based image registration and/or training deep neural networks may be implemented in network-based cloud computing system may be performed by the network-based cloud-computing system or by a local client computer in any combination.
  • one or more trained artificial intelligent agents are stored on a server or other remote computer device in the network-based cloud-computing system, and a client computer electronically transmits two or more medical images to be registered to the server of the network-based cloud-computing system.
  • the server and/or other computers/processors in the network-based cloud-computing system performs the registration of the medical images using the one or more trained artificial agents and electronically transmits the registration results to the client computer, which then displays the registration results on a display of the client computer.
  • the method steps to perform the registration in the network-based cloud-computing system can be performed by a single computer device in the network-based cloud-computing system (e.g., the server) or may be distributed on multiple computer devices or processors in the network-based cloud-computing system.

Abstract

Methods and systems for image registration using an intelligent artificial agent are disclosed. In an intelligent artificial agent based registration method, a current state observation of an artificial agent is determined based on the medical images to be registered and current transformation parameters. Action-values are calculated for a plurality of actions available to the artificial agent based on the current state observation using a machine learning based model, such as a trained deep neural network (DNN). The actions correspond to predetermined adjustments of the transformation parameters. An action having a highest action-value is selected from the plurality of actions and the transformation parameters are adjusted by the predetermined adjustment corresponding to the selected action. The determining, calculating, and selecting steps are repeated for a plurality of iterations, and the medical images are registered using final transformation parameters resulting from the plurality of iterations.

Description

  • This application claims the benefit of U.S. Provisional Application No. 62/338,059, filed May 18, 2016, U.S. Provisional Application No. 62/344,125, filed Jun. 1, 2016, and U.S. Provisional Application No. 62/401,977, filed Sep. 30, 2016, the disclosures of which are herein incorporated by reference in its entirety.
  • BACKGROUND OF THE INVENTION
  • The present invention relates to registration of medical images, and more particularly to computer-based automated medical image registration using an intelligent artificial agent.
  • Medical image registration is an important technology that enables image guided therapy, automated diagnosis imaging, and machine-driven image understanding. The goal of image registration is to recover correspondences between two or more medical images, which can be acquired from different patients/objects, the same patient at different time points, different medical imaging modalities, etc., in order to align or fuse the images. The aligned images can provide important information in various applications. For example, the aligned images can be used to guide minimally invasive therapy by fusing pre-operative volume scans (e.g., MRI) with interventional imaging (e.g., DynaCT or fluoroscopy), or to provide complimentary diagnostic information (e.g., co-registration of cardiac ultrasound and MRI for joint function, flow and substrate analysis) or longitudinal analysis (e.g., radiotherapy monitoring through longitudinal CT scans). Due to the vast range of applications to which image registration can be applied, it is challenging to develop a general image registration method that works robustly for all uses.
  • Image registration problems are typically treated as optimization problems in which a generic matching metric (e.g., Mutual Information, Cross Correlation, etc.) is defined to measure the similarity of image pairs to be registered, and transformation parameters between the two images (e.g., rigid body, affine, deformable) are then estimated by an optimizer (e.g., Simplex, Powell, trust region optimization, etc.) via maximization of the defined matching metric. To work robustly, such optimization-based image registration methods typically require extensive hand-crafted engineering for individual registration tasks by incorporating prior knowledge about the specific anatomies, imaging modalities and expected artifacts, or workflows at hand. One reason for this is that a generic matching metric does not guarantee a good representation of the accuracy of the alignment of the data at hand for all use cases, in all circumstances. That is, the global maximum of the matching metric does not necessarily correspond to the correct alignment of the images, for example when the data is noisy, partially occluded, or with drastic different appearances due to different imaging physics. Therefore, the hand-crafted engineering of the matching metric is often required for the specific registration task, for example, by introducing a task-specific region-of-interest (ROI) or calculating hand-crafted features. In addition, a generic matching metric is often non-convex for a given registration task and generic optimizers typically perform poorly on non-convex optimization problems. To avoid being trapped into local minima, prior knowledge is often incorporated to develop optimization schemes for specific registration tasks, for example by prior knowledge-driven initialization/seeding, hierarchical optimization or application specific regularizers.
  • BRIEF SUMMARY OF THE INVENTION
  • The present invention provides methods and systems for registration of medical images using an intelligent artificial agent. Embodiments of the present invention provide artificial intelligent agents that learn registration tasks from training data, and achieve better and better registration results as more data are used for training.
  • In one embodiment of the present invention, a current state observation of an artificial agent is determined based on a plurality of medical images and current transformation parameters. Action-values are calculated for a plurality of actions available to the artificial agent based on the current state observation using a machine learning based model trained based on a plurality of registered training images, wherein the plurality of actions correspond to predetermined adjustments of the transformation parameters. An action having a highest action-value is selected from the plurality of actions and the transformation parameters are adjusted by the predetermined adjustment corresponding to the selected action. The determining, calculating, and selecting steps are repeated for a plurality of iterations, and the plurality of medical images are registered using final transformation parameters resulting from the plurality of iterations.
  • The machine learning based model used to calculate the action-values based on the current state observation may be a trained deep neural network (DNN). The trained DNN may be trained to predict action-values for the plurality of actions based on a plurality of training image pairs with known ground truth transformation parameters using reinforcement learning in which, for each training image pair, a reward for each action of the plurality of actions at a given state is used to train the DNN to learn an optimal registration policy. The trained DNN may be trained using supervised reinforcement learning, in which the reward for each action of the plurality of actions at a given state is directly calculated based on a decrease in a geodesic distance between the transformation parameters at the given a state and the ground truth transformation parameters for the training image pair resulting from applying the action. The plurality of training image pairs may include training image pairs synthetically generated by artificially de-aligning other training image pairs using randomly generated perturbations to the ground truth transformation parameters of the other training image pairs. The plurality of training image pairs may include training image pairs synthetically generated from other training image pairs to provide dense sampling close to the ground truth transformation parameters by co-deforming the other training image pairs by randomly generated affine transformations within a range of the ground truth transformation parameters. The plurality of training image pairs includes training image pairs synthetically generated by artificially altering image appearance of other training image pairs using synthetic image generators.
  • In an embodiment, the transformation may be a rigid-body transformation and the plurality of actions may comprise actions corresponding to predetermined adjustments of translation and rotation rigid-body transformation parameters. The plurality of actions may comprise respective actions corresponding to increasing and decreasing each of the translation and rotation rigid-body parameters by a constant or time-varying step size. The rigid-body transformation may be a 3D rigid-body transformation with rigid-body transformation parameters including three translation parameters and three rotation parameters. The rigid-body transformation may be a 2D rigid-body transformation with rigid-body transformation parameters including two translation parameters and one rotation parameter.
  • In an embodiment, wherein the plurality of medical images may include a first medical image and a second medical image and the transformation may be a dense deformation model that models deformations applied to the second medical image. The plurality of actions may comprise actions corresponding to adjustments to parameters of the deformation model. The first and second medical images may be registered by generating a dense deformation field corresponding to final parameters of the deformation model resulting from the plurality of iterations and warping the second medical image to register the second medical image with the first medical image by applying the dense deformation field corresponding to the final parameters of the deformation model.
  • In an embodiment, the intelligent artificial agent based image registration may be performed hierarchically using multi-scale image data. The determination of the current state observation, calculation of the action-values, and selection of the action having the highest action value may be repeated using a first image resolution of the plurality of medical images and a first machine learning based model trained using training images at the first image resolution until a first stop condition is reached. The determination of the current state observation, calculation of the action-values, and selection of the action having the highest action value may then be repeated using a second image resolution of the plurality of medical images and a second machine learning based model trained using training images at the second image resolution until a second stop condition is reached.
  • In another embodiment, a method for training an intelligent artificial agent to perform image registration according may include obtaining training image pairs and generating synthetic training image pairs, defining a state observation input for the artificial agent, defining a set of possible actions available to the artificial agent, defining a reward mechanism for learning a registration policy, and training a deep neural network (DNN) to predict action-values for the set of possible actions based on the state observation input using the training image pairs.
  • These and other advantages of the invention will be apparent to those of ordinary skill in the art by reference to the following detailed description and the accompanying drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIGS. 1A and 1B illustrate examples of medical image registration problems that present challenges to traditional optimization based image registration techniques;
  • FIG. 2 illustrates a method for training an intelligent artificial agent to perform image registration according to an embodiment of the present invention;
  • FIG. 3 illustrates a method for image registration using an intelligent artificial agent according to an embodiment of the present invention;
  • FIG. 4 illustrates a framework for artificial agent image registration using the method of FIG. 3 according to an embodiment of the present invention;
  • FIG. 5 illustrates a method for rigid registration of medical images using an intelligent artificial agent according to an embodiment of the present invention;
  • FIG. 6 illustrates and exemplary deep neural network (DNN) for rigid 3D/3D medical image registration according to an embodiment of the present invention;
  • FIG. 7 illustrates a method of hierarchical image registration using multi-scale image data according to an embodiment of the present invention;
  • FIG. 8 illustrates exemplary registration results for spine computed tomography (CT) and cone beam CT (CBCT) registration and cardiac CT and CBCT registration;
  • FIG. 9 illustrates a comparison of supervised deep reinforcement learning and unsupervised deep reinforcement learning for 2D/2D spine image registration;
  • FIG. 10 illustrates a method for deformable registration of medical images using an intelligent artificial agent according to an embodiment of the present invention;
  • FIG. 11 an exemplary idealized deformation trajectory between a moving image and a fixed image;
  • FIG. 12 illustrates an exemplary DNN for deformable image registration according to an embodiment of the present invention;
  • FIG. 13 illustrates a Gaussian radial basis function (RBF);
  • FIG. 14 illustrates a framework for decoding encoded flow actions output from a trained DNN according to an embodiment of the present invention;
  • FIG. 15 illustrates examples of deforming a contour of an organ of interest using Gaussian kernels;
  • FIG. 16 illustrates patch extraction from landmark positions of a moving image;
  • FIG. 17 illustrates a framework for performing a second stage of a multi-stage deformable registration according to an embodiment of the present invention;
  • FIG. 18 illustrates a method of autonomous intelligent artificial agent based cognitive image fusion according to an embodiment of the present invention;
  • FIG. 19 illustrates a method of training a DNN for deformable image registration using weakly supervised deep dense correspondence learning according to an embodiment of the present invention;
  • FIG. 20 illustrates an exemplary network architecture for a Convolutional Encoder-Decoder Network for dense correspondence estimation;
  • FIG. 21 illustrates an exemplary Flow-Net like deep architecture for dense correspondence estimation;
  • FIG. 22 illustrates a method for deformable registration of medical images according to an embodiment of the present invention; and
  • FIG. 23 is a high-level block diagram of a computer capable of implementing the present invention.
  • DETAILED DESCRIPTION
  • The present invention relates to registration of medical images using an intelligent artificial agent. Embodiments of the present invention are described herein to give a visual understanding of the image registration methods. A digital image is often composed of digital representations of one or more objects (or shapes). The digital representation of an object is often described herein in terms of identifying and manipulating the objects. Such manipulations are virtual manipulations accomplished in the memory or other circuitry/hardware of a computer system. Accordingly, is to be understood that embodiments of the present invention may be performed within a computer system using data stored within the computer system or available through a network system.
  • Image registration refers to the alignment of two or more images or transformation of different images into one coordinate system. For example, medical image registration can be used to recover correspondences between two or more images acquired using from different patients, the same patient at different times, different medical imaging modalities (e.g., computed tomography (CT), magnetic resonance imaging (MRI), positron emission tomography (PET), ultrasound, etc.), and/or statistical or geometric models. The images can be brought into the same coordinate system via various transformation models, such as, but not limited to, rigid-body transformation, affine transformation, parametric splines, and dense motion fields. Medical image registration is utilized in many different applications. For example, the aligned images can provide complimentary information for fusion-based decision making (e.g., cardiac assessment from MR, CT and ultrasound), motion compensation in image reconstruction (e.g., to cope with breathing motion), enabling longitudinal change analysis (e.g., follow-up reading of cancer patients by registering new and past images of the same patient), or guiding minimally invasive therapy (e.g., fusion of preoperative medical image data into the surgical coordinate system for guidance).
  • In the following, embodiments of the present invention will be described using the pair-wise image registration case. However, the embodiments described herein can easily be extended to the case of group-based registration, where a group of images are all registered to the same reference image (e.g., one of the images in the group or an atlas image) simultaneously. Image registration has typically been formulated as an optimization problem, where a generic matching metric is defined to measure the similarity of the image pairs to be registered. The transformation parameters between the image pairs are then estimated via maximization of a pre-defined matching metric using an optimizer. This formulation faces significant challenges in registering medical images. FIGS. 1A and 1B illustrate examples of medical image registration problems that present challenges to traditional optimization based image registration techniques. One challenge of traditional optimization based image registration is that a generic matching metric is often non-convex for medical images, and generic optimizers typically perform poorly on non-convex problems. This may be an issue in cases in which there is a large difference in the field of view (FOV) of the images to be registered, such as example shown in FIG. 1A. FIG. 1A shows an overlay of a spine CT volume 102 and a Cone-beam CT (CBCT) volume 104 before registration (100) and after registration (105), with a large difference in FOV between the CT and CBCT volumes 102 and 104. The large difference in the FOV between the CT and CBCT volumes 102 and 104 to be registered can result an optimizer finding a number of local maximums of the matching metric due to the repetitive nature of the vertebra. Another challenge of traditional optimization based image registration is that a generic matching metric is often sensitive to image noise and artifacts, such as in the example of FIG. 1B, to partial coverage of the organ of interest (due to field of view or prior surgery for instance), and to drastically different appearances due to different imaging physics (e.g. ultrasound/MRI registration for cardiac image fusion). FIG. 1B shows a cardiac CT volume 110 with contrast enhanced vessels and a cardiac CBCT volume 112 with soft tissue contrast and streaking artifacts. This use case of registering cardiac CT and cardiac CBCT is very common for cardiac interventions, and is often used in a catheterization laboratory (cath lab) to fuse preoperative information to interventional images for guidance of valve replacement, PCI, and other interventions. In FIG. 1B, the differences in the CT and CBCT volumes 110 and 112 due to the streaking artifacts and the difference in appearance of the vessels may result in an inaccurate registration of the CT and CBCT volumes 110 and 112 using traditional optimization based image registration.
  • Embodiments of the present invention reformulate medical image registrations problems by training an intelligent artificial agent to more closely mimic how humans perform image registration as a process of sequential actions of object recognition and manipulation. Embodiments of the present invention train an intelligent artificial agent to register image pairs through a combination of supervised learning and hierarchical image processing. The embodiments can be extended to the registration of groups of images by different strategies, such as extending the number of actions so that there is a set of actions for each image or by running a set of agents together, at the same time and cooperatively. The intelligent artificial agent is implemented on one or more computers or processors by executing computer program instructions (code) loaded into memory. The intelligent artificial agent observes its environment (i.e., the images to be registered) and autonomously acts on that environment to register the images using a registration strategy (herein referred to as a “policy” to be consistent with reinforcement learning terminology) learned using machine learning. The agent can be triggered on demand, or be persistent and always running in the background, ensuring robust image registration even if one of the image moves, for instance due to patient motion in the case of real-time imaging. The input for the intelligent artificial agent (the fixed image and the updated moving image as it is transformed by the current estimate of the transformation parameters) is used to determine its “state”, which corresponds to the current alignment of the raw image pairs. The output of the intelligent artificial agent is an “action” to improve the alignment between the fixed and moving image by changing the parameters that define the transformation between the images. In an advantageous embodiment of the present invention, a deep neural network (DNN) is used to approximate the state-value function which returns for each state, a value for each possible action which indicates how good that action is. Other approximation functions can be used in alternative embodiments. During training, the intelligent artificial agent learns a registration policy via a DNN that maps the current state to the optimal action that best improves the accuracy of the alignment. During testing, in order to register newly input images, the trained artificial agent applies the learned registration policy to improve the alignment of the images until it converges to the correct pose. By using this deep learning model, embodiments of the present invention are advantageous in that the intelligent artificial agent can inherently learn both a data-driven matching metric and a task-driven policy from raw image data without hand-crafted engineering and perform robust medical image registration.
  • FIG. 2 illustrates a method for training an intelligent artificial agent to perform image registration according to an embodiment of the present invention. When registering a pair of images, one image is designated as the reference image or fixed image and the other image is designated as the moving image or floating image. The reference and moving images to be registered can be denoted as Iref:RN→R and Imov:RM→R, respectively, where N and M are the respective dimensionalities of the two images to be registered. In medical images N and M are typically equal to 2, 3, or 4 (3D+time). For example, in 3D/3D registration, N=M=3. In 2D/2D registration, N=M=2. In 2D/3D, N=2 and M=3. Time series could be registered globally as well. In this case, N=M=4. The goal of image registration is to estimate a transformation, denoted as T: RN→RM, that transforms the moving image Imov to be aligned with the reference image Iref. The transformation, in general terms, can represent a curved manifold and can be modeled using K transformation parameters. The goal of the registration then becomes estimating the K transformation parameters such that the similarity between Iref and Imov is maximized. In some embodiment, various types of transformation models, such as an affine transformation, spline model, or biomechanical model, can be used to model the transformation of the moving image Imov to be aligned with the reference image Iref. In other embodiments, a non-parametric deformable registration can be estimated with a dense motion field. In this case, motion parameters can be estimated for each pixel/voxel on the grid, and the motion parameters can be the displacement directly, stationary velocities, or kinetic moments, or any subspace representation of those (e.g., PCA modes among others). The method of FIG. 2 trains a deep neural network (DNN) to predict action-values for a set of actions corresponding to the transformation parameters. By parameterizing the action space, the DNN can be trained to perform any registration task using the method of FIG. 2.
  • Referring to FIG. 2, at step 202, training images are obtained and/or generated. In an advantageous embodiment, the DNN can be trained for a particular registration task based on training image pairs (i.e., a reference image Iref and a moving image Imov) corresponding to the particular registration task and known ground truth transformations between the training image pairs. For example, the training image pairs may be medical images acquired using different imaging modalities (e.g., CT, MRI, ultrasound, PET, etc.), medical images acquired from the same patient (using the same imaging modality) at different times, or medical images acquired from different patients. The training image pairs can be pairs of 3D medical images for 3D/3D registration, pairs of 2D medical images for 2D/2D registration, or a 3D image and a 2D image for 2D/3D registration. The training image pairs can include actual medical images of patients acquired from medical image acquisition devices, such as a CT scanner, MRI scanner, C-arm image acquisition device, ultrasound device, etc. These training image pairs can be obtained by receiving the images from the image acquisition devices used to acquire the images or by loading previously acquired images from a storage or memory of a computer system. The ground truth transformation for these image pairs can be acquired by manual expert annotation or by applying an existing image registration method. In cases in which an existing registration method is used to compute the ground truth transformation for a training image pair, the reference image in the training image pair may be replaced with the transformed moving image using the computed transformation resulting from the existing registration method, in order to compensate for inaccuracies in the existing registration method.
  • The training image pairs can also include synthetic images which are generated by augmenting other training images rather than being acquired in a scan of a patient using an image acquisition device. It may be difficult to obtain a large number of labeled training pairs of actual scanned medical images of patients. According to an advantageous embodiment, a large training set can be synthetically generated from a small number of labeled training pairs. In order to generate synthetic training image pairs, each of the aligned training image pairs with a known ground truth transformation can be artificially de-aligned by transforming one of the images with a set of transformations sampled in the space of the transformation parameters. Since the original ground truth transformation is known, each of these transformation results in an additional training image pair for which the ground transformation is also known. In addition, in order to provide denser sampling of the transformation parameter space close to the ground truth transformation, each aligned training image pair can be co-deformed by a number of perturbations within specified range of the ground truth transformation parameters to generate additional training image pairs that are close to the ground truth transformation in the transformation parameter space. Additional synthetic training images may also be generated by altering the appearance of one or both images in existing training pairs by adding noise and/or artifacts, or even by simulating other imaging modalities from unpaired training images using synthetic image generators including, but not limited to, physics-based simulators (e.g., MRI simulator, ultrasound simulator) or machine learning based simulators (e.g., training a generative adversarial network (GAN) to simulate an imaging modality from another one, the GAN being trained on paired images).
  • At step 204, the state input to the DNN is defined for the particular image registration task for which the DNN is being trained. The state that is input to the DNN is an observation of the current alignment of the two images. Given two images Iref and Imov to be registered and current and previous transformation parameters {mt}t=−τ, . . . , 0, the intelligent artificial agent extracts an observed state of the current alignment, and inputs the state to the DNN. The state observation can include the current and previous transformation parameters {mt}t=−τ, . . . , 0; the whole image or a region of interest (ROI) of the reference image Iref; the whole image or ROIs of the moving image Imov transformed by the transformations specified by {mt}t=−τ, . . . , 0; the whole image or ROIs of J transformed by transformations derived from {mt}t=−τ, . . . , 0, e.g., via perturbation; in the case of 2D/3D registration (N=2, M=3), the whole image or ROIs of 2D projections of Imov transformed by the transformation specified by {mt}t=−τ, . . . , 0; in the case of 2D/3D registration (N=2, M=3), the whole image or ROIs of 3D projections of Imov transformed by transformations derived from {mt}t=−τ, . . . , 0, e.g., via perturbation; and/or the above images post-processed by operations such as filtering, subtraction, etc. In one embodiment, the state of the current alignment can be computed by comparing the transformed moving image Imov using current transformation parameters (or projections of the transformed moving image in the case of 2D/3D registration) and the reference image Iref using a specified function or operation. For example, in an advantageous implementation, the state of the current alignment can be computed by subtracting the transformed moving image Imov using current transformation parameters from the reference image Iref. In another embodiment, the reference image Iref and the transformed moving image Imov using current transformation parameters are passed to the neural network which features a multi-channel architecture (for example, in FIG. 12). The comparison of the images is then automatically performed and learned by the neural network.
  • At step 206, a set of possible actions are defined for the registration task. The set of possible actions include a set of actions that alter the transformation parameters. For example, to perform rigid registration, the set of actions can include actions that alter rigid body transformation parameters (e.g., 3 translation parameters and 3 orientation parameters for 3D/3D registration). For deformable registration, a deformation model can be used to model the deformation field and the set of actions can include actions that alter parameters of the deformation model used to model the deformation field. The set of possible actions can include actions to change (increase and decrease) individual transformation parameters by a certain step size δ, actions to change multiple (l) transformation parameters simultaneously by certain step sizes δ1 . . . δl, actions to optimize the step size δi (i=1 . . . k) for each transformation parameter, an action that determines that the correct registration is achieved and terminates the registration process, and an action that compounds the transformation resulting from the previously selected actions in the curved manifold of registration parameters to ensure that the entire path of corrective actions stay within the manifold. It is to be understood that the present invention is not limited to the above described actions and other actions may be included based on the particular registration task for which the intelligent artificial agent is trained.
  • At step 208, a reward mechanism is defined. A reward mechanism is defined that assigns a reward for each action, depending on the effect of that action of the resulting alignment of the images. A higher reward is assigned for actions that lead to more accurate alignment, while a lower reward or penalty (negative reward) is assigned for actions that lead to less accurate alignment. Reinforcement learning can then be used to train the agent. In an advantageous embodiment, the reward mechanism can be defined as the reduction of L2 error in the transformation parameters:

  • reward=∥m after −m gt2 −∥m before −m gt2  (1)
  • where mbefore and mafter are the transformation parameters before and after taking the action, and mgt is the ground truth parameters for the correct registration. In this case, the agent training is fully supervised and this type of reward can be thought of as a “GPS”, which directly tells the agent what is the best action to take. After performing several trials and learning from the “GPS” (training phase), one can turn off the “GPS” (testing phase) and the registration strategy learned from the “GPS” in the training phase can be applied to register images for which the transformation is not yet known.
  • At step 210, a DNN is trained to predict action-values for the set of possible actions based on the state input using the training images. A core part of the intelligent agent is a DNN (or possibly multiple DNNs) that takes the state as input and outputs action-values for all possible actions with proper parametrization for the particular image registration task. The action-value for an action can represent the highest possible future reward if an action is taken discounted based on the number of steps to reach the ground truth transformation parameters. With the reward mechanism described above (e.g., reduction of L2 error in the transformation parameters), a higher action-value indicates that the action is driving the registration toward the correct direction (i.e., toward the correct final registration result) and therefore is the preferred action to be taken. The action space can be discrete or continuous with parameterization.
  • According to an advantageous embodiment of the present invention, the DNN can be trained using a supervised Deep Reinforcement Learning (DRL) technique. Reinforcement Learning (RL) is a type of machine learning in which a software based artificial agent uses reward feedback to automatically learn ideal behavior in a specific context and for a specific task. In DRL, which combines DNNs with RL and has recently been applied to train an artificial agent to play Atari and Go games, a policy learning process is formulated as a RL problem and the action-value function (also known as Q function) is estimated following the Bellman equation as an iterative update. In DRL and RL, the training of the agent is typically unguided and the agent is free to evolve in its environment according to its current policy estimate. However, training efficiency of the agent and iterative update of the Q function is relatively low, which could render DRL impractical or impossible for image registration tasks where the input data and the parameter space is large, such as registration of 3D medical image volumes. In an advantageous embodiment of the present invention the DRL training of the DNN is supervised based on the known ground truth transformations of the training image pairs using a greedy search strategy in which the action-values are approximated analytically. This supervised DRL has an advantage that the Q-network guarantees to converge and the behavior distribution can be arbitrarily smooth without the need to deploy a memory replay strategy, which can have a very high memory footprint. The DNN can be trained in an end-to-end fashion, or can be trained separately for image feature extraction layers and policy learning layers in order to bring more robustness to variations in image content if needed.
  • In an alternative embodiment, the trained DNN can be trained to predict the action-values based on the training image pairs using evolution strategies (ES) to learn the parameters (e.g., weights) of the DNN. In ES, instead of using reinforcement learning, genetic-like algorithms or sampling strategies are used to optimize the parameters of the DNN directly based on the training data.
  • The trained DNN is stored, for example, in memory or storage of a computer system or on a remote “cloud-based” computer system. The trained DNN can then be loaded and used when registering two newly received images to iteratively predict action-values for each of the set of possible actions based on the current state of the images to be registered and execute the selected action having the highest predicted action-value.
  • FIG. 3 illustrates a method for image registration using an intelligent artificial agent according to an embodiment of the present invention. At step 302, first and second medical images are received. For example, the first and second medical images can be images acquired using different imaging modalities (e.g., CT, MRI, ultrasound, PET, DynaCT, etc.), images of the same patient acquired at different times, or images acquired from different patients. The first and second medical images may both be 3D medical images (volumes) (3D/3D registration), both be 2D medical images (2D/2D registration), or may be one 2D medical image and one 3D medical image (2D/3D registration). The first sand second medical images may also be image sequences (3D+t or 2D+t), which can also be registered using the method of FIG. 3. The first and second medical images may be received directly from image acquisition devices, such as a CT scanner, MRI scanner, ultrasound device, C-arm image acquisition device, etc., or may be received by loading previously stored medical images from a memory or storage of a computer system or receiving the medical images in an electronic transmission from another computer system. One of the medical images is designated as the reference image Iref and the other is designated as the moving image Imov.
  • At step 304, the current state is determined based on the first and second medical images and the current set of transformation parameters. At step 306, the current state is input to the trained deep neural network and action-values for the set of possible actions are calculated based on the current state using the trained DNN. The possible actions can correspond to adjustments to the transformation parameters. At step 308, the action with the highest action-value is selected and performed. The selected action will typically adjust at least one transformation parameter, which changes the transformation of the moving image Imov. At step 310, it is determined if a stop condition has been met. For example, the stop condition can be reached if it is determined that a correct registration is achieved or when a maximum number of iterations have been performed. It can be determined that a correct registration is achieved when a “stop registration” action is selected as the action with the highest action-score or by comparing an error value between the reference image and the transformed moving image to a predetermined threshold value. If the stop condition is not met, the method returns to step 304 and repeats steps 304-310. Accordingly, steps 304-310 are iterated until the stop condition is met. When the stop condition is met, the method proceeds to step 312. At step 312, the registration result is output. The moving image is transformed by the final transformation parameters, and the transformed moving image and the reference image can be displayed on a display of a computer system. For example, the transformed moving image can be overlaid on the reference image and be displayed as a fused image. The registered images can also be electronically transmitted to a remote computer system to be displayed on a display of the remote computer system. Although FIG. 3 only shows outputting the final registration result, it is also possible that the incremental registration results using the current transformation parameters can be output (e.g., displayed on a display of a computer system) as each iteration of steps 304-310 is performed. Such incremental results would allow a user to view each adjustment made by the intelligent artificial agent in real-time as the artificial agent is registering the images.
  • FIG. 4 illustrates a framework for artificial agent image registration using the method of FIG. 3 according to an embodiment of the present invention. As shown in FIG. 4, two medical images 402 to be registered are received. The state 404, which is an observation of the current alignment of the images 402, is determined based on the current transformation parameters. For example the state observation 404 can be an image resulting from subtracting the moving image as transformed by the current transformation parameters from the reference image. The state observation 404 is input to the trained DNN 406, which acts as the “brain” of the intelligent artificial agent. The trained DNN 406 calculates a respective action-score for each action in a set of possible actions, and the action 408 with the highest action score is selected. The action 408 is performed to adjust at least one of the transformation parameters, and registration is performed with the updated transformation parameters, resulting in an updated state observation 404. This process is iterated until a stop condition is met. For example, the process can be iterated until a stop registration action is selected or until a predetermined maximum number of iterations are performed.
  • In a possible embodiment, multiple agents can be trained in a coordinated fashion to register multiple objects (e.g., organs, anatomical structures, etc.) in the images simultaneously. In one exemplary implementation, each object is aligned independently in each step by a corresponding agent (with a corresponding DNN) trained to register that object, and a master agent is trained to coordinate the actions suggest from the individual object-specific agents and output the optimal overall action given the constraints of the relative positions of the multiple objects to be registered. In another possible implementation, each agent can be trained to collaborate with other agents and/or unknown teammates. Similarly, multiple agents can also be trained in a coordinated fashion to register multiple images simultaneously by incorporating the constraints of the relative transformations among the images.
  • In another possible embodiment, a multi-task agent can be trained to perform registration of a variety of images coming from different sources, e.g., varying modalities, imaging protocols, and organs. Transfer and hierarchical reinforcement learning can be applied to train the multi-task agent by transferring the knowledge from agents trained for individual registration tasks (e.g., specific organs) and abstracting the features and policies for registration. New task-specific agents can also be trained more efficiently by transferring the knowledge from previously trained agents on similar registration tasks. In this context, in an advantageous implementation, an actor-mimic approach can be used to train a single intelligent artificial agent capable of registering multiple different types of image pairs. Let N be the number of registration tasks, between N number of imaging modality pairs. N different artificial agents Ai (iε1 . . . N) can be trained as described above, each corresponding to a specific registration task between a specific type of imaging modality pairs. A “mimic” agent Am is then trained whose objective is to be as accurate and robust as all of the artificial agents Ai in their respective tasks. In one embodiment, Am is trained using N different types of training image pairs and the loss function to minimize is the sum of the least-square differences between the action-values of Am and those of the respective individual agents Ai. Other loss functions can be designed to enable domain transfer and task generalization. With such a framework, the mimic agent will learn generic features present in all imaging modalities that will enable the registration task.
  • In another possible embodiment, an intelligent artificial agent can be trained to register the medical images in a coarse-to-fine manner using different transformation models. For example, the intelligent artificial agent may first adjust rigid-body transformation parameters until a first stop condition is reached (e.g., a predetermined number of iterations or a switch transformation model action is selected), then affine transformation parameters until a second stop condition is reached (e.g., a predetermined number of iterations or a switch transformation model action is selected), then polyaffine transformation parameters until a third stop condition is reached (e.g., a predetermined number of iterations or a switch transformation model action is selected), and then dense deformation parameters until a final stop condition is reached (e.g., a predetermined number of iterations or a stop registration action is selected). In this case, the trained DNN can be trained to calculate the action-values for actions corresponding to adjustments to all of the various transformation models (e.g., rigid, affine, polyaffine, dense) as well as action-values for actions corresponding switching between the various transformation models.
  • Rigid Registration Using an Intelligent Artificial Agent
  • In an advantageous embodiment of the present invention, an intelligent artificial agent is used to perform rigid registration of medical images. The rigid registration is described herein for 3D/3D medical image registration. It is to be understood that the methods for rigid registration described herein are not limited to t 3D/3D medical image registration and can be similarly applied to 2D/2D medical image registration as well. The methods for rigid registration described herein register images using a rigid-body transformation including three translation parameters and three rotation parameters. Such methods can be similarly applied to perform registration using other types of parametric transformations, such as a nine parameter affine transformation (three translation parameters, three rotation parameters, and three scale parameters), a 12 parameter affine transformation (three translation parameters, three rotation parameters, three scale parameters, and three shear parameters), or a poly-affine transformation.
  • FIG. 5 illustrates a method for rigid registration of medical images using an intelligent artificial agent according to an embodiment of the present invention. At step 500 of FIG. 5, a DNN is trained to predict action-values for a set of actions corresponding to rigid-body transformation parameters. Step 500 is performed in an offline training stage prior to the on-line image registration for newly input/received images performed in steps 502-514.
  • Given a reference image Iref and a moving (floating) image Imov, the goal of 3D/3D rigid-body image registration is to transform Imov using rigid-body transformation parameters m=[tx, ty, tz, θx, θy, θz], represented by a column-wise homogenous transformation matrix as:
  • T m = [ 1 0 0 t x 0 cos θ x - sin θ x t y 0 sin θ x cos θ x t z 0 0 0 1 ] × [ cos θ y 0 sin θ y 0 0 0 0 0 - sin θ y 0 cos θ y 0 0 0 0 1 ] × [ cos θ z - sin θ z 0 0 sin θ z cos θ z 0 0 0 0 1 0 0 0 0 1 ] ( 2 )
  • so that Iref and Imov are anatomically aligned. Here, a point in Imov, represented in homogenous coordinates as {right arrow over (p)}=[x, y, z, 1]T, is transformed to location Tm{right arrow over (p)} in Imov(Tm). The problem can be cast as a Markov Decision Process (MDP), and the goal is to train an intelligent artificial agent that learns the policy to register two images by making a series of decisions based on the observation (state) of the current alignment. In an advantageous implementation, the current observation, or state st, is the intensity difference between Iref and Imov transformed using the current transformation parameters mt:

  • s t =I ref −I mov(T m t ).  (3)
  • In an advantageous implementation, the decision, or action at is a change in one of the six transformation parameters (three translation parameters tx, ty, tz and three orientation parameters θx, θy, θz) in mt, e.g., at=[1, 0, 0, 0, 0, 0] for translation along the x-axis by 1 mm and at=[0, 0, 0, 1, 0, 0] for rotation around the x-axis by 1°. The next state st+1 is calculated via updating the moving image Imov by applying the action at:

  • m t+1 =m t +a t,

  • s t+1 =I ref −I mov(T m t+1 ).  (4)
  • During training (step 500 of FIG. 5), the agent learns a registration policy using a DNN that maps the current state st to the optimal action at,o. During testing (steps 502-514 of FIG. 5), the agent applies the learned policy in a sequence of N actions {a1,o, . . . , aN,o}, to improve the alignment until it converges to the correct pose or a maximum number of iterations is reached.
  • The DNN can be trained using reinforcement learning. In an advantageous embodiment of the present invention, supervised DRL is used for training the DNN to predict action-values for the set of actions. The core problem is to find a policy that guides the decision process of the artificial agent. In previous DRL gaming applications, the policy learning is formulated as a RL problem with the action-value (also known as Q) function estimated following the Bellman equation as an iterative update. However, when relying on unguided explorations of the agent and iterative update of Q, training efficiency is relatively low. Indeed, this iterative update of Q is known to be unstable or even diverge when a nonlinear approximator is used to represent it, and there is no theoretical proof that such as convolutional neural network (CNN)-based deep network guarantees to converge to the correct Q-function for general applications. In an advantageous embodiment of the present invention, the training is supervised by instructing the agent to follow a greedy registration path, mimicking how humans register two objects in a most efficient manner. In particular, given the current transformation mt and the ground truth registration pose mg, the optimal action at,o at time t along the supervised registration path is defined as the action that minimizes the Euclidean distance between mt+at,i and mg:

  • a t,o=mina t,iεA ∥m g−(m t +a t,i)∥2.  (5)
  • In an advantageous implementation, there are 6×2=12 candidate actions in the action set A, corresponding to the change of ±1 mm for each of the translation parameters [tx, ty, tz] and +1° for the orientation (rotation) parameters [θx, θy, θz], meaning that the movement of the agent is restricted to be on the grid with a step size 1 in the transformation parameter space. By teaching the agent the registration path, the need to estimate Q through iterative update following the Bellman equation can be removed. Instead, the action-value Q function can be calculated explicitly via the following recursive function, assuming that the agent is allowed to run a sufficient number of steps to reach the ground truth pose mg following the supervised greedy path specified in Equation (5):
  • Q ( s t , a t , i ) = { r ( s t , a t , i ) + γ Q ( s t + 1 , i , a t + 1 , o ) if m t + 1 , i - m g 2 > ε r ( s t , a t , i ) + R otherwise ( 6 )
  • where

  • m t+1,i =m t +a t,i,

  • s t+1,i =I ref −I mov(T m t+1,i ).  (7)
  • and the immediate reward r(st, at,i) for action at,i(i=1, 2, . . . 12) is proportional to the decrease of the distance of the transformation parameters to the ground truth pose mg by applying action at,i,

  • r(s t ,a t,i)=∥m g −m t2 −∥m g −m t+12.  (8)
  • The reward is discounted by a predetermined factor γ (e.g., γ=0.9) to implicitly include the number of steps as a cost, and the agent is considered to reach the correct pose and receives a bonus reward R (e.g., R=10) when it is within a predetermined tolerance ε (e.g., ε=0.5) of the ground truth pose mg in the transformation parameter space.
  • It can be shown that if the agent is allowed to take the greediest path off the grid in the transformation parameter space to approach the ground truth pose mg with step size 1, i.e., the only constraint on the action is ∥at,i2=1, and the agent receives a proper bonus reward R when it reaches the target, then Q(st,at,o)≧Q(st,at,i), meaning the trained agent can perform registration by simply choosing the action with the largest Q in the testing phase. In addition, Q(st,at,o) monotonically decreases with the increase in distance between mt and mg, and approaches to a fixed value as the distance goes to infinity. Experimentally, it can be observed the following simplification of the Q-function: Q(st,at,i)=r(st,at,i) without recursive calculation works equally well due to this property, i.e., the term Q(st+1,i,at+1,o) in Equation (6) quickly approaches to approximately a fixed offset that does not affect the selection of the optimal action at a given state.
  • A DNN is trained to represent the action-value Q-function in Equation (6). In an advantageous implementation, the DNN is a deep convolutional neural network CNN). The input to the DNN is the current state st, the output of the DNN has 12 nodes, each corresponding to one of the 12 actions in the action set A, and the loss function can be defined as:

  • Loss=Σk=1 MΣi=1 12 ∥y i(s k)−Q(s k ,a k,i)∥2  (9)
  • Where yi(sk) is the i-th output of the CNN for the k-th sample among M training samples. The DNN (e.g., deep CNN) can be trained using a gradient descent algorithm and backpropagation to learn weights for layers of the DNN that minimize the loss function over all of the training samples. The DNN training scheme described herein has various advantages as compared to previous RL and DRL used in gaming applications. For example, the target Q-function described in the present disclosure is given analytically without iterative estimation, such that the network can be trained much more efficiently and with a more stable convergence property. In addition, the target Q calculation described herein does not require the exploration history of the agent, meaning that the data can be sampled arbitrarily randomly with little correlation, without the need for a large memory storage required by experience replay. These advantages are important to make this framework practical for 3D/3D registration applications dealing with large 3D volumes as input.
  • Labeled training image pairs (i.e., training image pairs with known ground truth transformations) of medical image may not be easy to obtain. According to an advantageous embodiment of the present invention, a limited set of labeled training data can be augmented to generate a large number synthetic training image pairs that can be used as training samples for training the DNN. In order to generate synthetic training image pairs, each of the aligned training image pairs with a known ground truth transformation can be artificially de-aligned by transforming one of the images with a set of transformations sampled in the transformation parameter space. For example, for each aligned training image pair, a predetermined number of rigid-body perturbations can be randomly generated within a predetermined range of each rigid-body transformation parameter. Since the original ground truth transformation is known, each of these transformation results in an additional training image pair for which the ground transformation is also known. In addition, for each aligned pair, denser sampling of the transformation parameter space close to the ground truth pose mg can be performed in order to train the DNN efficiently and reliably, because the network input-output mapping in this space is more complicated and less smooth as compared to that for the transformation parameter space far away from the ground truth aligned pose. In order to provide denser sampling of the transformation parameter space close to the ground truth transformation, each of the aligned training image pairs can be geometrically co-deformed by affine transformations TA:
  • T A = I + [ t 11 t 12 t 13 0 t 21 t 22 t 23 0 t 31 t 32 t 33 0 0 0 0 1 ] ( 10 )
  • where I is the 4×4 identity matrix and all the elements in [tij]i=1,2,3,j=1,2,3 for shearing are independently and randomly generated within a predetermined range (e.g., [−0.25, 0.25]), to cover possible anatomical variations among patients in terms of sizes and shapes.
  • The training (step 500 of FIG. 5) results in a trained DNN that predicts action-values (Q) for each of the set of actions based on the current state of the alignment of the images to be registered. FIG. 6 illustrates an exemplary DNN for rigid 3D/3D medical image registration according to an embodiment of the present invention. As shown in FIG. 6, the DNN 600 is a deep CNN with multiple hidden layers that inputs a 3D volume corresponding to the state of the alignment of the images Ir and If to be registered. In FIG. 6, Ir is the reference image and If is the moving or floating image. The input 3D volume is generated by subtracting If transformed by the current rigid-body transformation parameters from Ir. In the example of FIG. 6, the DNN 600 has five 3D convolutional layers followed by three fully connected layers, but the present invention is not limited to this particular network structure. The output layer of the DNN 600 has 12 nodes, each corresponding to an action in the action set. In particular, the 12 output nodes represent the 12 candidate actions corresponding to adjustments of ±1 mm for each of the translation parameters [tx, ty, tz] and ±1° for the orientation (rotation) parameters [θx, θy, θz]. The output nodes of the DNN 600 output predicted action-values/rewards for the corresponding actions. The intelligent artificial agent then selects the action with the highest predicted action-value and performs the action (i.e., the corresponding adjustment to the transformation) to update the transformation of the moving image If.
  • Returning to FIG. 5, at step 502, first and second medical images are received. For example, the first and second medical images can be images acquired using different imaging modalities (e.g., CT, MRI, ultrasound, PET, DynaCT, etc.), images of the same patient acquired at different times, or images acquired from different patients. The first and second medical images may be received directly from image acquisition devices, such as a CT scanner, MRI scanner, ultrasound device, C-arm image acquisition device, etc., or may be received by loading previously stored medical images from a memory or storage of a computer system or receiving the medical images in an electronic transmission from another computer system. One of the medical images is designated as the reference image Iref and the other is designated as the moving image Imov. The method of FIG. 5 is described for 3D/3D registration, but can be similarly applied for 2D/2D registration as well.
  • At step 504, the current state is determined based on the current rigid body transformation parameters. In an advantageous implementation, the current state is determined by subtracting the moving image transformed by the current rigid-body transformation parameters, Imov(Tm t ) from the reference image Iref, i.e., using Equation (3). This results in a 3D state volume st. In the first iteration, before any of the rigid-body transformation parameters have been adjusted, the initial state can be calculated by subtracting the moving image Imov from the reference image Iref since no transformation has been applied to the moving image yet.
  • At step 506, action-values are calculated for each action in the set of possible actions based on the current state using the trained DNN. As described above, the set of possible actions includes actions corresponding to adjustments to each of the rigid-body transformation parameters. In particular, the set of possible actions can include 12 action actions corresponding to adjustments of ±1 mm for each of the translation parameters [tx, ty, tz] and ±1° for the orientation (rotation) parameters [θx, θy, θz]. The trained DNN, such as the trained DNN 600 shown in FIG. 6, inputs the current state volume st and calculates a predicted action-value (Q value) for each action.
  • At step 508, the action with the highest predicted action value is selected and the rigid-body transformation parameters are updated by performing the selected action. In an advantageous implementation, the selected action will adjust one of the translation parameters by ±1 mm or adjust one of the orientation parameters by ±1°. At step 510, the moving image is transformed using the updated rigid-body transformation parameters. At step 512, it is determined whether a stop condition has been met. For example, the stop condition can be met when it is determined that the transformation parameters have converged to a correct pose or when a predetermined maximum number of iterations have been performed. If the stop condition has not been met, the method returns to step 504 and repeats steps 504-512. Accordingly steps 504-512 are repeated until the stop condition is met. When the stop condition is met, the method proceeds to step 514.
  • At step 514, the registration result is output. The first and second medical images are registered by transforming the moving image by the final rigid-body transformation to align the moving image with the reference image. The registered first and second medical images can be output by displaying the registered first and second medical images on a display of a computer system. For example, the registered first and second images can be overlaid and displayed as a fused image. The registered first and second medical images can also be output by electronically transmitting the registered first and second images to a remote computer system to be displayed on a display of the remote computer system.
  • Although FIG. 5 only shows outputting the final registration result, it is also possible that the incremental registration results using the current rigid-body transformation parameters can be output (e.g., displayed on a display of a computer system) as each iteration of steps 504-512 is performed. Such incremental results would allow a user to view each adjustment made by the intelligent artificial agent to iteratively align the moving image to the fixed image in real-time as the artificial agent is registering the images.
  • Hierarchical Image Registration.
  • Since, in the case of 3D/3D registration, the input to the DNN is a large 3D volume instead of 2D images, the size of the volume is important for practical use. In order to achieve efficiency and accuracy, hierarchical image registration using multiple scale image data can be used. According to an advantageous embodiment of the present invention, two or more separate DNNs (e.g., deep CNNs) can trained at different image resolutions. For example, in the case of two resolutions, a first DNN can be trained with down-sampled volumes with a lower resolution, and a second DNN can be trained using high-resolution (i.e., original resolution without down-sampling) volumes. In an advantageous implementation, both of the first and second trained DNNs are trained using the same grid size (e.g., 64×64×64) volumes as the input but with different resolutions. The first DNN is trained for coarse alignment using down-sampled training volumes with a lower resolution, and focuses on robust alignment of an object in the images to be registered even when the initial displacement is large. The second DNN is trained using high-resolution training volumes with a limited field of view (FOV) and focuses on aligning the object as accurate as possible despite the limited FOV. The first and second DNNs are each trained as described above in connection FIG. 1 and/or FIG. 5. For example, each of the first and second trained DNNs can have network architecture similar to the DNN 600 of FIG. 6. The trained first and second DNNs are then used for hierarchical image registration of newly received images. This hierarchical registration can be applied to any type of registration approach (e.g., rigid, affine, deformable) and any dimensionality (e.g., 2D/2D, 3D/3D. etc.).
  • FIG. 7 illustrates a method of hierarchical image registration using multi-scale image data according to an embodiment of the present invention. At step 702, first and second medical images are received. For example, the first and second medical images can be images acquired using different imaging modalities (e.g., CT, MRI, ultrasound, PET, DynaCT, etc.), images of the same patient acquired at different times, or images acquired from different patients. The first and second medical images may be received directly from image acquisition devices, such as a CT scanner, MRI scanner, ultrasound device, C-arm image acquisition device, etc., or may be received by loading previously stored medical images from a memory or storage of a computer system or receiving the medical images in an electronic transmission from another computer system. One of the medical images is designated as the reference image Iref and the other is designated as the moving image Imov. The method of FIG. 7 is described for 3D/3D registration using a rigid-body transformation, but can be similarly applied for 2D/2D registration and/or for other types of transformation parameterizations.
  • At step 704, the first and second medical images are down-sampled to generate first and second reduced-resolution images. Various well-known down-sampling techniques can be used to down-sample the first and second medical images.
  • At step 706, a current reduced resolution state volume is generated using the first and second reduced resolution images and the current transformation parameters. The current reduced resolution state volume can be determined by subtracting the reduced-resolution moving image transformed by the current transformation parameters from the reduced resolution reference image. At step 708, action-values are calculated for each of the actions in the set of possible actions using the first trained DNN. As described above, the set of possible actions can include actions corresponding to adjustments to each of the rigid-body transformation parameters. In particular, the set of possible actions can include 12 action actions corresponding to adjustments of ±1 mm for each of the translation parameters [tx, ty, tz] and ±1° for the orientation (rotation) parameters [θx, θy, θz]d. The first trained DNN, which is trained using down-sampled reduced resolution images, inputs the current reduced-resolution state volume calculated from the first and second reduced resolution images and calculates a predicted action-value (Q value) for each action. At step 710, the action with the highest action value is selected and the transformation parameters are updated by performing the selected action.
  • At step 712, it is determined whether a stop condition is met. In an exemplary embodiment, the stop condition for registration using the first trained DNN is met when a first predetermined number N1 of iterations using the first trained DNN have been performed. In this case, if N1 iterations using the first trained DNN have not yet been performed, the stop conditions has not yet been met and the method returns to step 706 and repeats steps 706-712. Accordingly, steps 706-712 are repeated until N1 iterations using the first trained DNN are performed. When N1 iterations have been performed using the first trained DNN, the stop condition has been met and the method proceeds to step 714. In an exemplary implementation, N1=200, such that 200 sequential actions are applied based on the reduced resolution images using the first trained DNN in order to roughly align the images. In other embodiments, other stop conditions for the registration using the first trained DNN may be used. For example, the stop condition may be met when the registration using first trained DNN achieves a predetermined accuracy threshold (e.g., an error value between the reduced-resolution reference image and the transformed reduced-resolution moving image is less than a threshold), when the registration using the first trained DNN converges (e.g., the transformation parameters converge or an error value between the reduced-resolution reference image and the transformed reduced-resolution moving image converges), or in a case in which the first trained DNN includes a “stop” action, when the “stop” action is the selected as the action having the highest action value. Other stopping criteria, such as when loops in actions are detected, may be used as well.
  • At step 714, a region of interest (ROI) is defined based on the first trained DNN. The reduced resolution state volume is updated based on the current transformation parameters. A single back-propagation pass is then used to compute the derivative of the sum of the outputs of the layers of the first trained DNN with respect to the input reduced resolution state volume, resulting in a saliency map of the input reduced resolution state image. The resulting saliency map is then thresholded, for example using 95th percentile thresholding, and the weighted mean is calculated on the thresholded saliency map to localize an ROI to be used for the input to the second trained DNN.
  • At step 716, a current high-resolution state volume is generated using the first and second medical images and the current transformation parameters and the defined ROI is extracted from the current high-resolution state volume. The current high-resolution state volume can be generate by subtracting the original resolution moving image transformed by the current transformation parameters from the original resolution reference image. The ROI defined step 714 is then extracted from the current high-resolution state volume. At step 718, action-values are calculated for each of the actions in the set of possible actions using the second trained DNN based on the ROI of the current high-resolution state volume. The set of possible actions can include actions corresponding to adjustments to each of the rigid-body transformation parameters. In particular, the set of possible actions can include 12 actions corresponding to adjustments of ±1 mm for each of the translation parameters [tx, ty, tz] and ±1° for the orientation (rotation) parameters [θx, θy, θz]. The second trained DNN, which is trained using high resolution images in a limited FOV, inputs only a portion of the current high-resolution state volume corresponding to the ROI defined in step 714, and calculates a predicted action-value (Q value) for each action. Since the ROI is defined based on the saliency map of the first trained DNN, the ROI focuses the second trained DNN on the most salient portions of the input volume for image registration. At step 720, the action with the highest action value is selected and the transformation parameters are updated by performing the selected action.
  • At step 722, it is determined whether a stop condition has been met. In an exemplary embodiment, the stop condition for registration using the second trained DNN is met when a second predetermined number N2 of iterations have been performed using the second trained DNN. In this case, if N2 iterations have not yet been performed using the second trained DNN, the stop condition has not been met and the method returns to step 716 and repeats steps 716-722. Accordingly, steps 716-722 are repeated until N2 iterations are performed using the second trained DNN. When N2 iterations have been performed using the second trained DNN, the stop condition has been met and the method proceeds to step 724. In an exemplary implementation, N2=100, such that 100 sequential actions are applied using the second trained DNN. In other embodiments, other stop conditions for the registration using the second trained DNN may be used. For example, the stop condition may be met when the registration using second trained DNN achieves a predetermined accuracy threshold (e.g., an error value between the reference image and the transformed moving image is less than a threshold), when the registration using the second trained DNN converges (e.g., the transformation parameters converge or an error value between the reference image and the transformed moving image converges), or in a case in which the second trained DNN includes a “stop” action, when the “stop” action is the selected as the action having the highest action value. Other stopping criteria, such as when loops in actions are detected, may be used as well.
  • At step 724, the registration result is output. The first and second medical images are registered by transforming the moving image by the final rigid-body transformation to align the moving image with the reference image. The registered first and second medical images can be output by displaying the registered first and second medical images on a display of a computer system. For example, the registered first and second images can be overlaid and displayed as a fused image. The registered first and second medical images can also be output by electronically transmitting the registered first and second images to a remote computer system to be displayed on a display of the remote computer system.
  • Although FIG. 7 only shows outputting the final registration result, it is also possible that the incremental registration results using the current transformation parameters can be output (e.g., displayed on a display of a computer system) as each iteration of steps 704-710 and each iteration of steps 716-722 is performed. Such incremental results would allow a user to view each adjustment made by the intelligent artificial agent in real-time as the artificial agent is registering the images
  • The present inventors evaluated the above described embodiments for 3D/3D rigid medical image registration on registration of abdominal spine CT and cone beam CT (CBCT) images and registration of cardiac CT and CBCT images. In abdominal spine CT and CBCT registration, a main challenge is that CT has a much larger view that CBCT, leading to many local optima in the registration space due to the repetitive nature of the spine. 87 image pairs were used for the abdominal spine CT and CBCT registration, with 82 image pairs used for training and 5 pairs used for testing. In cardiac CT and CBCT registration, a main challenge is the poor quality of CBCT with severe streaking artifacts and weak soft tissue contrast at the boundary of the object to be registered, i.e., the epicardium. 97 image pairs were used for the cardiac Ct and CBCT registration, with 92 pairs used for training and 5 pairs used for testing. Expert manual annotations on spine landmarks and mesh delineation at the epicardium were used to run iterative closest point (ICP)-based registration followed by visual inspection to provide the ground truth alignment for the spine and cardiac image pairs, respectively.
  • The same network architecture was used for the DNN for both the spine and cardiac applications and both registration layers, as follows. The DNN trained for each application includes five convolutional layers followed by three fully connected layers. The five convolutional layers use 8, 32, 32, 128, and 128 filters, respectively, all with 3×3×3 kernels. The first two convolutional layers are each followed by a 2×2×2 max-pooling layer. The three fully connected have 512, 512, and 64 activation neurons, respectively, and the output has 12 nodes corresponding to the twelve possible actions in A. Each layer is followed by a non-linear rectified layer, and batch normalization is applied to each layer. During training, each training image pair was augmented 64,000 times, leading to >5M training samples for each application. To train the DNN for coarse registration, rigid-body perturbation was randomly generated within [±30 mm, ±30 mm, ±30 mm, ±30°, ±30°, ±30° ] for the cardiac data and [±30 mm, ±30 mm, ±150 mm, ±30°, ±30°, ±30° ] for the spine data to cover the large FOV in the head-foot direction in the spine CT. In addition, with a probability of 0.2, rigid-body perturbation was randomly generated within [±10 mm, ±10 mm, ±10 mm, ±10°, ±10°, ±10° ], in order to more densely sample the transformation parameter space close to the correct alignment. To train the CNN for refinement registration, the rigid-body perturbation range was reduced to [±5 mm, ±5 mm, ±5 mm, ±5°, ±5°, ±5° ]. The DNN (both for coarse alignment and for refinement registration) was trained using RMSprop gradient descent update without momentum and a batch size of 32. The learning rate was 0.00006 with a decay of 0.7 every 1000 mini-batch based back propagations. For each case, the training took approximately 24 hours on a single GeForce Titan X.
  • FIG. 8 illustrates exemplary registration results for spine CT and CBCT registration and cardiac CT and CBCT registration. FIG. 8 shows registration results for three spine CT and CBCT registration examples and two cardiac CT and CBCT registration examples. As shown in FIG. 8, images 800 and 802 show the difference between the reference and moving images (i.e., the state image) before and after registration, respectively, for a first spine CT and CBCT registration example. Images 810 and 812 show the difference between the reference and moving images before and after registration, respectively, for a second spine CT and CBCT registration example. Images 820 and 822 show the difference between the reference and moving images before and after registration, respectively, for a third spine CT and CBCT registration example. Arrows 805, 815, and 825 in the first, second, and third spine CT and CBCT registration examples, point to the kidney, the black background outside the imaged volume, and deployed stent graphs, respectively, showing the high robustness of the intelligent agent in dealing with various interfering objects and artifacts during registration. Hierarchical registration using the method of FIG. 7 was applied to each spine example, and boxes 807, 817, and 827 show the detected ROI at which the attention was focused in the refinement registration using the second trained DNN for the first, second, and third spine CT and CBCT examples, respectively. As shown in FIG. 8 images 830 and 832 show the difference between the reference and moving images before and after registration, respectively, for a first cardiac CT and CBCT registration example. Images 834 and 836 show the mesh overlay for the epicardium meshes in the CT and CBCT images before and after registration, respectively, for the first cardiac CT and CBCT registration example. Images 840 and 842 show the difference between the reference and moving images before and after registration, respectively, for a second cardiac CT and CBCT registration example. Images 844 and 846 show the mesh overlay for the epicardium meshes in the CT and CBCT images before and after registration, respectively, for the second cardiac CT and CBCT registration example.
  • The present inventors evaluated the efficiency of the supervised deep RL described herein in comparison with deep RL without supervision on a modified 2D/2D registration problem using the spine data, due to the un-affordable large memory storage required by experience replay in the unsupervised RL for 3D/3D registration. In the modified 2D/2D registration problem, 2D multiplanar reconstruction (MPR) image pairs were extracted from 82 aligned CT and CBCT spine pairs using various MPR cuttings. For the supervised learning (SL), these 2D images were artificially de-aligned in 2D by random perturbations within [±30 mm, ±30 mm, ±30° ] to generate 2M image pairs for training. The network architecture was modified from the 3D/3D registration examples to take 128×128 2D images as the input, and the output has six nodes corresponding to 6 possible actions in changing [tx, ty, θz]. The network architecture and training hyper-parameters were the same for SL and unsupervised RL. For SL, the present inventors also compared the training efficiency and accuracy with and without denser sampling of the transformation parameter space close to the ground truth alignments. FIG. 9 illustrates a comparison of supervised deep reinforcement learning (SL) and unsupervised deep reinforcement learning for the 2D/2D spine registration. FIG. 9 shows a plot of success rate versus the number of training steps for the 2D/2D spine registration using supervised learning (SL) 902, supervised learning without denser sampling 904, and unsupervised RL 906. It can be seen from FIG. 9 that SL is much more efficient than unsupervised RL and achieves significantly better results when the same number of training data are used. In addition, proper data distribution in training samples (by denser sampling close to the ground truth alignments) is shown to be beneficial for the agent to learn the complex mapping efficiently.
  • Deformable Registration of Medical Images Used Intelligent Artificial Agent
  • In an advantageous embodiment of the present invention, an intelligent artificial agent is used to perform deformable registration of medical images. A major difficulty in medical image registration is incorporating non-rigid deformations, which frequently occur in medical imaging due to different patient positioning, pathological and physiological changes (e.g. empty vs. filled bladder), and/or cardiac and respiratory motion. Current deformable registration algorithms based on variational approaches try to counter such problems in finding a transformation that maps one image to the other by solving an objective function typically consisting of a similarity measure (e.g., SSD, mutual information) and differently motivated regularizers (e.g., elastic, fluid, curvature-based models). However, these approaches have difficulties in modeling large deformations and sliding motions along the boundaries of different tissues, such as the lungs. Furthermore, the existing algorithms are engineered for specific applications and do not generalize well for other registration applications. By using deep-learning systems with automated feature design, embodiments of the present invention can overcome these limitations through learning the intrinsic and complex mapping in the spatial and intensity domain. Embodiments of the present invention utilize artificial deep neural networks to extract features from both objects/images to be registered and compare them to assess the similarity, followed by a prediction of an optimal way to change the features of one image should be changed to get closer to the other image. This pattern is applied in a step-by-step approach to approximate the registration in a coarse-to-fine manner. Embodiments of the present invention combine an efficient encoding/decoding of the deformation field with an intelligent artificial agent based trajectory learning that predicts parametric actions operating on the encoded parameters.
  • FIG. 10 illustrates a method for deformable registration of medical images using an intelligent artificial agent according to an embodiment of the present invention. The method of FIG. 10 can be applied to any deformable (non-rigid) registration task. For example, the method of FIG. 10 can be used for applications including, but not limited to, contour propagation between different medical imaging modalities or atlas-based segmentation (in which segmentation is performed by registering an atlas to a medical image a patient) for therapy planning. In contrast to other current deformable registration methods, the method of FIG. 10 is not engineered for a specific application and can be used to perform deformable registration in problematic cases, such as cases involving large deformations or sliding motion.
  • Referring to FIG. 10, at step 1000, a DNN is trained to predict action-values for a set of actions corresponding to parameters of a deformation model used to encode a deformation field. Step 1000 is performed in an offline training stage prior to the on-line image registration for newly input/received images performed in steps 1002-1014. The DNN is trained to predict action-values for the actions based on a reward function that evaluates the current state of the system.
  • In an advantageous embodiment of the present invention, supervised trajectory learning is used to train the DNN. As the ground truth flow field (image deformation field) between two images is typically not available for real medical images, supervised trajectory learning is used to train the DNN to learn steps on the trajectory of small deformations to reach the final deformation field and reward actions corresponding to the small deformations on the learned trajectory. Thus, at each iteration, a small deformation flow field (δDF), which is the outcome of taking one of the best actions based the rewards (action-values) computed by the agent-based network (DNN) and applying the action in the encoded dense deformation field space, is applied to the moving image to warp the moving image to be closer to the fixed image. In an advantageous implementation, the action selection can be performed in a probabilistic way to enable the intelligent artificial agent to not always select the best action in a specific state and allow circuitous routes to avoid local minima (ε-greedy selection). In this case, the probability of each action being selected is based on the action-value (reward) predicted for that action by the DNN. The composition of all of the δDFs over all of the iterations results in the final flow field that warps the moving image to register the moving image to the fixed image. FIG. 11 an exemplary idealized deformation trajectory between a moving image M and a fixed image F. As shown in FIG. 11, in each iteration the artificial intelligent agent selects an action based on the current reward determined by the DNN and applies the δDF corresponding to the selected action to warp the moving image M. In FIG. 11, the moving image is registered to the fixed image F over four iterations, and the composition of δDF1, δDF2, δDF3, and δDF4 result in the final flow field that registers the moving image M with the fixed image F.
  • FIG. 12 illustrates an exemplary DNN for deformable image registration according to an embodiment of the present invention. As shown in FIG. 12, the DNN 1200 inputs a fixed image F and a moving image M. In the embodiment of FIG. 12, the DNN 1200 takes 64×64 2D images as input, but the present invention is not limited thereto. Features of both images are extracted with separate but identical, in points of network architecture, processing streams. In particular, the CNN 1200 uses multiple convolutional layers, whose weights are not shared between the fixed/reference image stream and the moving image stream, for feature extraction. That is separate weights are learned for the convolutional layers that extract features from the fixed image F and the convolutional layers that extracted features from the moving image M. In the embodiment of FIG. 12, the CNN 1200 includes three convolutional layers that use 128, 256, and 512 filters, respectively, in each of the fixed image stream and the moving image stream. The convolutional layers used for feature extraction are followed by fully connected layers, at which the extracted features from the fixed and moving images are compared at a higher level to compute reward values for various possible actions. The fully connected layers act as a learned similarity measure based on the features extracted from the fixed and moving images. In the embodiment of FIG. 12, the CNN 1200 includes three fully connected layers having 500, 128, and 64 activation neurons, respectively. An action prediction is retrieved/selected based on the action-values (rewards) output by the DNN 1200, and the action prediction can then be transferred to a low-dimensional deformation encoding. Dense prediction follows from decoding the few parameters of the encoded action to a delta dense deformation field (δDF) with help from a chose parametric deformation model. The δDF is then directly applied on the moving image to warp the moving image. Dependent on the training data, the reward may be computed directly at the level of encoded components. The DNN framework is trained to predict the next best action based on the current state (determined by the feature extraction from the fixed and moving images) from available actions corresponding to +/− adjustments on the encoded parameters (i.e., the parameters of the parametric deformation model). Coarse-to-fine support may be given by joining actions that are “similar” in parametric space to enable faster convergence, avoid unrealistic states, and reduce dimensionality of otherwise huge parametric spaces. For training of the DNN, the loss function can be computed as:
  • Loss ( s , a , s ) = 1 B i B ( r GT - R ( s , a , s ) ) 2 ( 11 )
  • where s is the current state, a is the action, s′ is the next state if the agent takes action a from state s, rGT is the known ground truth reward, R is the learned reward computed by the trained DNN, and B is the number of landmarks or mesh points on the contour or surface being deformed. The DNN can be trained using a gradient descent algorithm and backpropagation to learn weights for layers of the DNN that minimize the loss function over all of the training samples.
  • A deformation model is used to encode the dense deformation field in order to reduce the dimensionality of the dense deformation field. In one embodiment of the present invention, a parametric deformation model whose parameters are applicable for smooth transitions from one state to its neighbor can be used to encode the dense deformation field. Radial Basis Functions (RBF) are an example of such a parametric deformation model that can be used to encode dense deformation field. FIG. 13 illustrates a Gaussian RBF. As shown in FIG. 13, the Gaussian RBF 1302 can be expressed as
  • φ ( x ) = d * e - ( x - p x ) 2 + ( x - p y ) 2 2 * σ 2 .
  • The Gaussian RBF 1302 models the deformation of pixels in a neighborhood of a location (px,py) as having a Gaussian distribution 1304 centered as (px,py) with magnitude d=(dx,dy) and standard deviation σ. The Gaussian RBF 1302 computes the deformation φ at a point x based on the distance of x from the location (px,py), the magnitude d=(dx,dy), and the standard deviation σ. In a possible implementation, multiple Gaussian RBFs can be used to approximate the dense deformation field for the moving image. In this case the set of possible actions can correspond to adjustments to the Gaussian parameters, such as location and magnitude, for each of the Gaussian RBFs.
  • In another embodiment, a free form deformation (FFD) approach can be used to encode the dense deformation field. In this case, the DDN learns the desired action for moving a limited number of points of a uniform hull while preserving topology. Dense deformation inside the whole is approximated, for example using B-splines or thin plate splines (TPS). In this model, the agent can learn actions that directly modify the B-splines parameters (e.g., direction, amplitude, and extent of the displacement at the B-spline node). In another embodiment, all B-spline parameters are projected onto a subspace representation (e.g. via PCA or any other manifold learning method) and the agent directly modify the transformation parameters in the subspace (e.g. by normalized increment/decrement along a PCA mode).
  • In another embodiment, various types of manifolds that reduce the dimensionality of the action space through encoding the dense deformation field can be used as the deformation model. In this case, the action space of deformations is learned from a set of training examples using manifold learning (linear or non-linear). Multiple possibilities exist for this purpose, such as principal component analysis (PCA), stacked denoising autoencoder, or deep learning based generative models. The agent actions are performed directly on the subspace parameters (e.g., via PCA or any other manifold learning method). The agent directly modifies the transformation parameters in the subspace (e.g., by normalized increment/decrement along a PCA mode) and the resulting deformation field is reconstructed from the updated subspace parameterization.
  • The deformation model can be dense model or a correspondence based model. In one embodiment, the deformation model is a dense model. In this case, the deformation model encodes the entire dense deformation field of the moving image (or image region/image patch) to be registered. For a dense model, in additional the two input images (i.e., fixed and moving images) in each training image pair, ground truth delta deformation fields δDFs are used to train the encoder/decoder and the DNN. δDFs can be small deformation updates or, in the case of diffeomorphic deformation, stationary velocity field updates or even kinetic moment updates in the case of time-varying velocity field representation of the deformation, like in the Large Deformation Diffeormophic Metric Mapping (LDDMM). As only small deformation fields (resulting from applying one action) are required, the ground truth δDFs can be obtained from existing deformable registration algorithms. According to an advantageous implementation, to avoid learning possible insufficient deformation fields resulting from such existing deformable registration algorithms, the algorithm's output can be used to produce a new warped version of the fixed image (generated by warping the moving image by the deformation field output by the algorithm) which can be used in the training instead of the original fixed image. The new fixed training image may be different that the original fixe training image, but by applying this scheme it is guaranteed that the real ground truth deformation field is used for training, as long as any reasonable deformation field is provided by the existing deformable registration algorithm used.
  • In another embodiment, the deformation model is a correspondence based model. In this case, the deformation model encodes a sparse deformation field corresponding to deformations of the particular points (e.g., landmarks or control points) or structures in the moving image that having correspondences in the fixed image. In addition to the two input images (i.e., fixed and moving images) of each training image pair, point correspondences or segmentations of deforming structures in the images are also used for training. The segmentations can be performed manually, automatically, or semi-automatically using various segmentation algorithms. In the case of precise segmentations, exact point correspondences exist and can be used as the ground truth spare deformation field for training. It is also possible to establish point correspondences between the fixed and moving images in the training image pairs using a ray-shooting algorithm. Given a segmentation mask, in order to retrieve distributed points on the surface of a convex structure of interest, a ray-shooting algorithm can be performed that shoots rays from the center of mass to the surface with equal angular distances and selects the points where the rays leave the mask as landmark positions. By applying this procedure, the topology of landmarks is guaranteed. If the landmark positions are extracted in the same way across the whole training data set, the DNN learns the topology in training and may learn to connect rewards to specific landmarks. Once the sparse deformation field for the point correspondences is obtained, the dense flow field can be obtained by interpolation and extrapolation, for example using RBFs. In another possible implementation, a neural network, such as an autoencoder, can be trained to find a mapping between the sparse deformation field and the dense flow field.
  • FIG. 14 illustrates a framework for decoding encoded flow actions output from a trained DNN according to an embodiment of the present invention. As shown in FIG. 14, a fixed image F and a moving image M are input to the trained DNN 1400. The DNN 1400 extracts features from the fixed and moving images and predicts best flow action based on the current state of the fixed and moving images. A deformation model is selected for flow encoding, and the trained DNN 1400 outputs encoded flow actions 1402 in the parameter space/component space of the deformation model used for the flow encoding. The encoded flow action 1402 performed to adjust the encoded flow 1404, and flow decoding is performed to decode the encoded flow 1404 into a decoded flow field 1406. Path A represents the case in which the deformation model is a dense model. In this case, the decoded flow field 1406 is a dense deformation field and is applied to the moving image generate the warped moving image 1408, which is input to the DNN 1400 as the moving image M for the next iteration. Path B represents the case in which the deformation model is correspondence based model. In this case, the decoded flow field is used to transform a mesh (1410) of a segmented structure in the moving image. A dense deformation field is then computed based on the deformation of the mesh using interpolation and/or extrapolation (1412), and the dense deformation field is applied to the moving image to generate the warped moving image 1408, which is input the DNN 1400 as the moving image M for the next iteration.
  • The action design is highly related to the chose encoding method. In particular, for each type of deformation model that can be used for the encoding, a set of actions can be defined to by small adjustments (+/−) to various parameters or components of that deformable model. In an advantageous embodiment, RBFs, such as Gaussian functions are used and a subspace of their parameters space is used as the action space. For example, the location (mean) and magnitude parameters of the Gaussian RBFs can be used to define the set of actions. In particular, the set of actions can be include +/− predetermined adjustments of the location (mean) and magnitude parameters for each Gaussian RBF used to model the deformation of the moving image.
  • In an advantageous implementation, Gaussian kernels can be placed at landmark positions in the moving image, such as at control points on a contour or mesh of a segmented organ of interest, and the agent can be driven to align the contours of the of the organ of interest as in an FFD approach. FIG. 15 illustrates examples of deforming a contour of an organ of interest using Gaussian kernels. As shown in example 1500 of FIG. 15, contour F represents a contour of a segmented target organ in the fixed image and contour M represents a contour of the segmented target organ in the moving images. Gaussian kernels are placed at each of a plurality of landmarks 1501, 1502, 1503, 1504, 1505, 1506, 1507, and 1508 and the parameters of the Gaussian kernels are adjusted to iteratively deform contour M to align contour M with contour F. Coarse-to-fine registration can be reached by combining or joining neighboring Gaussian kernels and moving them together, such as in Example 1510 of FIG. 15. That is when an action is selected to move a Gaussian kernel at one landmark point, the Gaussian kernels at neighboring landmark points may be moved as well. The magnitude of movement can be controlled by a separate Gaussian process such that movements of the neighboring Gaussian kernels decrease with a distance to the chose Gaussian kernel. Neighbor relationships can be fixed by predefining the topology. In order to achieve coarse-to-fine registration, when the contour in the moving image is close to the contour in the fixed, the Gaussian kernels may be moved without the joint movement of the neighboring Gaussian kernels. As shown in example 1510 of FIG. 15, in a first stage of deforming the contour M in the moving image, when an action is selected to move the Gaussian kernel at landmark 1511 by 1 mm, the Gaussian kernels at landmarks 1512 and 1513 are each moved by 0.72 mm and the Gaussian kernels at landmarks 1514 and 1515 are each moved by 0.26 mm. In the last stage, when the contour M is close to the contour F, the same action only causes the Gaussian kernel at landmark 1511 to move 1 mm.
  • In an advantageous implementation, a fixed step size can be used for action values on a pixel level (e.g., +/−1 pixel movements) that can be adjusted through the above described joint action method. The step size can be used to control convergence speed in contrast to precision. Bigger step sizes may be used in early deformations to roughly approach the fixed structure and then smaller step sizes can be used combined with smaller impact on the neighbors and joining less actions to reach a finer registration. In another possible implementation, adjusted step sizes may be applied in relation to the reward either using a fixed set of step sizes or using continuous action spaces. In the case of using principal components actions, which are either increasing or decreasing the components action values, a fixed number (e.g., +/−1) can be used to increase or decrease the components action values. As the action space can be large, and a greedy agent exploration scheme may be computationally inefficient and lead to long training times, sparse exploration schemes can be used. This can be achieved by applying a hierarchical approach on the parameters (multi-scale parameter representation). Different DNNs are trained with different step sizes. The first DNN with largest step size starts, the second takes over, and so on. Alternatively, an actor-critic architecture can be applied, where the actor (a first trained DNN) decides which action to take and the critic (a second trained DNN) evaluates the value of the action. This makes the training of continuous step sizes possible, as the two deep neural networks are trained iteratively, one to pick the action, the second one to choose the step size.
  • Various types of reward functions can be used for training the DNN. The reward function may be based on human-level recognition or may be a reward that is directly obtained from available ground truth deformation fields. Direct reward functions require the knowledge of the ground truth deformation field during training. As such ground truth data is not typically available for real medical image data, artificially generated deformation fields which are generated under controlled conditions can be used instead. In one embodiment, in the case of parametric deformation models, the distance of encoded components: ∥cgt−ccurrent2−∥cgt−cnext2 can be directly used as the reward, where cgt is the ground truth encoded components or parameters of the deformation model.
  • Reward functions can also be designed based on human-level recognition, that is, the deformation is considered satisfactory when the image content (anatomic structures and boundaries) are well aligned. In one embodiment, the image difference: −Σ|Iref−Imov(t)|2 can be used as the reward function. In another embodiment, landmark/mesh differences: Σ∥Pref−Pmov(t)2 can be used as the reward function, where Pref denotes a set of landmarks or mesh points in the reference image and Pmov(t) denotes the set of landmarks or mesh points in the current moving image. In another embodiment, the image difference+the landmark/mesh differences: Σ∥Pref−Pmov(t)2−μΣ|Iref−Imov(t)) can be used as the reward function, where μ is a weighting factor. In another embodiment, instead of the image difference, a deep learned similarity measure: sim(Iref−Imov(t) can be used as the reward function.
  • Once the DNN is trained (in step 1000 of FIG. 10) to predict action values for the actions in the defined action space, the trained DNN is then used in the registration of newly received images. Returning the FIG. 10, at step 1002, first and second medical images are received. For example, the first and second medical images can be 2D or 3D images acquired using different imaging modalities (e.g., CT, MRI, ultrasound, PET, DynaCT, etc.), images of the same patient acquired at different times, or images acquired from different patients. The first and second medical images may be received directly from image acquisition devices, such as a CT scanner, MRI scanner, ultrasound device, C-arm image acquisition device, etc., or may be received by loading previously stored medical images from a memory or storage of a computer system or receiving the medical images in an electronic transmission from another computer system. One of the medical images is designated as the reference/fixed image Iref and the other is designated as the moving image Imov.
  • A pre-processing stage may be performed before proceeding to step 1004. For example, one or more anatomical objects of interest (e.g., organs) can be segmented in each of the first and second medical images in order to define contours or meshes in the first and second images so that a contour or mesh in the moving image can be deformed to match the contour or mesh in the fixed image. A landmark or point of interest, such as a center of mass of an organ of interest, can be detected in each of the first and second images, and the first and second images can be initially aligned with respect to the point of interest.
  • At step 1004, action values for possible actions can be calculated based on the current by the trained DNN. The trained DNN inputs the fixed image and the current moving image which is warped based on a current dense deformation field. In the first iteration, the current moving image is not yet warped. The trained DNN extracts features from the fixed and moving images, which provide an observation of the current state, and computes action-values for actions corresponding to adjustments of parameters in a parameter space of a deformation model used to encode the deformation of the moving image. The trained DNN may apply a probabilistic model to determine the action-values for the set of actions.
  • At step 1006, an action is selected based on the action-values calculated by the trained DNN. The action corresponds to an adjustment to a parameter in the parameter space of the deformation model used to encode the deformation of the moving image. The action with the highest predicted action-value may be selected.
  • At step 1008, a delta dense deformation field δDF is computed based on the selected action and the deformable model used to encode the deformation of the moving image. The δDF is a small dense deformation field corresponding to the selected action that provides an incremental adjustment to the overall deformation of the moving image. In one embodiment, when the deformation model used to encode the deformation of the moving image is a dense model that represents an entire deformation field of the moving image, the adjustment to the parameter in the parameter space of the deformation model corresponding to the selected action is performed and the change in the parameter space of the deformation model is decoded which maps the change in the parameter space of the deformation model to the δDF. In another embodiment, the deformation model used to encode the deformation of the moving image is a correspondence-based model that models deformation of landmark or control points of a structure of interest in the moving image. In this case, the adjustment to the parameter in the parameter space of the deformation model corresponding to the selected action is performed and the change is the parameter space of the deformation model is decoded, which maps the change in the parameter space of the deformation model to movement of the landmarks or control points to deform a contour or mesh of the structure of interest in the moving image. The δDF for the entire moving image is then calculated from the deformation of the contour or mesh in the moving image, for example using interpolation and/or extrapolation.
  • At step 1010, the delta dense deformation field δDF is applied to the moving image to warp the moving image. At step 1012, it is determined whether a stop condition has been met. For example, the stop condition can be met a predetermined maximum number of iterations has been performed, when the moving image converges, or when an error value between the fixed and moving images or between landmarks or mesh points in the fixed and moving images is less than a threshold value. If the stop condition has not yet been met, the method returns to step 1004 and repeats steps 1004-1012. Accordingly, the moving image is iteratively warped until the stop condition is met. When the stop condition is met, the method proceeds to step 1014.
  • At step 1014, the registration result is output. The registered first and second medical images (i.e., the fixed image and the final warped moving image) can be output by displaying the registered first and second medical images on a display of a computer system. The registered first and second medical images can be displayed in the same coordinate system. For example, the registered first and second images can be overlaid and displayed as a fused image. In addition to displaying the registered first and second images, the final dense deformation field used to warp the moving image, which is a composition of all the δapplied to the moving image, can be visualized and displayed on the display of the computer system. The individual δDFs applied to the moving image can also be visualized and displayed. The registered first and second medical images, as well as the final dense deformation field and/or the individual δDFs can also be output by being electronically transmitting to a remote computer system to be displayed on a display of the remote computer system.
  • Although FIG. 10 only shows outputting the final registration result, it is also possible that the incremental registration results resulting from warping the moving image using the current dense deformation field can be output (e.g., displayed on a display of a computer system) as each iteration of steps 1004-1012 is performed. Such incremental results would allow a user to view each adjustment made by the intelligent artificial agent to the deformation of the moving image in real-time as the artificial agent iteratively warps the moving image to register the moving image to the fixed image.
  • As described above the point correspondence between the fixed and the moving images can be established using a ray-shooting algorithm that, given a segmentation mask of an object of interest shoots rays from the center of mass of the object to the surface with equal angular distances and selects the points where the rays leave the segmentation mask and landmark positions. This ray-shooting algorithm can be performed in the pre-processing phase prior to performing the registration. As these landmarks are not anatomically distinct points and are more utilized as regions for which actions can be selected by the trained DNN, this approach can be extended to shoot rays multiple times for the same landmark region with slightly varying angles from the center. That is, in a possible embodiment, the registration using the method of FIG. 10 can be performed multiple times using different landmarks acquired by ray-shooting with different angles to drive the registration to provide better representation of the shape of the object of interest in between the original landmark locations. The registration results from the various registrations can then be combined. After running the registration process individually for each set of landmarks, multiples of the number of landmarks can be retrieved as point correspondences, which can result in a more accurate registration.
  • Hierarchical Deformable Registration Using Higher Resolution Image Patches.
  • In an advantageous embodiment, in order to increase the registration accuracy a hierarchical (multi-stage) framework based on higher resolution image patches may be utilized to perform deformable image registration. The first stage of the multi-stage deformable image registration method is performed on the full moving and fixed images using a first trained DNN. For example, the first stage can be performed as describe above in the method of FIG. 10. A second stage performs refinement then performs refinement of the registration estimated in the first stage based on higher resolution image patches extracted from the fixed and moving images using a second trained DNN. After the first stage registration on the full moving and fixed images, patches centered at the new landmark positions of the moving image are extracted from higher resolution versions of the fixed image and the current moving image. For example the higher resolution versions of the fixed and moving images can high-resolution images generated by up-sampling the fixed and moving images. Alternatively, the first stage can be performed on reduced resolution fixed and moving images generated by down-sampling the fixed and moving images acquired from the image acquisition devices, and the higher resolution versions are the original resolution fixed and moving images acquired from the image acquisition devices. FIG. 16 illustrates patch extraction from landmark positions of the moving image. As shown in FIG. 16, patches 1602, 1604, 1606, 1608, and 1610 are extracted at landmark locations in the moving image. Each of the image patches 1602, 1604, 1606, 1608, and 1610 is centered at the respective landmark and is oriented to be aligned with a normal of the contour at the respective landmark location. The image patches are extracted from the same locations in the fixed image.
  • FIG. 17 illustrates a framework for performing the second stage of the multi-stage deformable registration according to an embodiment of the present invention. As shown in FIG. 17, the image patches (F) extracted from the higher resolution fixed image and the image patches (M) extracted from the higher resolution moving image are input to a trained DNN 1700. All of the patches are input to the DNN 1700 in different channels. The deformation model, reward system, and action definitions can stay the same as in the first stage. However, the image warping is executed on a higher resolution. As the network is likely slower due to patch extraction, multiple channels, and higher resolution warping, in an advantageous embodiment, the method of FIG. 17 is used as a refinement step after the first stage is used to perform rough registration. The DNN 1700 extracts features from the input higher resolution image patches and calculates rewards for the set of actions. The best action is selected and the higher resolution moving image is warped based on the selected action, which results in movement of one or more of the landmark positions. Updated image patches are obtained from the higher resolution moving image at the locations of the updated landmark locations in the moving image, and updated image patches at the same locations are obtained from the higher resolution fixed image. This process then repeats until a stop condition is reached. The step size for the actions can also be increased to reduce the number of steps at the higher resolution or decreased to reach a higher accuracy.
  • Clinical Workflow for Autonomous Artificial Intelligence Based Image Registration
  • Due to the vast number of applications to which image fusion/registration can be applied, it is challenging to develop a general method that works robustly for all cases. According to an advantageous embodiment of the present invention, various trained intelligent artificial agents can be used to autonomously implement a clinical workflow that recognizes the registration task or tasks to be performed to achieve fusion of two or more medical images without the need of user interaction. FIG. 18 illustrates a method of autonomous artificial intelligence (AI)-based image registration according to an embodiment of the present invention. The method of FIG. 18 utilizes various intelligent artificial agents to perform various tasks related to image registration. Each intelligent artificial agent is implemented by one or more processors of one or more computer systems executing computer program instructions defining operations the artificial agent. In order to perform its designated task, each artificial agent autonomously observes a state of its environment and performs actions learned via artificial intelligence/machine learning techniques.
  • At step 1802, two or more medical images to be fused are received. The medical images may be 2D or 3D images acquired using any medical imaging modality. The medical images may be received directly from one or more medical image acquisition devices, received by loading previously stored medical images from a memory or storage of a computer system, or received via one or more electronic transmissions from a remote computer system.
  • At step 1804, an imaging modality is detected for each of the medical images to be fused. A trained intelligent artificial agent automatically detects the imaging modality (e.g., CT, MRI, Ultrasound, DynaCT, PET, etc.) for each image to be fused. In an exemplary implementation, the trained artificial agent can utilize one or more trained deep neural networks (DNN) to detect the imaging modalities of the medical images.
  • At step 1806, body parts are detected in each of the medical images to be fused. For each image, a trained intelligent artificial agent corresponding to the imaging modality detected for that agent automatically detects a region of the body (e.g., cardiac, abdominal, head, full body, etc.) present in that image. In an alternative implementation, it is possible that an intelligent artificial agent can be trained to perform the modality and the body region detection in the same step. In addition to detecting the body region present in each image, a trained artificial agent corresponding to the detected body region and imaging modality can be used to detect and/or segment specific anatomical structures, such as organs, vessels, landmarks, etc. in each image. For example, a trained artificial agent may use deep learning based segmentation.
  • At step 1808, rigid-body alignment of the medical images is performed by a trained artificial agent. In particular, for a given pair of images (or given configuration of multiple images), a trained artificial agent corresponding to the right configuration (i.e., imaging modality and anatomy) aligns the images by performing rigid-body registration of the images, as described above in the embodiment FIG. 5. Multiple intelligent artificial agents can be trained using supervised deep reinforcement learning as described above, each trained to perform rigid-body registration for a different configuration of imaging modality and anatomy. In a possible implementation, the medical images can be down-sampled to generate reduced-resolution images can the rigid registration can be performed based on the reduced-resolution images for fast alignment.
  • At step 1810, an accuracy analysis of the alignment is performed and it is determined whether the accuracy of the alignment is sufficient. A trained artificial agent can evaluate the accuracy of the alignment and assign an accuracy measure or confidence value to the alignment. The trained artificial agent may also be trained to consider whether the alignment would be improved by deformable registration. If the accuracy of the alignment is determined to be sufficient, the method proceeds to step 1818. If the accuracy of the alignment is determined to be insufficient and considered to be improvable by deformable registration, the method proceeds to step 1812.
  • At step 1812, sparse parametric alignment of the medical images is performed by a trained artificial agent. In particular, a trained artificial agent corresponding to the right configuration (i.e., imaging modality and anatomy) can perform the sparse deformable image registration of the medical images using the method of FIG. 10 with a sparse parametric deformation model. For example, a thin plate spline (TPS) model can be used as the deformation model. The trained artificial agent can be trained using supervised trajectory learning, as described above, with the movement (action) in each step corresponding to movement of one of the TPS control points. The total maximum movement of a TPS control point can be constrained since rigid-body alignment is performed beforehand.
  • At step 1814, an accuracy analysis of the alignment is performed and it is determined whether the accuracy of the alignment is sufficient. A trained artificial agent can evaluate the accuracy of the alignment and assign an accuracy measure or confidence value to the alignment. The trained artificial agent may also be trained to consider whether the alignment would be improved by dense deformable registration. If the accuracy of the alignment is determined to be sufficient, the method proceeds to step 1818. If the accuracy of the alignment is determined to be insufficient and considered to be improvable by dense deformable registration, the method proceeds to step 1816.
  • At step 1816, dense field alignment of the medical images is performed by a trained artificial agent. In one embodiment, a trained artificial agent corresponding to the right configuration (i.e., imaging modality and anatomy) can perform the sparse deformable image registration of the medical images using the method of FIG. 10 with a dense deformation model. Although not shown in FIG. 18, in a possible implementation, a final alignment accuracy evaluation may be performed by a trained artificial agent before proceeding to step 1818.
  • At step 1818, modality and anatomy specific fusion is performed for the aligned medical images. A trained intelligent artificial agent fuses the information from all of the images using learned proper visualization and quantification methods according to the detected modality, anatomy, and alignment accuracy. For example, the trained artificial agent can tune cinematic rendering based on the fused knowledge for hyper-realistic accuracy. Data enhancement can be achieved by the trained artificial agent selecting, for each voxel, which information (from which of the aligned medical images) to show, such as an ultrasound image with scar tissue highlighted.
  • At step 1820, procedure phase detection and phase-specific fusion are performed. A trained intelligent artificial agent detects the phase of the procedure (e.g., surgery/intervention) based on the medical images and fuses information associated with the current phase in the aligned medical images.
  • At step 1822, longitudinal trend analysis and prediction are performed. A trained intelligent artificial agent learns trends from longitudinal data and predicts a next step to guide the procedure based on the learned trends.
  • According to an advantageous implementation, a “master” intelligent artificial agent can be trained to control the workflow of FIG. 18. Various trained artificial agents for various registration tasks and various modality/anatomy configurations can be stored in one or more databases on a storage device of a computer system or networked “cloud”-based storage. The master intelligent artificial agent is trained to determine which other trained intelligent artificial agents are retrieved to perform the various steps of the method of FIG. 18.
  • Direct Regression of Regression Parameters
  • In the above described embodiments, in order to perform registration of medical images, artificial agents use trained DNNs to iteratively select actions that adjust the transformation parameters. In an alternative embodiment, medical image registrations can be performed using a DNN that is trained to directly regress the transformation parameters given two (or more) input images to be registered. In this case, instead of iteratively selecting registration actions, the first and second images (i.e., fixed and moving images) are input into the trained DNN and the trained DNN directly outputs the transformation parameters that register the moving image to the fixed image. In an exemplary implementation, for a parametric transformation (e.g., rigid transformation, affine transformation, or parametric deformation), a deep neural network for regression can be trained based on training pairs with known ground truth transformations with the loss function being calculated on the transformation parameters. In another exemplary implementation, for dense deformation fields, a deep encoder-decoder can be trained based on trained pairs with known ground truth transformations to directly output the deformation field. A method for training such a deep neural network using weakly supervised deep dense correspondence learning is described below.
  • Deformable Image Registration Using Weakly Supervised Deep Dense Correspondence Learning
  • Deep Learning methods typically require a large number of labeled images to supervise the training, and such data is often not available and difficult to obtain for deformable image registration (i.e., estimating dense correspondence between images). Specifically, the “label” for deformable image registration is a dense vector field, which is very difficult to manually annotate. In an advantageous embodiment of the present invention, cycle consistency of dense correspondences among multiple images is exploited to supervise deep dense correspondence learning for deformable image registration without the need for manually annotated labels.
  • FIG. 19 illustrates a method of training a deep neural network (DNN) for deformable image registration using weakly supervised deep dense correspondence learning according to an embodiment of the present invention. At step 1902, a first pool (A) of medical images with unknown dense correspondence is obtained. The medical images in pool A can be obtained from a database of stored medical images. Alternatively, the medical images in pool A can be obtained by acquiring images using an image acquisition device. The medical images in pool A can be any type of medical images depending on a registration task for which the DNN is being trained.
  • At step 1904, a second pool (B) of medical image pairs with known dense correspondence is obtained. The medical image pairs can be acquired using one or more image acquisition devices or can be obtained from a database of stored medical images. The size of pool B will likely be much smaller than the data pool (A) with unknown dense correspondence. Various techniques can be used to obtain the dense correspondence between the medical image pairs in pool B. In one possible implementation, the dense correspondence between the medical image pairs can be manually annotated. In another possible implementation, one medical image can be artificially deformed via a deformation model, such as a spline model, statistical model biomechanical model, etc. The original image and the deformed image form a pair with a known dense correspondence. In another possible implementation, image pairs can from the same patient can be acquired with a certain motion tracking method that provides the dense correspondence between the images. For example, tagged cardiac MRIs, or CT/MR images acquired with respiratory and/or cardiac phase recorded as a surrogate signal of the underlying motion can be used to obtain paired images with known dense correspondence.
  • At step 1906, M image pairs that form a loop are selected from pool B. For example, image pairs (Ia,Ib), (Ib,Ic), (Ic,Ia) can be selected from pool B if M=3. The dense correspondences for the selected image pairs (Ia,Ib), (Ib,Ic), (Ic,Ia) are denoted as Fab, Fbc, Fca, respectively. The image pairs can be picked randomly. M can be any integer greater than 1. In an exemplary implementation, M=3 and the method of FIG. 19 is described using the example of M=3.
  • At step 1908, N images are selected from pool A. The N images selected from pool A are referred to as I1,I2, . . . , IN. These images can be selected randomly from pool A.
  • At step 1910, the image pairs selected from pool B and the images selected from pool A are combined to form a training dataset. In particular, the images from pool B and pool A are combined and order as Ia, I1, . . . , Ik, Ib, Ik+1, . . . , Ij, Ic, Ij+1, . . . , IN. The images Ia, I1, . . . , Ik, Ib, Ik+1, . . . , Ij, Ic, Ij+1, . . . , IN and the dense correspondences Fab, Fbc, Fca form one training dataset.
  • At step 1912, steps 1906-1910 are repeated multiple times to generate a plurality of training datasets. Each of these training datasets formed by combining the image pairs selected from pool B and the images selected from pool A is treated as a single training example for training the DNN. Accordingly, in an advantageous implementation, steps 1906-1910 are repeated a large number of times to generate a large number of training datasets (training examples).
  • At step 1914, a deep neural network (DNN) is constructed. In particular, a DNN is constructed that takes two images an input, and outputs a 3D vector field, which represents the dense correspondence between the two input images. In one embodiment, the DNN can be a Convolutional Encoder-Decoder Network. FIG. 20 illustrates an exemplary network architecture for Convolutional Encoder-Decoder Network 2000 for dense correspondence estimation. In another embodiment, the DNN can use a FlowNet-like architecture. FIG. 21 illustrates an exemplary FlowNet-like deep architecture 2100 for dense correspondence estimation. Other types of DNN architectures can be used as well. The DNN is denoted as FW(•,•), where W denotes the weights of the DNN that are learned in training.
  • Returning to FIG. 19, at step 1916, for each training dataset, dense correspondences are estimated for each neighboring pair of images in the training dataset. For a given training dataset, the DNN is applied on every neighboring pair of images in (Ia, I1, . . . Ik, Ib, Ik+1, . . . , Ij, Ic, Ij+1, . . . , IN), resulting in estimated an estimated dense correspondence being calculated by the DNN for each neighboring image pair in the training dataset, i.e., FW(Ia,I1), FW(I1,I2), . . . , FW(IN−1,IN), FW(IN,Ia). The first and last images in the training dataset Ia and IN, respectively, are considered to be neighboring.
  • At step 1918, distance metrics are computed to compare the known dense correspondences for the image pairs selected from pool B in each training dataset to dense correspondences artificially generated by concatenating the dense correspondences estimated by the DNN. In a given training dataset, the dense correspondences estimated for the neighboring image pairs by the DNN in step 1916 can be concatenated to artificially generate dense correspondences for the image pairs (Ia,Ib), (Ib,Ic), (Ic,Ia) selected from pool B in the training dataset. In particular, the artificially generated dense correspondence for (Ia,Ib) can be calculated by concatenating the estimated dense correspondences for the neighboring image pairs from Ia to Ib, FW(Ia,I1), FW(I1,I2), . . . , FW(Ik−1,Ik), FW(Ik,Ib). The artificially generated dense correspondence for (Ib,Ic) can be calculated by concatenating the estimated dense correspondences for the neighboring image pairs from Ib to Ic, FW(Ib,Ik+1), FW(Ik+1,Ik+2), . . . , FW(Ij−1,Ij), FW(Ij,Ic). The artificially generated dense correspondence for (Ic,Ia) can be calculated by concatenating the estimated dense correspondences for the neighboring image pairs from Ic to Ia, FW(Ic,Ij+1), FW(Ij+1,Ij+2), . . . , FW(IN−1,IN), FW(IN,Ia). Distance metrics are calculated to compare to known dense correspondences Fab, Fbc, Fca for (Ia,Ib), (Ib,Ic), (Ic,Ia) with the artificially generated dense correspondences for (Ia,Ib), (Ib,Ic), (Ic,Ia):

  • D ab =D(F ab,(F W(I a ,I 1),F W(I 1 ,I 2), . . . ,F W(I k−1 ,I k),F W(I k ,I b))).

  • D bc =D(F bc,(F W(I b ,I k+1),F W(I k+1 ,I k+2), . . . ,F W(I j−1 ,I j),F W(I j ,I c))).

  • D ca =D(F ca,(F W(I c ,I j+1),F W(I j+1 ,I j+2), . . . ,F W(I N−1 ,I N),F W(I N ,I a))).
  • At step 1920, the DNN is trained based on a loss function constructed from the distance metrics. If the dense correspondences produced by the DNN are correct, i.e., they represent the real correspondences between the images, the concatenated dense correspondence between two images should be the same as the real dense correspondence. This property is referred to as Cycle Consistency. The distance metrics Dab, Dbc, Dca are used to construct the loss function for training the DNN. For example, the loss function can be a linear combination of Dab, Dbc, Dca. The DNN is then trained (e.g., using backpropagation and gradient descent techniques) to learn weights that minimize the loss function over all of the training datasets.
  • FIG. 22 illustrates a method for deformable registration of medical images according to an embodiment of the present invention. At step 2202, first and second medical images are received. At step 2204, a DNN trained using weakly supervised deep dense correspondence learning calculates a dense correspondence between the first and second medical images. In particular, the first and second medical images are input to a DNN trained using the method of FIG. 19, and the trained DNN outputs a 3D vector field representing the dense correspondence (i.e., dense deformation field) between the first and second images. At step 2206, the first and second medical images are registered using the dense correspondence. For example, the first image can be registered to the second image by warping the first medical image using the dense correspondence between the first and second medical images. At step 2208, registration result is output, for example, by displaying the registered first and second images on a display of a computer system.
  • The above described methods for intelligent agent based image registration can be applied to many different medical image registration applications. Various registrations applications for which the above described methods for intelligent agent based image registration can be advantageously applied are described herein. It is to be understood that the above described methods are not limited to these examples and may be applied for other medical imaged registrations applications as well.
  • In one embodiment, the above described methods for intelligent agent based image registration may be applied to perform fusion of ultrasound and volumetric cardiac images (e.g., MR or CT) for joint function, flow, anatomy, and coronary assessment. Previous registrations techniques require explicitly defining a segmentation of a coronary structure (e.g., full heart, chamber(s), valve) or anatomical landmarks. The above described methods enable the fusion of ultrasound and volumetric cardiac image (e.g., MR or CT) without the need to explicitly define a segmentation or landmarks.
  • In another embodiment, the above described methods of intelligent agent based image registration may be applied to register CT or MR images to ConeBeam CT images. Interventional ConeBeam CT is often used to guide surgeries or interventions (e.g., cardiac interventions, spine surgery, etc.). Interventional ConeBeam CT often suffers from image artifacts dues to device, which makes robust image registration difficult using previous registration techniques. The above described methods provides robust registration of CT or MR images to Conebeam CT images even in images with noise or image artifacts, which allows to image data from the pre-operative CT or MR images to be fused with the interventional ConeBeam CT images used for guiding an intervention or surgical procedure.
  • In another embodiment, the above described methods of intelligent agent based image registration may be applied to perform automatic change detection through deformable registration. For example, a previously acquired medical image of a patient and a follow-up medical image of the same patient can be registered using the above described methods for intelligent agent based deformable image registration in order to determine a deformation field that shows changes to the anatomy of the patient over time. Such automatic change detection can be used in cancer screening to detect tumor growth or reduction over time, to track changes in stenoses, plaques, or other abnormalities in coronary arteries or other vessels, or to track changes in any other anatomical structure of patient due to disease or treatment.
  • In another embodiment, the above described methods for intelligent agent based image registration can be used to perform motion compensation. For example, when registering a pre-operative medical image (e.g., CT or MR) to intraoperative/interventional medical images acquired during a procedure, a trained intelligent artificial agent for image registration can be kept always running in a perpetual state as real-time interventional medical images are acquired and the images are always registered even when movement occurs in the real-time interventional images (e.g., due to patient movement, respiratory motion, or cardiac motion). Using the above described methods, the intelligent artificial agent automatically re-registers the images as the interventional medical images are received in real-time, such that the registration immediately “catches up” to the movement in the real-time images, thus allowing for real-time perpetual compensation of respiratory motion.
  • The above-described methods for intelligent agent based image registration can be implemented on one or more computers using well-known computer processors, memory units, storage devices, computer software, and other components. A high-level block diagram of such a computer is illustrated in FIG. 23. Computer 2302 contains a processor 2304, which controls the overall operation of the computer 2302 by executing computer program instructions which define such operation. The computer program instructions may be stored in a storage device 2312 (e.g., magnetic disk) and loaded into memory 2310 when execution of the computer program instructions is desired. Thus, the method steps illustrated in FIGS. 2, 3, 4, 5, 6, 7, 10, 12, 14, 17, 18, 19, and 22 may be defined by the computer program instructions stored in the memory 2310 and/or storage 2312 and controlled by the processor 2304 executing the computer program instructions. One or more image acquisition devices 2320, such as a CT scanning device, C-arm image acquisition device, MR scanning device, Ultrasound device, etc., can be connected to the computer 2302 to input image data to the computer 2302. It is possible that the computer and one or more of the image acquisition devices 2320 may be implemented as one device. It is also possible that the image acquisition devices 2320 and the computer 2302 communicate wirelessly through a network or wireless communication protocol. In a possible embodiment, the computer 2302 may be located remotely with respect to the image acquisition devices 2320 and may perform some or all of the method steps of FIGS. 2, 3, 4, 5, 6, 7, 10, 12, 14, 17, 18, 19, and 22 as part of a server or cloud based service. The computer 2302 also includes one or more network interfaces 2306 for communicating with other devices via a network. The computer 2302 also includes other input/output devices 908 that enable user interaction with the computer 2302 (e.g., display, keyboard, mouse, speakers, buttons, etc.). One skilled in the art will recognize that an implementation of an actual computer could contain other components as well, and that FIG. 9 is a high level representation of some of the components of such a computer for illustrative purposes.
  • The above described methods for intelligent artificial agent based image registration and/or training deep neural networks may be implemented in network-based cloud computing system. In such a network-based cloud computing system, a server communicates with one or more client computers via a network. A client computer may communicate with the server via a network browser application residing and operating on the client computer, for example. A client computer may store data on the server and access the data via the network. A client computer may transmit requests for data, or requests for online services, to the server via the network. The server may perform requested services and provide data to the client computer(s). The server may also transmit data adapted to cause a client computer to perform a specified function, e.g., to perform a calculation, to display specified data on a screen, etc. Certain steps of the above described methods may be performed by a server or by other computers/processors in the network-based cloud-computing system. Certain steps of the above described methods may be performed locally by a client computer in a network-based cloud computing system. The steps of the above described methods for intelligent artificial agent based image registration and/or training deep neural networks may be implemented in network-based cloud computing system may be performed by the network-based cloud-computing system or by a local client computer in any combination. In one embodiment, one or more trained artificial intelligent agents are stored on a server or other remote computer device in the network-based cloud-computing system, and a client computer electronically transmits two or more medical images to be registered to the server of the network-based cloud-computing system. The server and/or other computers/processors in the network-based cloud-computing system performs the registration of the medical images using the one or more trained artificial agents and electronically transmits the registration results to the client computer, which then displays the registration results on a display of the client computer. The method steps to perform the registration in the network-based cloud-computing system can be performed by a single computer device in the network-based cloud-computing system (e.g., the server) or may be distributed on multiple computer devices or processors in the network-based cloud-computing system.
  • The foregoing Detailed Description is to be understood as being in every respect illustrative and exemplary, but not restrictive, and the scope of the invention disclosed herein is not to be determined from the Detailed Description, but rather from the claims as interpreted according to the full breadth permitted by the patent laws. It is to be understood that the embodiments shown and described herein are only illustrative of the principles of the present invention and that various modifications may be implemented by those skilled in the art without departing from the scope and spirit of the invention. Those skilled in the art could implement various other feature combinations without departing from the scope and spirit of the invention.

Claims (39)

1. A method for intelligent artificial agent based registration of medical images, comprising:
determining a current state observation of an artificial agent based on a plurality of medical images and current transformation parameters;
calculating action-values for a plurality of actions available to the artificial agent based on the current state observation using a machine learning based model trained based on a plurality of registered training images, wherein the plurality of actions correspond to predetermined adjustments of the transformation parameters;
selecting an action having a highest action-value from the plurality of actions and adjusting the transformation parameters by the predetermined adjustment corresponding to the selected action;
repeating the determining, calculating, and selecting steps for a plurality of iterations; and
registering the plurality of medical images using final transformation parameters resulting from the plurality of iterations.
2. The method of claim 1, wherein the plurality of medical images includes a first medical image and a second medical image and registering the plurality of medical images using final transformation parameters resulting from the plurality of iterations comprises:
transforming the second medical image using the final transformation parameters to align the second medical image with the first medical image.
3. The method of claim 1, wherein the plurality of medical images includes a first medical image and a second medical image and determining a current state observation of an artificial agent based on a plurality of medical images and current transformation parameters comprises:
determining the current state observation by comparing the second medical image transformed by the current transformation parameters and the first medical image.
4. The method of claim 3, wherein determining the current state observation by comparing the second medical image transformed by the current transformation parameters and the first medical image comprises:
determining the current state observation by subtracting the second medical image transformed by the current transformation parameters from the first medical image.
5. The method of claim 1, wherein the transformation is a rigid-body transformation and the plurality of actions comprises actions corresponding to predetermined adjustments of translation and rotation rigid-body transformation parameters.
6. The method of claim 5, wherein the plurality of actions comprises respective actions corresponding to increasing and decreasing each of the translation and rotation rigid-body parameters by a constant or time-varying step size.
7. The method of claim 6, wherein the transformation is a 3D rigid-body transformation and the rigid-body transformation parameters include three translation parameters and three rotation parameters.
8. The method of claim 6, wherein the transformation is a 2D rigid-body transformation and the rigid-body transformation parameters include two translation parameters and one rotation parameter.
9. The method of claim 1, wherein calculating action-values for a plurality of actions available to the artificial agent based on the current state observation using a machine learning based model trained based on a plurality of registered training images comprises:
calculating action-values for the plurality of actions available to the artificial agent based on the current state observation using a trained deep neural network (DNN).
10. The method of claim 9, wherein the trained DNN is trained to predict action-values for the plurality of actions based on a plurality of training image pairs with known ground truth transformation parameters using reinforcement learning in which, for each training image pair, a reward for each action of the plurality of actions at a given state is used to train the DNN to learn an optimal registration policy.
11. The method of claim 10, wherein the trained DNN is trained using supervised reinforcement learning, in which the reward for each action of the plurality of actions at a given state is directly calculated based on a decrease in a geodesic distance between the transformation parameters at the given a state and the ground truth transformation parameters for the training image pair resulting from applying the action.
12. The method of claim 11, wherein the plurality of training image pairs includes training image pairs synthetically generated by artificially de-aligning other training image pairs using randomly generated perturbations to the ground truth transformation parameters of the other training image pairs.
13. The method of claim 11, wherein the plurality of training image pairs includes training image pairs synthetically generated from other training image pairs to provide dense sampling close to the ground truth transformation parameters by co-deforming the other training image pairs by randomly generated affine transformations within a range of the ground truth transformation parameters.
14. The method of claim 11, wherein the plurality of training image pairs includes training image pairs synthetically generated by artificially altering image appearance of other training image pairs using synthetic image generators.
15. The method of claim 1, wherein the plurality of medical images includes a first medical image and a second medical image, the transformation is a dense deformation model that models deformations applied to the second medical image, the plurality of actions comprises actions corresponding to adjustments to parameters of the deformation model, and registering the plurality of medical images using final transformation parameters resulting from the plurality of iterations comprises:
generating a dense deformation field corresponding to final parameters of the deformation model resulting from the plurality of iterations; and
warping the second medical image to register the second medical image with the first medical image by applying the dense deformation field corresponding to the final parameters of the deformation model.
16. The method of claim 1, wherein calculating action-values for a plurality of actions available to the artificial agent based on the current state observation using a machine learning based model trained based on a plurality of registered training images comprises:
calculating action-values for each action in an action set including actions to adjust each of the transformation parameters individually by a respective step size, actions to adjust multiple transformation parameters simultaneously by their respective step sizes, and actions to optimize the respective step size for each of the transformation parameters.
17. The method of claim 1, wherein calculating action-values for a plurality of actions available to the artificial agent based on the current state observation using a machine learning based model trained based on a plurality of registered training images comprises:
calculating the action-values for the plurality of actions corresponding to the predetermined adjustments of the transformation parameters and an action-value for a stop registration action based on the current state observation using the trained machine learning based model.
18. The method of claim 17, wherein repeating the determining, calculating, and selecting steps for a plurality of iterations comprises:
repeating the determining, calculating, and selecting steps until the action-value calculated by the trained DNN for the stop registration action is higher than the action-values calculated by the trained DNN for each of the plurality of actions corresponding to the predetermined adjustments of the transformation parameters.
19. The method of claim 1, wherein repeating the determining, calculating, and selecting steps for a plurality of iterations comprises:
repeating the determining, calculating, and selecting steps using a first image resolution of the plurality of medical images and a first machine learning based model trained using training images at the first image resolution until a first stop condition is reached; and
repeating the determining, calculating, and selecting steps using a second image resolution of the plurality of medical images and a second machine learning based model trained using training images at the second image resolution until a second stop condition is reached.
20. An apparatus for intelligent artificial agent based registration of medical images, comprising:
a processor; and
a memory storing computer program instructions, which when executed by the processor cause the processor to perform operations comprising:
determining a current state observation of an artificial agent based on a plurality of medical images and current transformation parameters;
calculating action-values for a plurality of actions available to the artificial agent based on the current state observation using a machine learning based model trained based on a plurality of registered training images, wherein the plurality of actions correspond to predetermined adjustments of the transformation parameters;
selecting an action having a highest action-value from the plurality of actions and adjusting the transformation parameters by the predetermined adjustment corresponding to the selected action;
repeating the determining, calculating, and selecting operations for a plurality of iterations; and
registering the plurality of medical images using final transformation parameters resulting from the plurality of iterations.
21. The apparatus of claim 20, wherein the plurality of medical images includes a first medical image and a second medical image and determining a current state observation of an artificial agent based on a plurality of medical images and current transformation parameters comprises:
determining the current state observation by comparing the second medical image transformed by the current transformation parameters and the first medical image.
22. The apparatus of claim 20, wherein the transformation is a rigid-body transformation and the plurality of actions comprises actions corresponding to predetermined adjustments of translation and rotation rigid-body transformation parameters.
23. The apparatus of claim 20, wherein calculating action-values for a plurality of actions available to the artificial agent based on the current state observation using a machine learning based model trained based on a plurality of registered training images comprises:
calculating action-values for the plurality of actions available to the artificial agent based on the current state observation using a trained deep neural network (DNN).
24. The apparatus of claim 23, wherein the trained DNN is trained to predict action-values for the plurality of actions based on a plurality of training image pairs with known ground truth transformation parameters using reinforcement learning in which, for each training image pair, a reward for each action of the plurality of actions at a given state is used to train the DNN to learn an optimal registration policy.
25. The apparatus of claim 24, wherein the trained DNN is trained using supervised reinforcement learning, in which the reward for each action of the plurality of actions at a given state is directly calculated based on a decrease in a geodesic distance between the transformation parameters at the given a state and the ground truth transformation parameters for the training image pair resulting from applying the action.
26. The apparatus of claim 25, wherein the plurality of training image pairs includes training image pairs synthetically generated by artificially de-aligning other training image pairs using randomly generated perturbations to the ground truth transformation parameters of the other training image pairs.
27. The apparatus of claim 25, wherein the plurality of training image pairs includes training image pairs synthetically generated from other training image pairs to provide dense sampling close to the ground truth transformation parameters by co-deforming the other training image pairs by randomly generated affine transformations within a range of the ground truth transformation parameters.
28. The apparatus of claim 20, wherein the plurality of medical images includes a first medical image and a second medical image, the transformation is a dense deformation model that models deformations applied to the second medical image, the plurality of actions comprises actions corresponding to adjustments to parameters of the deformation model, and registering the plurality of medical images using final transformation parameters resulting from the plurality of iterations comprises:
generating a dense deformation field corresponding to final parameters of the deformation model resulting from the plurality of iterations; and
warping the second medical image to register the second medical image with the first medical image by applying the dense deformation field corresponding to the final parameters of the deformation model.
29. The apparatus of claim 20, wherein repeating the determining, calculating, and selecting operations for a plurality of iterations comprises:
repeating the determining, calculating, and selecting operations using a first image resolution of the plurality of medical images and a first machine learning based model trained using training images at the first image resolution until a first stop condition is reached; and
repeating the determining, calculating, and selecting operations using a second image resolution of the plurality of medical images and a second machine learning based model trained using training images at the second image resolution until a second stop condition is reached.
30. A non-transitory computer readable medium storing computer program instructions for intelligent artificial agent based registration of medical images, the computer program instructions defining operations comprising:
determining a current state observation of an artificial agent based on a plurality of medical images and current transformation parameters;
calculating action-values for a plurality of actions available to the artificial agent based on the current state observation using a machine learning based model trained based on a plurality of registered training images, wherein the plurality of actions correspond to predetermined adjustments of the transformation parameters;
selecting an action having a highest action-value from the plurality of actions and adjusting the transformation parameters by the predetermined adjustment corresponding to the selected action;
repeating the determining, calculating, and selecting operations for a plurality of iterations; and
registering the plurality of medical images using final transformation parameters resulting from the plurality of iterations.
31. The non-transitory computer readable medium of claim 30, wherein the plurality of medical images includes a first medical image and a second medical image and determining a current state observation of an artificial agent based on a plurality of medical images and current transformation parameters comprises:
determining the current state observation by comparing the second medical image transformed by the current transformation parameters and the first medical image.
32. The non-transitory computer readable medium of claim 30, wherein the transformation is a rigid-body transformation and the plurality of actions comprises actions corresponding to predetermined adjustments of translation and rotation rigid-body transformation parameters.
33. The non-transitory computer readable medium of claim 30, wherein calculating action-values for a plurality of actions available to the artificial agent based on the current state observation using a machine learning based model trained based on a plurality of registered training images comprises:
calculating action-values for the plurality of actions available to the artificial agent based on the current state observation using a trained deep neural network (DNN).
34. The non-transitory computer readable medium of claim 33, wherein the trained DNN is trained to predict action-values for the plurality of actions based on a plurality of training image pairs with known ground truth transformation parameters using reinforcement learning in which, for each training image pair, a reward for each action of the plurality of actions at a given state is used to train the DNN to learn an optimal registration policy.
35. The non-transitory computer readable medium of claim 34, wherein the trained DNN is trained using supervised reinforcement learning, in which the reward for each action of the plurality of actions at a given state is directly calculated based on a decrease in a geodesic distance between the transformation parameters at the given a state and the ground truth transformation parameters for the training image pair resulting from applying the action.
36. The non-transitory computer readable medium of claim 35, wherein the plurality of training image pairs includes training image pairs synthetically generated by artificially de-aligning other training image pairs using randomly generated perturbations to the ground truth transformation parameters of the other training image pairs.
37. The non-transitory computer readable medium of claim 35, wherein the plurality of training image pairs includes training image pairs synthetically generated from other training image pairs to provide dense sampling close to the ground truth transformation parameters by co-deforming the other training image pairs by randomly generated affine transformations within a range of the ground truth transformation parameters.
38. The non-transitory computer readable medium of claim 30, wherein the plurality of medical images includes a first medical image and a second medical image, the transformation is a dense deformation model that models deformations applied to the second medical image, the plurality of actions comprises actions corresponding to adjustments to parameters of the deformation model, and registering the plurality of medical images using final transformation parameters resulting from the plurality of iterations comprises:
generating a dense deformation field corresponding to final parameters of the deformation model resulting from the plurality of iterations; and
warping the second medical image to register the second medical image with the first medical image by applying the dense deformation field corresponding to the final parameters of the deformation model.
39. The non-transitory computer readable medium of claim 30, wherein repeating the determining, calculating, and selecting operations for a plurality of iterations comprises:
repeating the determining, calculating, and selecting operations using a first image resolution of the plurality of medical images and a first machine learning based model trained using training images at the first image resolution until a first stop condition is reached; and
repeating the determining, calculating, and selecting operations using a second image resolution of the plurality of medical images and a second machine learning based model trained using training images at the second image resolution until a second stop condition is reached.
US15/587,094 2016-05-18 2017-05-04 Method and System for Image Registration Using an Intelligent Artificial Agent Abandoned US20170337682A1 (en)

Priority Applications (6)

Application Number Priority Date Filing Date Title
US15/587,094 US20170337682A1 (en) 2016-05-18 2017-05-04 Method and System for Image Registration Using an Intelligent Artificial Agent
EP17171689.7A EP3246875A3 (en) 2016-05-18 2017-05-18 Method and system for image registration using an intelligent artificial agent
CN201710351955.5A CN107403446B (en) 2016-05-18 2017-05-18 Method and system for image registration using intelligent artificial agents
US16/861,353 US11557036B2 (en) 2016-05-18 2020-04-29 Method and system for image registration using an intelligent artificial agent
US18/064,366 US11741605B2 (en) 2016-05-18 2022-12-12 Method and system for image registration using an intelligent artificial agent
US18/351,548 US20230368383A1 (en) 2016-05-18 2023-07-13 Method and System for Image Registration Using an Intelligent Artificial Agent

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US201662338059P 2016-05-18 2016-05-18
US201662344125P 2016-06-01 2016-06-01
US201662401977P 2016-09-30 2016-09-30
US15/587,094 US20170337682A1 (en) 2016-05-18 2017-05-04 Method and System for Image Registration Using an Intelligent Artificial Agent

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US16/861,353 Continuation US11557036B2 (en) 2016-05-18 2020-04-29 Method and system for image registration using an intelligent artificial agent

Related Child Applications (2)

Application Number Title Priority Date Filing Date
US16/861,353 Continuation US11557036B2 (en) 2016-05-18 2020-04-29 Method and system for image registration using an intelligent artificial agent
US18/064,366 Continuation US11741605B2 (en) 2016-05-18 2022-12-12 Method and system for image registration using an intelligent artificial agent

Publications (1)

Publication Number Publication Date
US20170337682A1 true US20170337682A1 (en) 2017-11-23

Family

ID=59061771

Family Applications (4)

Application Number Title Priority Date Filing Date
US15/587,094 Abandoned US20170337682A1 (en) 2016-05-18 2017-05-04 Method and System for Image Registration Using an Intelligent Artificial Agent
US16/861,353 Active 2038-03-31 US11557036B2 (en) 2016-05-18 2020-04-29 Method and system for image registration using an intelligent artificial agent
US18/064,366 Active US11741605B2 (en) 2016-05-18 2022-12-12 Method and system for image registration using an intelligent artificial agent
US18/351,548 Pending US20230368383A1 (en) 2016-05-18 2023-07-13 Method and System for Image Registration Using an Intelligent Artificial Agent

Family Applications After (3)

Application Number Title Priority Date Filing Date
US16/861,353 Active 2038-03-31 US11557036B2 (en) 2016-05-18 2020-04-29 Method and system for image registration using an intelligent artificial agent
US18/064,366 Active US11741605B2 (en) 2016-05-18 2022-12-12 Method and system for image registration using an intelligent artificial agent
US18/351,548 Pending US20230368383A1 (en) 2016-05-18 2023-07-13 Method and System for Image Registration Using an Intelligent Artificial Agent

Country Status (3)

Country Link
US (4) US20170337682A1 (en)
EP (1) EP3246875A3 (en)
CN (1) CN107403446B (en)

Cited By (185)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170371017A1 (en) * 2016-06-23 2017-12-28 Siemens Healthcare Gmbh System and Method For Normalized Reference Database For MR Images Via Autoencoders
US20170371982A1 (en) * 2016-06-27 2017-12-28 Faro Technologies, Inc. Global optimization of networks of locally fitted objects
US10049323B1 (en) * 2017-10-13 2018-08-14 StradVision, Inc. Method and device for performing activation and convolution operation at the same time and learning method and learning device for the same
CN108492309A (en) * 2018-01-21 2018-09-04 西安电子科技大学 Magnetic resonance image medium sized vein blood vessel segmentation method based on migration convolutional neural networks
US10074038B2 (en) * 2016-11-23 2018-09-11 General Electric Company Deep learning medical systems and methods for image reconstruction and quality evaluation
US20180330511A1 (en) * 2017-05-11 2018-11-15 Kla-Tencor Corporation Learning based approach for aligning images acquired with different modalities
CN109035137A (en) * 2018-07-27 2018-12-18 重庆邮电大学 A kind of multimode medical image fusion method based on optimal transmission theory
CN109035316A (en) * 2018-08-28 2018-12-18 北京安德医智科技有限公司 The method for registering and equipment of nuclear magnetic resonance image sequence
US20180374194A1 (en) * 2017-06-21 2018-12-27 International Business Machines Corporation Automatic Image Registration Quality Assessment
US20180374193A1 (en) * 2017-06-21 2018-12-27 International Business Machines Corporation Automatic Image Registration Quality Assessment
US20180374216A1 (en) * 2017-05-02 2018-12-27 Aivitae LLC System and method for facilitating autonomous control of an imaging system
US20190043185A1 (en) * 2017-08-04 2019-02-07 Omron Corporation Image processing system
US20190057505A1 (en) * 2017-08-17 2019-02-21 Siemens Healthcare Gmbh Automatic change detection in medical images
US20190080205A1 (en) * 2016-09-16 2019-03-14 Deep Learning Analytics, LLC Systems and Methods for Deep Model Translation Generation
US20190083067A1 (en) * 2017-09-21 2019-03-21 General Electric Company Methods and systems for correction of one dimensional shear wave data
US10242264B1 (en) * 2018-04-25 2019-03-26 Imagry (Israel) Ltd. System and method for training a machine-learning model to identify real-world elements
US20190102878A1 (en) * 2017-09-30 2019-04-04 Baidu Online Network Technology (Beijing) Co., Ltd. Method and apparatus for analyzing medical image
CN109599109A (en) * 2018-12-26 2019-04-09 浙江大学 For the confrontation audio generation method and system of whitepack scene
CN109712175A (en) * 2018-12-19 2019-05-03 浙江大学常州工业技术研究院 The method for registering of CT picture
US20190138889A1 (en) * 2017-11-06 2019-05-09 Nvidia Corporation Multi-frame video interpolation using optical flow
CN109767459A (en) * 2019-01-17 2019-05-17 中南大学 Novel ocular base map method for registering
US10296828B2 (en) * 2017-04-05 2019-05-21 Here Global B.V. Learning a similarity measure for vision-based localization on a high definition (HD) map
EP3502978A1 (en) 2017-12-22 2019-06-26 Siemens Healthcare GmbH Meta-learning system
US20190206218A1 (en) * 2017-12-28 2019-07-04 Cerner Innovation, Inc. Utilizing artificial intelligence to detect objects or patient safety events in a patient room
US20190216409A1 (en) * 2018-01-15 2019-07-18 Siemens Healthcare Gmbh Method and System for 3D Reconstruction of X-ray CT Volume and Segmentation Mask from a Few X-ray Radiographs
US10366494B2 (en) * 2017-04-21 2019-07-30 Samsung Medison Co., Ltd. Image segmentation method and electronic device therefor
US20190236360A1 (en) * 2018-01-30 2019-08-01 Mashgin Inc. Feedback loop for image-based recognition
CN110074813A (en) * 2019-04-26 2019-08-02 深圳大学 A kind of ultrasonic image reconstruction method and system
US10388016B2 (en) 2016-12-30 2019-08-20 Cerner Innovation, Inc. Seizure detection
US10402969B2 (en) * 2017-03-10 2019-09-03 General Electric Company Methods and systems for model driven multi-modal medical imaging
WO2019166447A1 (en) * 2018-03-02 2019-09-06 Koninklijke Philips N.V. Image registration qualification
CN110211165A (en) * 2019-06-10 2019-09-06 成都信息工程大学 A kind of image multi-mode registration method based on the study of asynchronous deeply
US10410042B2 (en) 2015-12-31 2019-09-10 Cerner Innovation, Inc. Detecting unauthorized visitors
US20190279361A1 (en) * 2018-03-07 2019-09-12 University Of Virginia Patent Foundation Automatic quantification of cardiac mri for hypertrophic cardiomyopathy
US10475538B2 (en) * 2018-01-11 2019-11-12 Siemens Healthcare Gmbh System for predicting computed tomography data from surface data
US10482321B2 (en) 2017-12-29 2019-11-19 Cerner Innovation, Inc. Methods and systems for identifying the crossing of a virtual barrier
WO2019222289A1 (en) * 2018-05-14 2019-11-21 Tempus Labs, Inc. A generalizable and interpretable deep learning framework for predicting msi from histopathology slide images
US20190353741A1 (en) * 2018-05-16 2019-11-21 Siemens Healthcare Gmbh Deep Learning Reconstruction of Free Breathing Perfusion
US10489708B2 (en) * 2016-05-20 2019-11-26 Magic Leap, Inc. Method and system for performing convolutional image transformation estimation
US10489936B1 (en) * 2019-04-29 2019-11-26 Deep Render Ltd. System and method for lossy image and video compression utilizing a metanetwork
US20190377943A1 (en) * 2018-06-12 2019-12-12 Wipro Limited Method and system for correcting fabrication in a document
CN110580695A (en) * 2019-08-07 2019-12-17 深圳先进技术研究院 multi-mode three-dimensional medical image fusion method and system and electronic equipment
US10510443B2 (en) 2014-12-23 2019-12-17 Cerner Innovation, Inc. Methods and systems for determining whether a monitored individual's hand(s) have entered a virtual safety zone
CN110598765A (en) * 2019-08-28 2019-12-20 腾讯科技(深圳)有限公司 Sample generation method and device, computer equipment and storage medium
CN110634117A (en) * 2018-05-31 2019-12-31 中山大学肿瘤防治中心(中山大学附属肿瘤医院、中山大学肿瘤研究所) Method for quantitatively evaluating image by using three-dimensional parameters
US10524722B2 (en) 2014-12-26 2020-01-07 Cerner Innovation, Inc. Method and system for determining whether a caregiver takes appropriate measures to prevent patient bedsores
US10540798B1 (en) 2019-01-10 2020-01-21 Capital One Services, Llc Methods and arrangements to create images
US10540578B2 (en) 2017-12-21 2020-01-21 International Business Machines Corporation Adapting a generative adversarial network to new data sources for image classification
US10546481B2 (en) 2011-07-12 2020-01-28 Cerner Innovation, Inc. Method for determining whether an individual leaves a prescribed virtual perimeter
US10552714B2 (en) * 2018-03-16 2020-02-04 Ebay Inc. Generating a digital image using a generative adversarial network
US20200043179A1 (en) * 2018-08-03 2020-02-06 Logitech Europe S.A. Method and system for detecting peripheral device displacement
US20200038109A1 (en) * 2017-01-12 2020-02-06 Mazor Robotics Ltd. Image based pathology prediction using artificial intelligence
US10565476B1 (en) * 2018-09-04 2020-02-18 StradVision, Inc. Method and computing device for generating image data set for learning to be used for detection of obstruction in autonomous driving circumstances and learning method and learning device using the same
CN110827331A (en) * 2019-11-04 2020-02-21 上海联影智能医疗科技有限公司 Training method of image registration model, image registration method and computer equipment
US20200065673A1 (en) * 2017-05-10 2020-02-27 Telefonaktiebolaget Lm Ericsson (Publ) Pre-training system for self-learning agent in virtualized environment
US10592779B2 (en) * 2017-12-21 2020-03-17 International Business Machines Corporation Generative adversarial network medical image generation for training of a classifier
US10593051B2 (en) 2017-12-20 2020-03-17 International Business Machines Corporation Medical image registration guided by target lesion
US10629046B2 (en) 2015-06-01 2020-04-21 Cerner Innovation, Inc. Systems and methods for determining whether an individual enters a prescribed virtual zone using skeletal tracking and 3D blob detection
WO2020078725A1 (en) * 2018-10-19 2020-04-23 Mirada Medical Limited System and method for automatic delineation of scanned images
CN111144486A (en) * 2019-12-27 2020-05-12 电子科技大学 Heart nuclear magnetic resonance image key point detection method based on convolutional neural network
CN111161227A (en) * 2019-12-20 2020-05-15 成都数之联科技有限公司 Target positioning method and system based on deep neural network
US10655975B2 (en) * 2019-09-23 2020-05-19 Alibaba Group Holding Limited System and method for routing optimization
CN111199227A (en) * 2019-12-20 2020-05-26 广西柳州联耕科技有限公司 High-precision image identification method
WO2020113148A1 (en) * 2018-11-29 2020-06-04 The Board Of Trustees Of The Leland Sanford Junior University Single or a few views computed tomography imaging with deep neural network
WO2020118425A1 (en) * 2018-12-10 2020-06-18 Her Majesty The Queen In Right Of Canada, As Represented By The Minister Of National Defence Automated diagnostics in 3d ultrasound system and method
CN111311652A (en) * 2020-01-19 2020-06-19 杭州电子科技大学 Regression correction network-based sonar image registration method and system
US20200202515A1 (en) * 2018-12-21 2020-06-25 General Electric Company Systems and methods for deep learning based automated spine registration and label propagation
US20200202622A1 (en) * 2018-12-19 2020-06-25 Nvidia Corporation Mesh reconstruction using data-driven priors
CN111340859A (en) * 2018-12-19 2020-06-26 西门子医疗有限公司 Method for image registration, learning device and medical imaging device
CN111462146A (en) * 2020-04-16 2020-07-28 成都信息工程大学 Medical image multi-mode registration method based on space-time intelligent agent
US10733746B2 (en) * 2016-11-02 2020-08-04 Shanghai United Imaging Healthcare Co., Ltd. System and method for registering multi-modality images
US10748296B2 (en) * 2018-01-18 2020-08-18 Elekta, Inc. Methods and devices for surface motion tracking
US20200272853A1 (en) * 2019-02-25 2020-08-27 Fair Isaac Corporation Density based confidence measures of neural networks for reliable predictions
US10769791B2 (en) * 2017-10-13 2020-09-08 Beijing Keya Medical Technology Co., Ltd. Systems and methods for cross-modality image segmentation
US20200294217A1 (en) * 2017-11-09 2020-09-17 Samsung Electronics Co., Ltd. Method and apparatus for video super resolution using convolutional neural network with two-stage motion compensation
US10783634B2 (en) * 2017-11-22 2020-09-22 General Electric Company Systems and methods to deliver point of care alerts for radiological findings
WO2020198456A1 (en) * 2019-03-27 2020-10-01 The General Hospital Corporation Retrospective motion correction using a combined neural network and model-based image reconstruction of magnetic resonance data
CN111739614A (en) * 2019-03-25 2020-10-02 西门子医疗有限公司 Medical image enhancement
WO2020198854A1 (en) * 2019-03-29 2020-10-08 Polyvalor, Limited Partnership Method and system for producing medical images
US10799315B2 (en) * 2016-07-26 2020-10-13 7D Surgical Inc. Systems and methods for verification of fiducial correspondence during image-guided surgical procedures
US10818019B2 (en) * 2017-08-14 2020-10-27 Siemens Healthcare Gmbh Dilated fully convolutional network for multi-agent 2D/3D medical image registration
CN111950194A (en) * 2020-07-22 2020-11-17 西南大学 Newton momentum-based distributed acceleration composite optimization method and system
CN111950584A (en) * 2020-06-16 2020-11-17 江西中科九峰智慧医疗科技有限公司 Intelligent identification method and system for integrity of part in X-ray chest radiography
US10842445B2 (en) * 2018-11-08 2020-11-24 General Electric Company System and method for unsupervised deep learning for deformable image registration
US10861184B1 (en) * 2017-01-19 2020-12-08 X Development Llc Object pose neural network system
US10867385B2 (en) * 2016-01-21 2020-12-15 Elekta, Inc. Systems and methods for segmentation of intra-patient medical images
US10863946B2 (en) * 2016-11-03 2020-12-15 Siemens Medical Solutions Usa, Inc. Respiratory motion estimation in projection domain in nuclear medical imaging
US10874794B2 (en) 2011-06-20 2020-12-29 Cerner Innovation, Inc. Managing medication administration in clinical care room
CN112313668A (en) * 2018-06-15 2021-02-02 深透医疗公司 System and method for magnetic resonance imaging normalization using deep learning
US20210034974A1 (en) * 2019-07-31 2021-02-04 Royal Bank Of Canada Devices and methods for reinforcement learning visualization using immersive environments
CN112365422A (en) * 2020-11-17 2021-02-12 重庆邮电大学 Irregular missing image restoration method and system based on deep aggregation network
US10922936B2 (en) 2018-11-06 2021-02-16 Cerner Innovation, Inc. Methods and systems for detecting prohibited objects
US10929989B2 (en) 2017-11-01 2021-02-23 Siemens Healthcare Gmbh Determining a transformation between coordinate frames of sets of image data
US20210056717A1 (en) * 2018-05-01 2021-02-25 Koninklijke Philips N.V. Lower to higher resolution image fusion
US10937540B2 (en) 2017-12-21 2021-03-02 International Business Machines Coporation Medical image classification based on a generative adversarial network trained discriminator
US10937141B2 (en) * 2017-10-13 2021-03-02 Sualab Co., Ltd. Deep learning based image comparison device, method and computer program stored in computer readable medium
US10943353B1 (en) * 2019-09-11 2021-03-09 International Business Machines Corporation Handling untrainable conditions in a network architecture search
CN112494028A (en) * 2019-09-13 2021-03-16 通用电气公司 Biopsy workflow using multi-modality imaging
US10956785B2 (en) 2018-04-27 2021-03-23 Arizona Board Of Regents On Behalf Of Arizona State University Methods, systems, and media for selecting candidates for annotation for use in training classifiers
US10957041B2 (en) 2018-05-14 2021-03-23 Tempus Labs, Inc. Determining biomarkers from histopathology slide images
US10957031B1 (en) * 2019-09-06 2021-03-23 Accenture Global Solutions Limited Intelligent defect detection from image data
US10959780B2 (en) * 2015-06-26 2021-03-30 Therenva Method and system for helping to guide an endovascular tool in vascular structures
US10984905B2 (en) * 2017-11-03 2021-04-20 Siemens Healthcare Gmbh Artificial intelligence for physiological quantification in medical imaging
WO2021077009A1 (en) * 2019-10-16 2021-04-22 REHABILITATION INSTITUTE OF CHICAGO d/b/a Shirley Ryan AbilityLab Systems and methods for electromyogram-based control using neural networks
US10991097B2 (en) 2018-12-31 2021-04-27 Tempus Labs, Inc. Artificial intelligence segmentation of tissue images
US10996306B2 (en) * 2019-04-25 2021-05-04 General Electric Company MRI system and method using neural network for detection of patient motion
US10997462B2 (en) * 2018-04-04 2021-05-04 Canon Virginia, Inc. Devices, systems, and methods for clustering reference images for non-destructive testing
US20210133984A1 (en) * 2019-11-04 2021-05-06 Shanghai United Imaging Intelligence Co., Ltd. Systems and methods for machine learning based physiological motion measurement
US11003991B2 (en) * 2016-09-30 2021-05-11 Idemia Identity & Security France Methods for secure learning of parameters of a convolution neural network, and for secure input data classification
US11003982B2 (en) * 2017-06-27 2021-05-11 D5Ai Llc Aligned training of deep networks
CN112802046A (en) * 2021-01-28 2021-05-14 华南理工大学 Image generation system for generating pseudo CT from multi-sequence MR based on deep learning
US11023783B2 (en) 2019-09-11 2021-06-01 International Business Machines Corporation Network architecture search with global optimization
US20210170579A1 (en) * 2019-12-09 2021-06-10 Canon Kabushiki Kaisha Robot controller that controls robot, learned model, method of controlling robot, and storage medium
CN112990359A (en) * 2021-04-19 2021-06-18 深圳市深光粟科技有限公司 Image data processing method and device, computer and storage medium
CN113052879A (en) * 2021-04-08 2021-06-29 西安应用光学研究所 Multispectral image automatic registration method
US11054271B2 (en) 2018-12-04 2021-07-06 Here Global B.V. Method and apparatus for aligning drive segments based on user inputs
CN113096169A (en) * 2021-03-31 2021-07-09 华中科技大学 Non-rigid multimode medical image registration model establishing method and application thereof
US11067786B2 (en) * 2019-06-07 2021-07-20 Leica Microsystems Inc. Artifact regulation methods in deep model training for image transformation
CN113240699A (en) * 2021-05-20 2021-08-10 推想医疗科技股份有限公司 Image processing method and device, model training method and device, and electronic equipment
US11087185B2 (en) 2018-05-25 2021-08-10 Samsung Electronics Co., Ltd. Image processing method and apparatus with neural network adjustment
US20210274208A1 (en) * 2018-11-15 2021-09-02 Beijing Bytedance Network Technology Co., Ltd. Merge with mvd for affine
US11109945B2 (en) * 2017-06-09 2021-09-07 Dental Monitoring Method of evaluating an orthodontic aligner
US11127153B2 (en) * 2017-10-10 2021-09-21 Hitachi, Ltd. Radiation imaging device, image processing method, and image processing program
US11132792B2 (en) * 2018-02-22 2021-09-28 Siemens Healthcare Gmbh Cross domain medical image segmentation
US20210304457A1 (en) * 2020-03-31 2021-09-30 The Regents Of The University Of California Using neural networks to estimate motion vectors for motion corrected pet image reconstruction
WO2021202203A1 (en) * 2020-03-31 2021-10-07 Alibaba Group Holding Limited Medical image processing method and system and data processing method
US11151726B2 (en) * 2018-01-10 2021-10-19 Canon Medical Systems Corporation Medical image processing apparatus, X-ray diagnostic apparatus, and medical image processing method
US20210327063A1 (en) * 2018-12-21 2021-10-21 GE Precision Healthcare LLC Systems and methods for whole-body spine labeling
CN113538529A (en) * 2020-04-15 2021-10-22 杭州海康威视数字技术股份有限公司 Image registration apparatus
US11158069B2 (en) * 2018-12-11 2021-10-26 Siemens Healthcare Gmbh Unsupervised deformable registration for multi-modal images
US11164021B2 (en) * 2019-05-16 2021-11-02 Arizona Board Of Regents On Behalf Of Arizona State University Methods, systems, and media for discriminating and generating translated images
US20210366135A1 (en) * 2020-05-20 2021-11-25 Therenva Method and system for registering images containing anatomical structures
US11200416B2 (en) * 2017-06-14 2021-12-14 Beijing Sensetime Technology Development Co., Ltd Methods and apparatuses for image detection, electronic devices and storage media
US11200659B2 (en) 2019-11-18 2021-12-14 Stmicroelectronics (Rousset) Sas Neural network training device, system and method
US11200456B2 (en) * 2019-07-31 2021-12-14 GE Precision Healthcare LLC Systems and methods for generating augmented training data for machine learning models
US20210397886A1 (en) * 2020-06-22 2021-12-23 Shanghai United Imaging Intelligence Co., Ltd. Anatomy-aware motion estimation
WO2022005337A1 (en) * 2020-06-29 2022-01-06 Autonomous Non-Profit Organization For Higher Education «Skolkovo Institute Of Science And Technology» Veins mask projection alignment
US20220013231A1 (en) * 2020-07-13 2022-01-13 Corsmed Ab Method for ai applications in mri simulation
JP2022505498A (en) * 2018-12-19 2022-01-14 シャンハイ センスタイム インテリジェント テクノロジー カンパニー リミテッド Image processing methods, devices, electronic devices and computer readable storage media
US11250580B2 (en) * 2019-09-24 2022-02-15 Dentsply Sirona Inc. Method, system and computer readable storage media for registering intraoral measurements
US11263744B2 (en) * 2019-12-09 2022-03-01 Siemens Healthcare Gmbh Saliency mapping by feature reduction and perturbation modeling in medical imaging
US11270110B2 (en) 2019-09-17 2022-03-08 Boston Polarimetrics, Inc. Systems and methods for surface modeling using polarization cues
US20220092789A1 (en) * 2020-04-09 2022-03-24 Zhejiang Lab Automatic pancreas ct segmentation method based on a saliency-aware densely connected dilated convolutional neural network
CN114332509A (en) * 2021-12-29 2022-04-12 阿波罗智能技术(北京)有限公司 Image processing method, model training method, electronic device and automatic driving vehicle
US11302012B2 (en) * 2019-11-30 2022-04-12 Boston Polarimetrics, Inc. Systems and methods for transparent object segmentation using polarization cues
US11348239B2 (en) 2018-05-14 2022-05-31 Tempus Labs, Inc. Predicting total nucleic acid yield and dissection boundaries for histology slides
US11348661B2 (en) 2018-05-14 2022-05-31 Tempus Labs, Inc. Predicting total nucleic acid yield and dissection boundaries for histology slides
US11348240B2 (en) 2018-05-14 2022-05-31 Tempus Labs, Inc. Predicting total nucleic acid yield and dissection boundaries for histology slides
US11361442B2 (en) * 2017-11-16 2022-06-14 Oslo Universitetssykehus Histological image analysis
CN114663840A (en) * 2022-05-13 2022-06-24 浙江有色地球物理技术应用研究院有限公司 Tunnel environment inspection equipment monitoring method and system
US11380119B2 (en) * 2013-11-15 2022-07-05 Meta Platforms, Inc. Pose-aligned networks for deep attribute modeling
US20220225955A1 (en) * 2019-04-26 2022-07-21 Spectronic Ab Generating synthetic electron density images from magnetic resonance images
US11405637B2 (en) * 2019-10-29 2022-08-02 Samsung Electronics Co., Ltd. Image encoding method and apparatus and image decoding method and apparatus
US11436780B2 (en) * 2018-05-24 2022-09-06 Warner Bros. Entertainment Inc. Matching mouth shape and movement in digital video to alternative audio
WO2022192015A1 (en) * 2021-03-10 2022-09-15 Intrinsic Innovation Llc Systems and methods for high dynamic range image reconstruction
US11449759B2 (en) * 2018-01-03 2022-09-20 Siemens Heathcare Gmbh Medical imaging diffeomorphic registration based on machine learning
US20220318516A1 (en) * 2017-06-05 2022-10-06 Deepmind Technologies Limited Action selection based on environment observations and textual instructions
US11501424B2 (en) 2019-11-18 2022-11-15 Stmicroelectronics (Rousset) Sas Neural network training device, system and method
US11501157B2 (en) 2018-07-30 2022-11-15 International Business Machines Corporation Action shaping from demonstration for fast reinforcement learning
CN115393402A (en) * 2022-08-24 2022-11-25 北京医智影科技有限公司 Training method of image registration network model, image registration method and equipment
US11515030B2 (en) 2016-06-23 2022-11-29 Siemens Healthcare Gmbh System and method for artificial agent based cognitive operating rooms
WO2022247296A1 (en) * 2021-05-27 2022-12-01 广州柏视医疗科技有限公司 Mark point-based image registration method
WO2022247218A1 (en) * 2021-05-27 2022-12-01 广州柏视医疗科技有限公司 Image registration method based on automatic delineation
US11517197B2 (en) 2017-10-06 2022-12-06 Canon Medical Systems Corporation Apparatus and method for medical image reconstruction using deep learning for computed tomography (CT) image noise and artifacts reduction
US11521064B2 (en) * 2017-12-01 2022-12-06 Koninklijke Philips N.V. Training a neural network model
US11525906B2 (en) 2019-10-07 2022-12-13 Intrinsic Innovation Llc Systems and methods for augmentation of sensor systems and imaging systems with polarization
US11537872B2 (en) 2018-07-30 2022-12-27 International Business Machines Corporation Imitation learning by action shaping with antagonist reinforcement learning
US11551037B2 (en) 2018-04-23 2023-01-10 Siemens Aktiengesellschaft Method and apparatus for determining a physical shape, method for manufacturing a calculation device, calculation device, and use of the calculation device
US11568236B2 (en) 2018-01-25 2023-01-31 The Research Foundation For The State University Of New York Framework and methods of diverse exploration for fast and safe policy improvement
US11571809B1 (en) * 2019-09-15 2023-02-07 X Development Llc Robotic control using value distributions
CN115762722A (en) * 2022-11-22 2023-03-07 南方医科大学珠江医院 Image review system based on artificial intelligence
US20230076196A1 (en) * 2021-09-03 2023-03-09 Adobe Inc. Asynchronous multi-user real-time streaming of web-based image edits using generative adversarial network(s)
US11604941B1 (en) * 2017-10-27 2023-03-14 Deepmind Technologies Limited Training action-selection neural networks from demonstrations using multiple losses
US20230169659A1 (en) * 2021-11-30 2023-06-01 Shanghai United Imaging Intelligence Co., Ltd. Image segmentation and tracking based on statistical shape model
WO2023049211A3 (en) * 2021-09-21 2023-06-01 Hyperfine Operations, Inc. Contrastive multimodality image registration
US11710241B2 (en) 2018-02-14 2023-07-25 Elekta, Inc. Atlas-based segmentation using deep-learning
US11712208B2 (en) 2017-11-22 2023-08-01 General Electric Company Systems and methods to deliver point of care alerts for radiological findings
US11734575B2 (en) 2018-07-30 2023-08-22 International Business Machines Corporation Sequential learning of constraints for hierarchical reinforcement learning
US11756160B2 (en) * 2018-07-27 2023-09-12 Washington University ML-based methods for pseudo-CT and HR MR image estimation
US11756668B2 (en) * 2017-11-12 2023-09-12 Aleph Bot Ltd Systems methods devices circuits and computer executable code for tracking evaluating and facilitating a medical procedure
WO2023172400A1 (en) * 2022-03-10 2023-09-14 Varian Medical Systems, Inc. Artificial intelligence boosting dose calculations
US11847761B2 (en) * 2017-10-06 2023-12-19 Canon Medical Systems Corporation Medical image processing apparatus having a plurality of neural networks corresponding to different fields of view
US11853901B2 (en) 2019-07-26 2023-12-26 Samsung Electronics Co., Ltd. Learning method of AI model and electronic apparatus
US11875441B2 (en) * 2019-12-03 2024-01-16 Disney Enterprises, Inc. Data-driven extraction and composition of secondary dynamics in facial performance capture
US11941802B2 (en) 2020-03-31 2024-03-26 Alibaba Group Holding Limited Data processing method, means and system

Families Citing this family (75)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP4241694A3 (en) 2016-05-16 2023-12-20 Cathworks Ltd. Selection of vascular paths from images
US20170337682A1 (en) 2016-05-18 2017-11-23 Siemens Healthcare Gmbh Method and System for Image Registration Using an Intelligent Artificial Agent
CN116741353A (en) * 2017-11-24 2023-09-12 佳能医疗系统株式会社 Medical data processing device, magnetic resonance imaging device, and learning model generation method
GB201720059D0 (en) * 2017-12-01 2018-01-17 Ucb Biopharma Sprl Three-dimensional medical image analysis method and system for identification of vertebral fractures
CN111971751A (en) * 2018-02-08 2020-11-20 通用电气公司 System and method for evaluating dynamic data
CN108563692B (en) * 2018-03-16 2021-08-06 新智认知数据服务有限公司 Target duplicate checking method and device
US10902621B2 (en) 2018-03-25 2021-01-26 Varian Medical Systems International Ag Deformable image registration based on masked computed tomography (CT) image
US10540749B2 (en) 2018-03-29 2020-01-21 Mitsubishi Electric Research Laboratories, Inc. System and method for learning-based image super-resolution
CN111971688A (en) * 2018-04-09 2020-11-20 皇家飞利浦有限公司 Ultrasound system with artificial neural network for retrieving imaging parameter settings of relapsing patients
CN108615073B (en) * 2018-04-28 2020-11-03 京东数字科技控股有限公司 Image processing method and device, computer readable storage medium and electronic device
US11557390B2 (en) * 2018-04-30 2023-01-17 Elekta, Inc. Radiotherapy treatment plan modeling using generative adversarial networks
US11903771B2 (en) * 2018-05-16 2024-02-20 Koninklijke Philips N.V. Automated tumor identification during surgery using machine-learning
US10846875B2 (en) * 2018-06-07 2020-11-24 Siemens Healthcare Gmbh Adaptive nonlinear optimization of shape parameters for object localization in 3D medical images
CN108968916B (en) * 2018-06-19 2021-04-16 艾瑞迈迪医疗科技(北京)有限公司 Respiratory motion correction method, apparatus, computer device, and storage medium
US20210216878A1 (en) * 2018-08-24 2021-07-15 Arterys Inc. Deep learning-based coregistration
CN109350100A (en) * 2018-09-27 2019-02-19 上海联影医疗科技有限公司 Medical imaging procedure, medical imaging devices and computer readable storage medium
CN110969248B (en) * 2018-09-28 2023-04-07 中国石油化工股份有限公司 RBF-A2C-based reinforcement learning method and system
CN110969664B (en) * 2018-09-30 2023-10-24 北京魔门塔科技有限公司 Dynamic calibration method for external parameters of camera
US10799183B2 (en) 2018-11-07 2020-10-13 General Electric Company Methods and systems for whole body imaging
CN109658349A (en) * 2018-11-16 2019-04-19 聚时科技(上海)有限公司 A kind of image enchancing method and its application for supervised learning application
CN111325675B (en) * 2018-12-17 2023-12-26 中国科学院深圳先进技术研究院 Image processing method, device, equipment and storage medium
US11354791B2 (en) * 2018-12-19 2022-06-07 General Electric Company Methods and system for transforming medical images into different styled images with deep neural networks
CN109754396B (en) * 2018-12-29 2021-02-19 上海联影智能医疗科技有限公司 Image registration method and device, computer equipment and storage medium
CN109767460A (en) * 2018-12-27 2019-05-17 上海商汤智能科技有限公司 Image processing method, device, electronic equipment and computer readable storage medium
JP7073286B2 (en) * 2019-01-10 2022-05-23 株式会社日立製作所 Data generator, predictor learning device, data generation method, and learning method
EP3709263A1 (en) * 2019-03-14 2020-09-16 Siemens Healthcare GmbH Method and system for monitoring a biological process
CN110009669B (en) * 2019-03-22 2021-12-10 电子科技大学 3D/2D medical image registration method based on deep reinforcement learning
CN110163897B (en) * 2019-04-24 2021-06-29 艾瑞迈迪科技石家庄有限公司 Multi-modal image registration method based on synthetic ultrasound image
US11195277B2 (en) * 2019-04-25 2021-12-07 GE Precision Healthcare LLC Systems and methods for generating normative imaging data for medical image processing using deep learning
US11062183B2 (en) * 2019-05-21 2021-07-13 Wipro Limited System and method for automated 3D training content generation
TWI738001B (en) 2019-06-03 2021-09-01 睿傳數據股份有限公司 Method and device for identifying body part in medical image
CN110263845B (en) * 2019-06-18 2023-05-02 西安电子科技大学 SAR image change detection method based on semi-supervised countermeasure depth network
CN110288581B (en) * 2019-06-26 2022-11-04 电子科技大学 Segmentation method based on model for keeping shape convexity level set
US11410308B2 (en) * 2019-07-17 2022-08-09 Siemens Healthcare Gmbh 3D vessel centerline reconstruction from 2D medical images
CN110517225B (en) * 2019-07-19 2023-07-11 平安科技(深圳)有限公司 AI image recognition method, apparatus, device and storage medium
AU2020344704A1 (en) 2019-09-13 2022-04-14 Treace Medical Concepts, Inc. Patient-specific surgical methods and instrumentation
US20210089924A1 (en) * 2019-09-24 2021-03-25 Nec Laboratories America, Inc Learning weighted-average neighbor embeddings
CN110648276B (en) * 2019-09-25 2023-03-31 重庆大学 High-dimensional image data dimension reduction method based on manifold mapping and dictionary learning
CN112734814B (en) * 2019-10-28 2023-10-20 北京大学 Three-dimensional craniofacial cone beam CT image registration method
CN110838139B (en) * 2019-11-04 2022-10-25 上海联影智能医疗科技有限公司 Training method of image registration model, image registration method and computer equipment
US11809977B2 (en) 2019-11-14 2023-11-07 NEC Laboratories Europe GmbH Weakly supervised reinforcement learning
US20220270256A1 (en) * 2019-12-13 2022-08-25 Siemens Healthcare Gmbh Compensation of organ deformation for medical image registration
CN111260705B (en) * 2020-01-13 2022-03-15 武汉大学 Prostate MR image multi-task registration method based on deep convolutional neural network
EP3885990A1 (en) * 2020-03-26 2021-09-29 Siemens Healthcare GmbH Landmark detection
US11922628B2 (en) * 2020-04-08 2024-03-05 Arizona Board Of Regents On Behalf Of Arizona State University Systems, methods, and apparatuses for the generation of self-taught models genesis absent manual labeling for the processing of medical imaging
CN111626328B (en) * 2020-04-16 2023-12-15 湘潭大学 Image recognition method and device based on lightweight deep neural network
US20210334955A1 (en) * 2020-04-24 2021-10-28 Nvidia Corporation Image annotation using one or more neural networks
WO2021244754A1 (en) * 2020-06-05 2021-12-09 Huawei Technologies Co., Ltd. Image segmentation using multi-agent deep reinforcement learning
CN113808178A (en) * 2020-06-11 2021-12-17 通用电气精准医疗有限责任公司 Image registration method and model training method thereof
CN111832436B (en) * 2020-06-24 2023-06-16 五邑大学 Multi-task and weak supervision-based beauty prediction method and device and storage medium
CN111787222A (en) * 2020-06-30 2020-10-16 北京百度网讯科技有限公司 Method and device for adjusting camera, electronic equipment and storage medium
CN111968195B (en) * 2020-08-20 2022-09-02 太原科技大学 Dual-attention generation countermeasure network for low-dose CT image denoising and artifact removal
US20220084220A1 (en) * 2020-09-17 2022-03-17 Center For Deep Learning In Electronics Manufacturing, Inc. Methods and systems for registering images for electronic designs
CN112150524B (en) * 2020-09-29 2022-03-11 武汉大学 Two-dimensional and three-dimensional medical image registration method and system based on deep learning
CN112370161B (en) * 2020-10-12 2022-07-26 珠海横乐医学科技有限公司 Operation navigation method and medium based on ultrasonic image characteristic plane detection
CN112329848B (en) * 2020-11-04 2022-07-29 昆明理工大学 Image space mapping method based on advection vector field clustering
CN112419324B (en) * 2020-11-24 2022-04-19 山西三友和智慧信息技术股份有限公司 Medical image data expansion method based on semi-supervised task driving
KR102283673B1 (en) * 2020-11-30 2021-08-03 주식회사 코어라인소프트 Medical image reading assistant apparatus and method for adjusting threshold of diagnostic assistant information based on follow-up exam
CN112422946B (en) * 2020-11-30 2023-01-31 重庆邮电大学 Intelligent yoga action guidance system based on 3D reconstruction
CN112598028B (en) * 2020-12-10 2022-06-07 上海鹰瞳医疗科技有限公司 Eye fundus image registration model training method, eye fundus image registration method and eye fundus image registration device
CN112668757A (en) * 2020-12-15 2021-04-16 山东鲁能软件技术有限公司 Method, device and system for fusing GIS (geographic information System) graph and government control and regulation graph in power grid planning
CN112712507B (en) * 2020-12-31 2023-12-19 杭州依图医疗技术有限公司 Method and device for determining calcified region of coronary artery
CN112785670A (en) * 2021-02-01 2021-05-11 北京字节跳动网络技术有限公司 Image synthesis method, device, equipment and storage medium
CN113160290A (en) * 2021-03-31 2021-07-23 上海联影医疗科技股份有限公司 2D-3D medical image registration method and device and readable storage medium
EP4329875A1 (en) * 2021-04-28 2024-03-06 Elekta, Inc. Real-time anatomic position monitoring for radiotherapy treatment
US11679276B2 (en) 2021-04-28 2023-06-20 Elekta, Inc. Real-time anatomic position monitoring for radiotherapy treatment control
CN113298091A (en) * 2021-05-25 2021-08-24 商汤集团有限公司 Image processing method and device, electronic equipment and storage medium
US11823399B2 (en) * 2021-06-16 2023-11-21 Siemens Medical Solutions Usa, Inc. Multi-scan image processing
CN113742370B (en) * 2021-11-02 2022-04-19 阿里云计算有限公司 Data query method and statistical information ciphertext generation method of full-encryption database
WO2023115115A1 (en) * 2021-12-23 2023-06-29 Annalise-Ai Pty Ltd System and method for automatic registration of medical images
CN114046774B (en) * 2022-01-05 2022-04-08 中国测绘科学研究院 Ground deformation continuous monitoring method integrating CORS network and multi-source data
EP4273811A1 (en) * 2022-05-06 2023-11-08 Siemens Healthcare GmbH Technique for optimizing rendering parameters of overlays of medical images
CN115457259B (en) * 2022-09-14 2023-10-31 华洋通信科技股份有限公司 Image rapid saliency detection method based on multichannel activation optimization
CN115249245B (en) * 2022-09-22 2022-12-23 深圳华声医疗技术股份有限公司 Ultrasonic imaging parameter optimization method and device, computer equipment and storage medium
CN116942313B (en) * 2023-06-05 2024-01-16 北京长木谷医疗科技股份有限公司 Surgical robot registration method and device based on reinforcement learning and electronic equipment

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5580441A (en) 1993-12-16 1996-12-03 Kabushiki Kaisha Toshiba Method of measuring ion concentration and apparatus therefor
US8645294B1 (en) 2004-02-03 2014-02-04 Hrl Laboratories, Llc Method for image registration utilizing particle swarm optimization
US8068654B2 (en) 2007-02-02 2011-11-29 Siemens Akteingesellschaft Method and system for detection and registration of 3D objects using incremental parameter learning
US9547902B2 (en) 2008-09-18 2017-01-17 Siemens Healthcare Gmbh Method and system for physiological image registration and fusion
CN101916443B (en) * 2010-08-19 2012-10-17 中国科学院深圳先进技术研究院 Processing method and system of CT image
CN102385748B (en) 2010-08-31 2013-12-25 上海微创医疗器械(集团)有限公司 Image registration method
CN102411780B (en) * 2011-09-07 2014-07-02 华南理工大学 Registration-based CT (Computed Tomography) image total heart automatic cutting system
WO2014144019A1 (en) 2013-03-15 2014-09-18 The University Of North Carolina At Chapel Hill Methods, systems, and computer readable media for real-time 2d/3d deformable registration using metric learning
US10095917B2 (en) * 2013-11-04 2018-10-09 Facebook, Inc. Systems and methods for facial representation
KR102294734B1 (en) * 2014-09-30 2021-08-30 삼성전자주식회사 Method and apparatus for image registration, and ultrasonic diagnosis apparatus
CN104835112B (en) * 2015-05-07 2018-06-08 厦门大学 A kind of liver multiphase phase CT image interfusion methods
US20170032245A1 (en) 2015-07-01 2017-02-02 The Board Of Trustees Of The Leland Stanford Junior University Systems and Methods for Providing Reinforcement Learning in a Deep Learning System
US10235606B2 (en) 2015-07-22 2019-03-19 Siemens Healthcare Gmbh Method and system for convolutional neural network regression based 2D/3D image registration
US10521902B2 (en) * 2015-10-14 2019-12-31 The Regents Of The University Of California Automated segmentation of organ chambers using deep learning methods from medical imaging
US20170337682A1 (en) * 2016-05-18 2017-11-23 Siemens Healthcare Gmbh Method and System for Image Registration Using an Intelligent Artificial Agent
US10592743B2 (en) * 2017-08-24 2020-03-17 International Business Machines Corporation Machine learning to predict cognitive image composition

Cited By (279)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10874794B2 (en) 2011-06-20 2020-12-29 Cerner Innovation, Inc. Managing medication administration in clinical care room
US10546481B2 (en) 2011-07-12 2020-01-28 Cerner Innovation, Inc. Method for determining whether an individual leaves a prescribed virtual perimeter
US11380119B2 (en) * 2013-11-15 2022-07-05 Meta Platforms, Inc. Pose-aligned networks for deep attribute modeling
US10510443B2 (en) 2014-12-23 2019-12-17 Cerner Innovation, Inc. Methods and systems for determining whether a monitored individual's hand(s) have entered a virtual safety zone
US10524722B2 (en) 2014-12-26 2020-01-07 Cerner Innovation, Inc. Method and system for determining whether a caregiver takes appropriate measures to prevent patient bedsores
US10629046B2 (en) 2015-06-01 2020-04-21 Cerner Innovation, Inc. Systems and methods for determining whether an individual enters a prescribed virtual zone using skeletal tracking and 3D blob detection
US10959780B2 (en) * 2015-06-26 2021-03-30 Therenva Method and system for helping to guide an endovascular tool in vascular structures
US10643061B2 (en) 2015-12-31 2020-05-05 Cerner Innovation, Inc. Detecting unauthorized visitors
US10614288B2 (en) 2015-12-31 2020-04-07 Cerner Innovation, Inc. Methods and systems for detecting stroke symptoms
US11241169B2 (en) 2015-12-31 2022-02-08 Cerner Innovation, Inc. Methods and systems for detecting stroke symptoms
US10878220B2 (en) 2015-12-31 2020-12-29 Cerner Innovation, Inc. Methods and systems for assigning locations to devices
US10410042B2 (en) 2015-12-31 2019-09-10 Cerner Innovation, Inc. Detecting unauthorized visitors
US11937915B2 (en) 2015-12-31 2024-03-26 Cerner Innovation, Inc. Methods and systems for detecting stroke symptoms
US11363966B2 (en) 2015-12-31 2022-06-21 Cerner Innovation, Inc. Detecting unauthorized visitors
US11666246B2 (en) 2015-12-31 2023-06-06 Cerner Innovation, Inc. Methods and systems for assigning locations to devices
US11386557B2 (en) * 2016-01-21 2022-07-12 Elekta, Inc. Systems and methods for segmentation of intra-patient medical images
US10867385B2 (en) * 2016-01-21 2020-12-15 Elekta, Inc. Systems and methods for segmentation of intra-patient medical images
US11062209B2 (en) * 2016-05-20 2021-07-13 Magic Leap, Inc. Method and system for performing convolutional image transformation estimation
US10489708B2 (en) * 2016-05-20 2019-11-26 Magic Leap, Inc. Method and system for performing convolutional image transformation estimation
US20210365785A1 (en) * 2016-05-20 2021-11-25 Magic Leap, Inc. Method and system for performing convolutional image transformation estimation
US11593654B2 (en) * 2016-05-20 2023-02-28 Magic Leap, Inc. System for performing convolutional image transformation estimation
US11515030B2 (en) 2016-06-23 2022-11-29 Siemens Healthcare Gmbh System and method for artificial agent based cognitive operating rooms
US10794977B2 (en) * 2016-06-23 2020-10-06 Siemens Healthcare Gmbh System and method for normalized reference database for MR images via autoencoders
US20170371017A1 (en) * 2016-06-23 2017-12-28 Siemens Healthcare Gmbh System and Method For Normalized Reference Database For MR Images Via Autoencoders
US10628533B2 (en) * 2016-06-27 2020-04-21 Faro Technologies, Inc. Global optimization of networks of locally fitted objects
US20170371982A1 (en) * 2016-06-27 2017-12-28 Faro Technologies, Inc. Global optimization of networks of locally fitted objects
US10799315B2 (en) * 2016-07-26 2020-10-13 7D Surgical Inc. Systems and methods for verification of fiducial correspondence during image-guided surgical procedures
US10504004B2 (en) * 2016-09-16 2019-12-10 General Dynamics Mission Systems, Inc. Systems and methods for deep model translation generation
US20190080205A1 (en) * 2016-09-16 2019-03-14 Deep Learning Analytics, LLC Systems and Methods for Deep Model Translation Generation
US11003991B2 (en) * 2016-09-30 2021-05-11 Idemia Identity & Security France Methods for secure learning of parameters of a convolution neural network, and for secure input data classification
US11348258B2 (en) * 2016-11-02 2022-05-31 Shanghai United Imaging Healthcare Co., Ltd. System and method for registering multi-modality images
US10733746B2 (en) * 2016-11-02 2020-08-04 Shanghai United Imaging Healthcare Co., Ltd. System and method for registering multi-modality images
US10863946B2 (en) * 2016-11-03 2020-12-15 Siemens Medical Solutions Usa, Inc. Respiratory motion estimation in projection domain in nuclear medical imaging
US10896352B2 (en) * 2016-11-23 2021-01-19 General Electric Company Deep learning medical systems and methods for image reconstruction and quality evaluation
US20200097773A1 (en) * 2016-11-23 2020-03-26 General Electric Company Deep learning medical systems and methods for image reconstruction and quality evaluation
US10565477B2 (en) 2016-11-23 2020-02-18 General Electric Company Deep learning medical systems and methods for image reconstruction and quality evaluation
US10074038B2 (en) * 2016-11-23 2018-09-11 General Electric Company Deep learning medical systems and methods for image reconstruction and quality evaluation
US10354171B2 (en) 2016-11-23 2019-07-16 General Electric Company Deep learning medical systems and methods for image reconstruction and quality evaluation
US10388016B2 (en) 2016-12-30 2019-08-20 Cerner Innovation, Inc. Seizure detection
US10504226B2 (en) 2016-12-30 2019-12-10 Cerner Innovation, Inc. Seizure detection
US11571256B2 (en) * 2017-01-12 2023-02-07 Mazor Robotics Ltd. Image based pathology prediction using artificial intelligence
US20200038109A1 (en) * 2017-01-12 2020-02-06 Mazor Robotics Ltd. Image based pathology prediction using artificial intelligence
US11625852B1 (en) 2017-01-19 2023-04-11 X Development Llc Object pose neural network system
US10861184B1 (en) * 2017-01-19 2020-12-08 X Development Llc Object pose neural network system
US10402969B2 (en) * 2017-03-10 2019-09-03 General Electric Company Methods and systems for model driven multi-modal medical imaging
US10296828B2 (en) * 2017-04-05 2019-05-21 Here Global B.V. Learning a similarity measure for vision-based localization on a high definition (HD) map
US10366494B2 (en) * 2017-04-21 2019-07-30 Samsung Medison Co., Ltd. Image segmentation method and electronic device therefor
US20180374216A1 (en) * 2017-05-02 2018-12-27 Aivitae LLC System and method for facilitating autonomous control of an imaging system
US11276163B2 (en) * 2017-05-02 2022-03-15 Alvitae LLC System and method for facilitating autonomous control of an imaging system
US11321827B2 (en) * 2017-05-02 2022-05-03 Aivitae LLC System and method for facilitating autonomous control of an imaging system
US20200065673A1 (en) * 2017-05-10 2020-02-27 Telefonaktiebolaget Lm Ericsson (Publ) Pre-training system for self-learning agent in virtualized environment
US11586911B2 (en) * 2017-05-10 2023-02-21 Telefonaktiebolaget Lm Ericsson (Publ) Pre-training system for self-learning agent in virtualized environment
US20180330511A1 (en) * 2017-05-11 2018-11-15 Kla-Tencor Corporation Learning based approach for aligning images acquired with different modalities
US10733744B2 (en) * 2017-05-11 2020-08-04 Kla-Tencor Corp. Learning based approach for aligning images acquired with different modalities
US20220318516A1 (en) * 2017-06-05 2022-10-06 Deepmind Technologies Limited Action selection based on environment observations and textual instructions
US11109945B2 (en) * 2017-06-09 2021-09-07 Dental Monitoring Method of evaluating an orthodontic aligner
US11200416B2 (en) * 2017-06-14 2021-12-14 Beijing Sensetime Technology Development Co., Ltd Methods and apparatuses for image detection, electronic devices and storage media
US10424045B2 (en) * 2017-06-21 2019-09-24 International Business Machines Corporation Machine learning model for automatic image registration quality assessment and correction
US20180374193A1 (en) * 2017-06-21 2018-12-27 International Business Machines Corporation Automatic Image Registration Quality Assessment
US20180374194A1 (en) * 2017-06-21 2018-12-27 International Business Machines Corporation Automatic Image Registration Quality Assessment
US10417737B2 (en) * 2017-06-21 2019-09-17 International Business Machines Corporation Machine learning model for automatic image registration quality assessment and correction
US10679325B2 (en) * 2017-06-21 2020-06-09 International Business Machines Corporation Machine learning model for automatic image registration quality assessment and correction
US11003982B2 (en) * 2017-06-27 2021-05-11 D5Ai Llc Aligned training of deep networks
US20190043185A1 (en) * 2017-08-04 2019-02-07 Omron Corporation Image processing system
US10885620B2 (en) * 2017-08-04 2021-01-05 Omron Corporation Neural network image processing system
US10818019B2 (en) * 2017-08-14 2020-10-27 Siemens Healthcare Gmbh Dilated fully convolutional network for multi-agent 2D/3D medical image registration
US11354813B2 (en) 2017-08-14 2022-06-07 Siemens Healthcare Gmbh Dilated fully convolutional network for 2D/3D medical image registration
US20190057505A1 (en) * 2017-08-17 2019-02-21 Siemens Healthcare Gmbh Automatic change detection in medical images
US10699410B2 (en) * 2017-08-17 2020-06-30 Siemes Healthcare GmbH Automatic change detection in medical images
US20190083067A1 (en) * 2017-09-21 2019-03-21 General Electric Company Methods and systems for correction of one dimensional shear wave data
US10810735B2 (en) * 2017-09-30 2020-10-20 Baidu Online Network Technology (Beijing) Co., Ltd. Method and apparatus for analyzing medical image
US20190102878A1 (en) * 2017-09-30 2019-04-04 Baidu Online Network Technology (Beijing) Co., Ltd. Method and apparatus for analyzing medical image
US11847761B2 (en) * 2017-10-06 2023-12-19 Canon Medical Systems Corporation Medical image processing apparatus having a plurality of neural networks corresponding to different fields of view
US11517197B2 (en) 2017-10-06 2022-12-06 Canon Medical Systems Corporation Apparatus and method for medical image reconstruction using deep learning for computed tomography (CT) image noise and artifacts reduction
US11127153B2 (en) * 2017-10-10 2021-09-21 Hitachi, Ltd. Radiation imaging device, image processing method, and image processing program
KR102112997B1 (en) 2017-10-13 2020-06-02 주식회사 스트라드비젼 Method and device for performing activation and convolution operation at the same time, learning method and learning device for the same
KR20190041922A (en) * 2017-10-13 2019-04-23 주식회사 스트라드비젼 Method and device for performing activation and convolution operation at the same time, learning method and learning device for the same
US10049323B1 (en) * 2017-10-13 2018-08-14 StradVision, Inc. Method and device for performing activation and convolution operation at the same time and learning method and learning device for the same
US10769791B2 (en) * 2017-10-13 2020-09-08 Beijing Keya Medical Technology Co., Ltd. Systems and methods for cross-modality image segmentation
US10937141B2 (en) * 2017-10-13 2021-03-02 Sualab Co., Ltd. Deep learning based image comparison device, method and computer program stored in computer readable medium
US11604941B1 (en) * 2017-10-27 2023-03-14 Deepmind Technologies Limited Training action-selection neural networks from demonstrations using multiple losses
US10929989B2 (en) 2017-11-01 2021-02-23 Siemens Healthcare Gmbh Determining a transformation between coordinate frames of sets of image data
US10984905B2 (en) * 2017-11-03 2021-04-20 Siemens Healthcare Gmbh Artificial intelligence for physiological quantification in medical imaging
US10776688B2 (en) * 2017-11-06 2020-09-15 Nvidia Corporation Multi-frame video interpolation using optical flow
US20190138889A1 (en) * 2017-11-06 2019-05-09 Nvidia Corporation Multi-frame video interpolation using optical flow
US11599979B2 (en) * 2017-11-09 2023-03-07 Samsung Electronics Co., Ltd Method and apparatus for video super resolution using convolutional neural network with two-stage motion compensation
US20200294217A1 (en) * 2017-11-09 2020-09-17 Samsung Electronics Co., Ltd. Method and apparatus for video super resolution using convolutional neural network with two-stage motion compensation
US11756668B2 (en) * 2017-11-12 2023-09-12 Aleph Bot Ltd Systems methods devices circuits and computer executable code for tracking evaluating and facilitating a medical procedure
US11361442B2 (en) * 2017-11-16 2022-06-14 Oslo Universitetssykehus Histological image analysis
US11341646B2 (en) 2017-11-22 2022-05-24 General Electric Company Systems and methods to deliver point of care alerts for radiological findings
US10783634B2 (en) * 2017-11-22 2020-09-22 General Electric Company Systems and methods to deliver point of care alerts for radiological findings
US11712208B2 (en) 2017-11-22 2023-08-01 General Electric Company Systems and methods to deliver point of care alerts for radiological findings
US11521064B2 (en) * 2017-12-01 2022-12-06 Koninklijke Philips N.V. Training a neural network model
US11062467B2 (en) 2017-12-20 2021-07-13 International Business Machines Corporation Medical image registration guided by target lesion
US10593051B2 (en) 2017-12-20 2020-03-17 International Business Machines Corporation Medical image registration guided by target lesion
US11645833B2 (en) * 2017-12-21 2023-05-09 Merative Us L.P. Generative adversarial network medical image generation for training of a classifier
US20220076075A1 (en) * 2017-12-21 2022-03-10 International Business Machines Corporation Generative Adversarial Network Medical Image Generation for Training of a Classifier
US10937540B2 (en) 2017-12-21 2021-03-02 International Business Machines Coporation Medical image classification based on a generative adversarial network trained discriminator
US10540578B2 (en) 2017-12-21 2020-01-21 International Business Machines Corporation Adapting a generative adversarial network to new data sources for image classification
US10592779B2 (en) * 2017-12-21 2020-03-17 International Business Machines Corporation Generative adversarial network medical image generation for training of a classifier
EP3502978A1 (en) 2017-12-22 2019-06-26 Siemens Healthcare GmbH Meta-learning system
US11861500B2 (en) 2017-12-22 2024-01-02 Siemens Healthcare Gmbh Meta-learning system
US20220130225A1 (en) * 2017-12-28 2022-04-28 Cerner Innovation, Inc. Utilizing artificial intelligence to detect objects or patient safety events in a patient room
US20190206218A1 (en) * 2017-12-28 2019-07-04 Cerner Innovation, Inc. Utilizing artificial intelligence to detect objects or patient safety events in a patient room
US10643446B2 (en) * 2017-12-28 2020-05-05 Cerner Innovation, Inc. Utilizing artificial intelligence to detect objects or patient safety events in a patient room
US10922946B2 (en) * 2017-12-28 2021-02-16 Cerner Innovation, Inc. Utilizing artificial intelligence to detect objects or patient safety events in a patient room
US11276291B2 (en) * 2017-12-28 2022-03-15 Cerner Innovation, Inc. Utilizing artificial intelligence to detect objects or patient safety events in a patient room
US11721190B2 (en) * 2017-12-28 2023-08-08 Cerner Innovation, Inc. Utilizing artificial intelligence to detect objects or patient safety events in a patient room
US11544953B2 (en) 2017-12-29 2023-01-03 Cerner Innovation, Inc. Methods and systems for identifying the crossing of a virtual barrier
US10482321B2 (en) 2017-12-29 2019-11-19 Cerner Innovation, Inc. Methods and systems for identifying the crossing of a virtual barrier
US11074440B2 (en) 2017-12-29 2021-07-27 Cerner Innovation, Inc. Methods and systems for identifying the crossing of a virtual barrier
US11449759B2 (en) * 2018-01-03 2022-09-20 Siemens Heathcare Gmbh Medical imaging diffeomorphic registration based on machine learning
US11151726B2 (en) * 2018-01-10 2021-10-19 Canon Medical Systems Corporation Medical image processing apparatus, X-ray diagnostic apparatus, and medical image processing method
US10475538B2 (en) * 2018-01-11 2019-11-12 Siemens Healthcare Gmbh System for predicting computed tomography data from surface data
US10709394B2 (en) * 2018-01-15 2020-07-14 Siemens Healthcare Gmbh Method and system for 3D reconstruction of X-ray CT volume and segmentation mask from a few X-ray radiographs
US20190216409A1 (en) * 2018-01-15 2019-07-18 Siemens Healthcare Gmbh Method and System for 3D Reconstruction of X-ray CT Volume and Segmentation Mask from a Few X-ray Radiographs
US11263764B2 (en) 2018-01-18 2022-03-01 Elekta, Inc. Methods and devices for surface motion tracking
US10748296B2 (en) * 2018-01-18 2020-08-18 Elekta, Inc. Methods and devices for surface motion tracking
CN108492309A (en) * 2018-01-21 2018-09-04 西安电子科技大学 Magnetic resonance image medium sized vein blood vessel segmentation method based on migration convolutional neural networks
US11568236B2 (en) 2018-01-25 2023-01-31 The Research Foundation For The State University Of New York Framework and methods of diverse exploration for fast and safe policy improvement
US20190236360A1 (en) * 2018-01-30 2019-08-01 Mashgin Inc. Feedback loop for image-based recognition
US10540551B2 (en) 2018-01-30 2020-01-21 Mashgin Inc. Generation of two-dimensional and three-dimensional images of items for visual recognition in checkout apparatus
US11710241B2 (en) 2018-02-14 2023-07-25 Elekta, Inc. Atlas-based segmentation using deep-learning
US11132792B2 (en) * 2018-02-22 2021-09-28 Siemens Healthcare Gmbh Cross domain medical image segmentation
WO2019166447A1 (en) * 2018-03-02 2019-09-06 Koninklijke Philips N.V. Image registration qualification
US11380001B2 (en) * 2018-03-02 2022-07-05 Koninklijke Philips N.V. Image registration qualification
CN112055870A (en) * 2018-03-02 2020-12-08 皇家飞利浦有限公司 Image registration qualification
US11813047B2 (en) * 2018-03-07 2023-11-14 University Of Virginia Patent Foundation Automatic quantification of cardiac MRI for hypertrophic cardiomyopathy
US20210287367A1 (en) * 2018-03-07 2021-09-16 University Of Virginia Patent Foundation Automatic quantification of cardiac mri for hypertrophic cardiomyopathy
US11024025B2 (en) * 2018-03-07 2021-06-01 University Of Virginia Patent Foundation Automatic quantification of cardiac MRI for hypertrophic cardiomyopathy
US20190279361A1 (en) * 2018-03-07 2019-09-12 University Of Virginia Patent Foundation Automatic quantification of cardiac mri for hypertrophic cardiomyopathy
US11869057B2 (en) * 2018-03-16 2024-01-09 Ebay Inc. Generating a digital image using a generative adversarial network
US11222246B2 (en) * 2018-03-16 2022-01-11 Ebay Inc. Generating a digital image using a generative adversarial network
US20220092367A1 (en) * 2018-03-16 2022-03-24 Ebay Inc. Generating a digital image using a generative adversarial network
US10552714B2 (en) * 2018-03-16 2020-02-04 Ebay Inc. Generating a digital image using a generative adversarial network
US10997462B2 (en) * 2018-04-04 2021-05-04 Canon Virginia, Inc. Devices, systems, and methods for clustering reference images for non-destructive testing
US11551037B2 (en) 2018-04-23 2023-01-10 Siemens Aktiengesellschaft Method and apparatus for determining a physical shape, method for manufacturing a calculation device, calculation device, and use of the calculation device
US10242264B1 (en) * 2018-04-25 2019-03-26 Imagry (Israel) Ltd. System and method for training a machine-learning model to identify real-world elements
US10956785B2 (en) 2018-04-27 2021-03-23 Arizona Board Of Regents On Behalf Of Arizona State University Methods, systems, and media for selecting candidates for annotation for use in training classifiers
US20210056717A1 (en) * 2018-05-01 2021-02-25 Koninklijke Philips N.V. Lower to higher resolution image fusion
US10957041B2 (en) 2018-05-14 2021-03-23 Tempus Labs, Inc. Determining biomarkers from histopathology slide images
US11348661B2 (en) 2018-05-14 2022-05-31 Tempus Labs, Inc. Predicting total nucleic acid yield and dissection boundaries for histology slides
US11348240B2 (en) 2018-05-14 2022-05-31 Tempus Labs, Inc. Predicting total nucleic acid yield and dissection boundaries for histology slides
US11682098B2 (en) 2018-05-14 2023-06-20 Tempus Labs, Inc. Determining biomarkers from histopathology slide images
US11935152B2 (en) 2018-05-14 2024-03-19 Tempus Labs, Inc. Determining biomarkers from histopathology slide images
WO2019222289A1 (en) * 2018-05-14 2019-11-21 Tempus Labs, Inc. A generalizable and interpretable deep learning framework for predicting msi from histopathology slide images
US11610307B2 (en) 2018-05-14 2023-03-21 Tempus Labs, Inc. Determining biomarkers from histopathology slide images
US11348239B2 (en) 2018-05-14 2022-05-31 Tempus Labs, Inc. Predicting total nucleic acid yield and dissection boundaries for histology slides
US11263748B2 (en) 2018-05-14 2022-03-01 Tempus Labs, Inc. Determining biomarkers from histopathology slide images
US11741365B2 (en) 2018-05-14 2023-08-29 Tempus Labs, Inc. Generalizable and interpretable deep learning framework for predicting MSI from histopathology slide images
US20190353741A1 (en) * 2018-05-16 2019-11-21 Siemens Healthcare Gmbh Deep Learning Reconstruction of Free Breathing Perfusion
US10782378B2 (en) * 2018-05-16 2020-09-22 Siemens Healthcare Gmbh Deep learning reconstruction of free breathing perfusion
US11436780B2 (en) * 2018-05-24 2022-09-06 Warner Bros. Entertainment Inc. Matching mouth shape and movement in digital video to alternative audio
US11087185B2 (en) 2018-05-25 2021-08-10 Samsung Electronics Co., Ltd. Image processing method and apparatus with neural network adjustment
US11636698B2 (en) 2018-05-25 2023-04-25 Samsung Electronics Co., Ltd. Image processing method and apparatus with neural network adjustment
CN110634117A (en) * 2018-05-31 2019-12-31 中山大学肿瘤防治中心(中山大学附属肿瘤医院、中山大学肿瘤研究所) Method for quantitatively evaluating image by using three-dimensional parameters
US20190377943A1 (en) * 2018-06-12 2019-12-12 Wipro Limited Method and system for correcting fabrication in a document
US10769430B2 (en) * 2018-06-12 2020-09-08 Wipro Limited Method and system for correcting fabrication in a document
US11550011B2 (en) * 2018-06-15 2023-01-10 Subtle Medical, Inc. Systems and methods for magnetic resonance imaging standardization using deep learning
CN112313668A (en) * 2018-06-15 2021-02-02 深透医疗公司 System and method for magnetic resonance imaging normalization using deep learning
US11756160B2 (en) * 2018-07-27 2023-09-12 Washington University ML-based methods for pseudo-CT and HR MR image estimation
CN109035137A (en) * 2018-07-27 2018-12-18 重庆邮电大学 A kind of multimode medical image fusion method based on optimal transmission theory
US11734575B2 (en) 2018-07-30 2023-08-22 International Business Machines Corporation Sequential learning of constraints for hierarchical reinforcement learning
US11537872B2 (en) 2018-07-30 2022-12-27 International Business Machines Corporation Imitation learning by action shaping with antagonist reinforcement learning
US11501157B2 (en) 2018-07-30 2022-11-15 International Business Machines Corporation Action shaping from demonstration for fast reinforcement learning
US11568549B2 (en) 2018-08-03 2023-01-31 Logitech Europe S.A. Method and system for detecting peripheral device displacement
US11042992B2 (en) * 2018-08-03 2021-06-22 Logitech Europe S.A. Method and system for detecting peripheral device displacement
US20200043179A1 (en) * 2018-08-03 2020-02-06 Logitech Europe S.A. Method and system for detecting peripheral device displacement
CN110796259A (en) * 2018-08-03 2020-02-14 罗技欧洲公司 Method and system for determining peripheral device displacement
CN109035316A (en) * 2018-08-28 2018-12-18 北京安德医智科技有限公司 The method for registering and equipment of nuclear magnetic resonance image sequence
US10565476B1 (en) * 2018-09-04 2020-02-18 StradVision, Inc. Method and computing device for generating image data set for learning to be used for detection of obstruction in autonomous driving circumstances and learning method and learning device using the same
US11887301B2 (en) 2018-10-19 2024-01-30 Mirada Medical Limited System and method for automatic delineation of scanned images
WO2020078725A1 (en) * 2018-10-19 2020-04-23 Mirada Medical Limited System and method for automatic delineation of scanned images
US11443602B2 (en) 2018-11-06 2022-09-13 Cerner Innovation, Inc. Methods and systems for detecting prohibited objects
US10922936B2 (en) 2018-11-06 2021-02-16 Cerner Innovation, Inc. Methods and systems for detecting prohibited objects
US10842445B2 (en) * 2018-11-08 2020-11-24 General Electric Company System and method for unsupervised deep learning for deformable image registration
US20210274208A1 (en) * 2018-11-15 2021-09-02 Beijing Bytedance Network Technology Co., Ltd. Merge with mvd for affine
US11677973B2 (en) * 2018-11-15 2023-06-13 Beijing Bytedance Network Technology Co., Ltd Merge with MVD for affine
WO2020113148A1 (en) * 2018-11-29 2020-06-04 The Board Of Trustees Of The Leland Sanford Junior University Single or a few views computed tomography imaging with deep neural network
US11054271B2 (en) 2018-12-04 2021-07-06 Here Global B.V. Method and apparatus for aligning drive segments based on user inputs
WO2020118425A1 (en) * 2018-12-10 2020-06-18 Her Majesty The Queen In Right Of Canada, As Represented By The Minister Of National Defence Automated diagnostics in 3d ultrasound system and method
US11158069B2 (en) * 2018-12-11 2021-10-26 Siemens Healthcare Gmbh Unsupervised deformable registration for multi-modal images
CN109712175A (en) * 2018-12-19 2019-05-03 浙江大学常州工业技术研究院 The method for registering of CT picture
JP2022505498A (en) * 2018-12-19 2022-01-14 シャンハイ センスタイム インテリジェント テクノロジー カンパニー リミテッド Image processing methods, devices, electronic devices and computer readable storage media
CN111340859A (en) * 2018-12-19 2020-06-26 西门子医疗有限公司 Method for image registration, learning device and medical imaging device
US20200202622A1 (en) * 2018-12-19 2020-06-25 Nvidia Corporation Mesh reconstruction using data-driven priors
US11080849B2 (en) * 2018-12-21 2021-08-03 General Electric Company Systems and methods for deep learning based automated spine registration and label propagation
US20210327063A1 (en) * 2018-12-21 2021-10-21 GE Precision Healthcare LLC Systems and methods for whole-body spine labeling
US11475565B2 (en) * 2018-12-21 2022-10-18 GE Precision Healthcare LLC Systems and methods for whole-body spine labeling
US20200202515A1 (en) * 2018-12-21 2020-06-25 General Electric Company Systems and methods for deep learning based automated spine registration and label propagation
CN109599109A (en) * 2018-12-26 2019-04-09 浙江大学 For the confrontation audio generation method and system of whitepack scene
US10991097B2 (en) 2018-12-31 2021-04-27 Tempus Labs, Inc. Artificial intelligence segmentation of tissue images
US10540798B1 (en) 2019-01-10 2020-01-21 Capital One Services, Llc Methods and arrangements to create images
CN109767459A (en) * 2019-01-17 2019-05-17 中南大学 Novel ocular base map method for registering
US11468260B2 (en) 2019-02-25 2022-10-11 Fair Isaac Corporation Density based confidence measures of neural networks for reliable predictions
US11003947B2 (en) * 2019-02-25 2021-05-11 Fair Isaac Corporation Density based confidence measures of neural networks for reliable predictions
US20200272853A1 (en) * 2019-02-25 2020-08-27 Fair Isaac Corporation Density based confidence measures of neural networks for reliable predictions
CN111739614A (en) * 2019-03-25 2020-10-02 西门子医疗有限公司 Medical image enhancement
US11854158B2 (en) 2019-03-25 2023-12-26 Siemens Healthcare Gmbh Medical image enhancement
US11948311B2 (en) 2019-03-27 2024-04-02 The General Hospital Corporation Retrospective motion correction using a combined neural network and model-based image reconstruction of magnetic resonance data
WO2020198456A1 (en) * 2019-03-27 2020-10-01 The General Hospital Corporation Retrospective motion correction using a combined neural network and model-based image reconstruction of magnetic resonance data
WO2020198854A1 (en) * 2019-03-29 2020-10-08 Polyvalor, Limited Partnership Method and system for producing medical images
US10996306B2 (en) * 2019-04-25 2021-05-04 General Electric Company MRI system and method using neural network for detection of patient motion
CN110074813A (en) * 2019-04-26 2019-08-02 深圳大学 A kind of ultrasonic image reconstruction method and system
US20220225955A1 (en) * 2019-04-26 2022-07-21 Spectronic Ab Generating synthetic electron density images from magnetic resonance images
US10489936B1 (en) * 2019-04-29 2019-11-26 Deep Render Ltd. System and method for lossy image and video compression utilizing a metanetwork
US11164021B2 (en) * 2019-05-16 2021-11-02 Arizona Board Of Regents On Behalf Of Arizona State University Methods, systems, and media for discriminating and generating translated images
US11067786B2 (en) * 2019-06-07 2021-07-20 Leica Microsystems Inc. Artifact regulation methods in deep model training for image transformation
CN110211165A (en) * 2019-06-10 2019-09-06 成都信息工程大学 A kind of image multi-mode registration method based on the study of asynchronous deeply
US11853901B2 (en) 2019-07-26 2023-12-26 Samsung Electronics Co., Ltd. Learning method of AI model and electronic apparatus
US20210034974A1 (en) * 2019-07-31 2021-02-04 Royal Bank Of Canada Devices and methods for reinforcement learning visualization using immersive environments
US11200456B2 (en) * 2019-07-31 2021-12-14 GE Precision Healthcare LLC Systems and methods for generating augmented training data for machine learning models
US11720792B2 (en) * 2019-07-31 2023-08-08 Royal Bank Of Canada Devices and methods for reinforcement learning visualization using immersive environments
CN110580695A (en) * 2019-08-07 2019-12-17 深圳先进技术研究院 multi-mode three-dimensional medical image fusion method and system and electronic equipment
WO2021036471A1 (en) * 2019-08-28 2021-03-04 腾讯科技(深圳)有限公司 Sample generation method and apparatus, and computer device and storage medium
CN110598765A (en) * 2019-08-28 2019-12-20 腾讯科技(深圳)有限公司 Sample generation method and device, computer equipment and storage medium
US10957031B1 (en) * 2019-09-06 2021-03-23 Accenture Global Solutions Limited Intelligent defect detection from image data
US10943353B1 (en) * 2019-09-11 2021-03-09 International Business Machines Corporation Handling untrainable conditions in a network architecture search
US11023783B2 (en) 2019-09-11 2021-06-01 International Business Machines Corporation Network architecture search with global optimization
CN112494028A (en) * 2019-09-13 2021-03-16 通用电气公司 Biopsy workflow using multi-modality imaging
US11304683B2 (en) * 2019-09-13 2022-04-19 General Electric Company Biopsy workflow using multimodal imaging
US11571809B1 (en) * 2019-09-15 2023-02-07 X Development Llc Robotic control using value distributions
US11270110B2 (en) 2019-09-17 2022-03-08 Boston Polarimetrics, Inc. Systems and methods for surface modeling using polarization cues
US11699273B2 (en) 2019-09-17 2023-07-11 Intrinsic Innovation Llc Systems and methods for surface modeling using polarization cues
US10655975B2 (en) * 2019-09-23 2020-05-19 Alibaba Group Holding Limited System and method for routing optimization
US11250580B2 (en) * 2019-09-24 2022-02-15 Dentsply Sirona Inc. Method, system and computer readable storage media for registering intraoral measurements
US11525906B2 (en) 2019-10-07 2022-12-13 Intrinsic Innovation Llc Systems and methods for augmentation of sensor systems and imaging systems with polarization
WO2021077009A1 (en) * 2019-10-16 2021-04-22 REHABILITATION INSTITUTE OF CHICAGO d/b/a Shirley Ryan AbilityLab Systems and methods for electromyogram-based control using neural networks
US11405637B2 (en) * 2019-10-29 2022-08-02 Samsung Electronics Co., Ltd. Image encoding method and apparatus and image decoding method and apparatus
US20210133984A1 (en) * 2019-11-04 2021-05-06 Shanghai United Imaging Intelligence Co., Ltd. Systems and methods for machine learning based physiological motion measurement
US11710244B2 (en) * 2019-11-04 2023-07-25 Shanghai United Imaging Intelligence Co., Ltd. Systems and methods for machine learning based physiological motion measurement
CN110827331A (en) * 2019-11-04 2020-02-21 上海联影智能医疗科技有限公司 Training method of image registration model, image registration method and computer equipment
US11699224B2 (en) 2019-11-18 2023-07-11 Stmicroelectronics (Rousset) Sas Neural network training device, system and method
US11200659B2 (en) 2019-11-18 2021-12-14 Stmicroelectronics (Rousset) Sas Neural network training device, system and method
US11501424B2 (en) 2019-11-18 2022-11-15 Stmicroelectronics (Rousset) Sas Neural network training device, system and method
US11302012B2 (en) * 2019-11-30 2022-04-12 Boston Polarimetrics, Inc. Systems and methods for transparent object segmentation using polarization cues
US11842495B2 (en) * 2019-11-30 2023-12-12 Intrinsic Innovation Llc Systems and methods for transparent object segmentation using polarization cues
US20220198673A1 (en) * 2019-11-30 2022-06-23 Boston Polarimetrics, Inc. Systems and methods for transparent object segmentation using polarization cues
US11875441B2 (en) * 2019-12-03 2024-01-16 Disney Enterprises, Inc. Data-driven extraction and composition of secondary dynamics in facial performance capture
US11263744B2 (en) * 2019-12-09 2022-03-01 Siemens Healthcare Gmbh Saliency mapping by feature reduction and perturbation modeling in medical imaging
US20210170579A1 (en) * 2019-12-09 2021-06-10 Canon Kabushiki Kaisha Robot controller that controls robot, learned model, method of controlling robot, and storage medium
US11679496B2 (en) * 2019-12-09 2023-06-20 Canon Kabushiki Kaisha Robot controller that controls robot, learned model, method of controlling robot, and storage medium
CN111199227A (en) * 2019-12-20 2020-05-26 广西柳州联耕科技有限公司 High-precision image identification method
CN111161227A (en) * 2019-12-20 2020-05-15 成都数之联科技有限公司 Target positioning method and system based on deep neural network
CN111144486A (en) * 2019-12-27 2020-05-12 电子科技大学 Heart nuclear magnetic resonance image key point detection method based on convolutional neural network
CN111311652A (en) * 2020-01-19 2020-06-19 杭州电子科技大学 Regression correction network-based sonar image registration method and system
US20210304457A1 (en) * 2020-03-31 2021-09-30 The Regents Of The University Of California Using neural networks to estimate motion vectors for motion corrected pet image reconstruction
WO2021202203A1 (en) * 2020-03-31 2021-10-07 Alibaba Group Holding Limited Medical image processing method and system and data processing method
US11941802B2 (en) 2020-03-31 2024-03-26 Alibaba Group Holding Limited Data processing method, means and system
US20220092789A1 (en) * 2020-04-09 2022-03-24 Zhejiang Lab Automatic pancreas ct segmentation method based on a saliency-aware densely connected dilated convolutional neural network
US11562491B2 (en) * 2020-04-09 2023-01-24 Zhejiang Lab Automatic pancreas CT segmentation method based on a saliency-aware densely connected dilated convolutional neural network
CN113538529A (en) * 2020-04-15 2021-10-22 杭州海康威视数字技术股份有限公司 Image registration apparatus
CN111462146A (en) * 2020-04-16 2020-07-28 成都信息工程大学 Medical image multi-mode registration method based on space-time intelligent agent
US11900620B2 (en) * 2020-05-20 2024-02-13 Therenva Method and system for registering images containing anatomical structures
US20210366135A1 (en) * 2020-05-20 2021-11-25 Therenva Method and system for registering images containing anatomical structures
CN111950584A (en) * 2020-06-16 2020-11-17 江西中科九峰智慧医疗科技有限公司 Intelligent identification method and system for integrity of part in X-ray chest radiography
US20210397886A1 (en) * 2020-06-22 2021-12-23 Shanghai United Imaging Intelligence Co., Ltd. Anatomy-aware motion estimation
US11693919B2 (en) * 2020-06-22 2023-07-04 Shanghai United Imaging Intelligence Co., Ltd. Anatomy-aware motion estimation
WO2022005337A1 (en) * 2020-06-29 2022-01-06 Autonomous Non-Profit Organization For Higher Education «Skolkovo Institute Of Science And Technology» Veins mask projection alignment
US20220013231A1 (en) * 2020-07-13 2022-01-13 Corsmed Ab Method for ai applications in mri simulation
CN111950194A (en) * 2020-07-22 2020-11-17 西南大学 Newton momentum-based distributed acceleration composite optimization method and system
CN112365422A (en) * 2020-11-17 2021-02-12 重庆邮电大学 Irregular missing image restoration method and system based on deep aggregation network
CN112802046A (en) * 2021-01-28 2021-05-14 华南理工大学 Image generation system for generating pseudo CT from multi-sequence MR based on deep learning
WO2022192015A1 (en) * 2021-03-10 2022-09-15 Intrinsic Innovation Llc Systems and methods for high dynamic range image reconstruction
CN113096169A (en) * 2021-03-31 2021-07-09 华中科技大学 Non-rigid multimode medical image registration model establishing method and application thereof
CN113052879A (en) * 2021-04-08 2021-06-29 西安应用光学研究所 Multispectral image automatic registration method
CN112990359A (en) * 2021-04-19 2021-06-18 深圳市深光粟科技有限公司 Image data processing method and device, computer and storage medium
CN113240699A (en) * 2021-05-20 2021-08-10 推想医疗科技股份有限公司 Image processing method and device, model training method and device, and electronic equipment
WO2022247296A1 (en) * 2021-05-27 2022-12-01 广州柏视医疗科技有限公司 Mark point-based image registration method
WO2022247218A1 (en) * 2021-05-27 2022-12-01 广州柏视医疗科技有限公司 Image registration method based on automatic delineation
US20230076196A1 (en) * 2021-09-03 2023-03-09 Adobe Inc. Asynchronous multi-user real-time streaming of web-based image edits using generative adversarial network(s)
US11928762B2 (en) * 2021-09-03 2024-03-12 Adobe Inc. Asynchronous multi-user real-time streaming of web-based image edits using generative adversarial network(s)
WO2023049211A3 (en) * 2021-09-21 2023-06-01 Hyperfine Operations, Inc. Contrastive multimodality image registration
US20230169659A1 (en) * 2021-11-30 2023-06-01 Shanghai United Imaging Intelligence Co., Ltd. Image segmentation and tracking based on statistical shape model
CN114332509A (en) * 2021-12-29 2022-04-12 阿波罗智能技术(北京)有限公司 Image processing method, model training method, electronic device and automatic driving vehicle
WO2023172400A1 (en) * 2022-03-10 2023-09-14 Varian Medical Systems, Inc. Artificial intelligence boosting dose calculations
CN114663840A (en) * 2022-05-13 2022-06-24 浙江有色地球物理技术应用研究院有限公司 Tunnel environment inspection equipment monitoring method and system
CN115393402A (en) * 2022-08-24 2022-11-25 北京医智影科技有限公司 Training method of image registration network model, image registration method and equipment
CN115762722A (en) * 2022-11-22 2023-03-07 南方医科大学珠江医院 Image review system based on artificial intelligence

Also Published As

Publication number Publication date
EP3246875A3 (en) 2018-04-11
US11741605B2 (en) 2023-08-29
US20200258227A1 (en) 2020-08-13
EP3246875A2 (en) 2017-11-22
US11557036B2 (en) 2023-01-17
US20230368383A1 (en) 2023-11-16
US20230114934A1 (en) 2023-04-13
CN107403446A (en) 2017-11-28
CN107403446B (en) 2023-04-21

Similar Documents

Publication Publication Date Title
US11741605B2 (en) Method and system for image registration using an intelligent artificial agent
US11393229B2 (en) Method and system for artificial intelligence based medical image segmentation
Zhou et al. Deep reinforcement learning in medical imaging: A literature review
US11185231B2 (en) Intelligent multi-scale medical image landmark detection
Ghesu et al. Multi-scale deep reinforcement learning for real-time 3D-landmark detection in CT scans
Liao et al. An artificial agent for robust image registration
Al et al. Partial policy-based reinforcement learning for anatomical landmark localization in 3d medical images
KR101908520B1 (en) Landmark detection with spatial and temporal constraints in medical imaging
KR102442090B1 (en) Point registration method in surgical navigation system
CN114846524A (en) Medical image analysis using machine learning and anatomical vectors
Xu et al. Deep reinforcement learning and its applications in medical imaging and radiation therapy: a survey
Miao et al. Agent-based methods for medical image registration
Schwab et al. Multimodal medical image registration using particle swarm optimization with influence of the data's initial orientation
CN112825619A (en) Training machine learning algorithm using digitally reconstructed radiological images
US20220130524A1 (en) Method and Systems for Predicting a Stream of Virtual Topograms
Bermejo et al. Coral reef optimization for intensity-based medical image registration
Vázquez Romaguera et al. Personalized respiratory motion model using conditional generative networks for MR-guided radiotherapy
US20240115320A1 (en) Automatic planning and guidance of liver tumor thermal ablation using ai agents trained with deep reinforcement learning
Zhou et al. Deep reinforcement learning in medical imaging
Ghesu et al. Learning cardiac anatomy
EP4345830A1 (en) Automatic planning and guidance of liver tumor thermal ablation using ai agents trained with deep reinforcement learning
Zhang Deep Reinforcement Learning in Medical Object Detection and Segmentation
Guo Sensorless frame-to-volume multimodal image fusion via deep learning
Wan et al. Multi-target landmark detection with incomplete images via reinforcement learning and shape prior embedding
Valsecchi Intensity-based medical image registration using metaheuristics

Legal Events

Date Code Title Description
AS Assignment

Owner name: SIEMENS MEDICAL SOLUTIONS USA, INC., PENNSYLVANIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LIAO, RUI;GRBIC, SASA;MIAO, SHUN;AND OTHERS;SIGNING DATES FROM 20170509 TO 20170524;REEL/FRAME:042493/0373

AS Assignment

Owner name: SIEMENS HEALTHCARE GMBH, GERMANY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SIEMENS MEDICAL SOLUTIONS USA, INC.;REEL/FRAME:042656/0787

Effective date: 20170607

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION