WO2024138224A1 - Systems and methods for 3-dimensional anatomical reconstruction using monoscopic photogrammetry - Google Patents

Systems and methods for 3-dimensional anatomical reconstruction using monoscopic photogrammetry Download PDF

Info

Publication number
WO2024138224A1
WO2024138224A1 PCT/US2023/085928 US2023085928W WO2024138224A1 WO 2024138224 A1 WO2024138224 A1 WO 2024138224A1 US 2023085928 W US2023085928 W US 2023085928W WO 2024138224 A1 WO2024138224 A1 WO 2024138224A1
Authority
WO
WIPO (PCT)
Prior art keywords
dimensional
neural network
depth
image data
depth estimation
Prior art date
Application number
PCT/US2023/085928
Other languages
French (fr)
Inventor
Sahin HANALIOGLU
Giancarlo MIGNUCCI
Mark Preul
Nicolas I. GONZALEZ ROMO
Original Assignee
Dignity Health
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Dignity Health filed Critical Dignity Health
Publication of WO2024138224A1 publication Critical patent/WO2024138224A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/20Finite element generation, e.g. wire-frame surface description, tesselation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/04815Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04845Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation, expansion or change of colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H20/00ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance
    • G16H20/40ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance relating to mechanical, radiation or invasive therapies, e.g. surgery, laser therapy, dialysis or acupuncture
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H30/00ICT specially adapted for the handling or processing of medical images
    • G16H30/40ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/20ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/70ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for mining of medical data, e.g. analysing previous cases of other patients
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/10Computer-aided planning, simulation or modelling of surgical operations
    • A61B2034/101Computer-aided simulation of surgical operations
    • A61B2034/105Modelling of the patient, e.g. for ligaments or bones
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • A61B2090/364Correlation of different images or relation of image positions in respect to the body
    • A61B2090/365Correlation of different images or relation of image positions in respect to the body augmented reality, i.e. correlating a live optical image with another image
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • A61B2090/364Correlation of different images or relation of image positions in respect to the body
    • A61B2090/367Correlation of different images or relation of image positions in respect to the body creating a 3D dataset from 2D images using position information
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B2576/00Medical imaging apparatus involving image processing or analysis
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/25User interfaces for surgical systems
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/48Other medical applications
    • A61B5/4887Locating particular structures in or on the body
    • A61B5/4893Nerves
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/72Signal processing specially adapted for physiological signals or for diagnostic purposes
    • A61B5/7235Details of waveform analysis
    • A61B5/7264Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/30Devices for illuminating a surgical field, the devices having an interrelation with other surgical devices or with a surgical procedure
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • A61B90/361Image-producing devices, e.g. surgical cameras
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/90Identification means for patients or instruments, e.g. tags
    • A61B90/94Identification means for patients or instruments, e.g. tags coded with symbols, e.g. text
    • A61B90/96Identification means for patients or instruments, e.g. tags coded with symbols, e.g. text using barcodes
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C11/00Photogrammetry or videogrammetry, e.g. stereogrammetry; Photographic surveying
    • G01C11/02Picture taking arrangements specially adapted for photogrammetry or photographic surveying, e.g. controlling overlapping of pictures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30016Brain
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30101Blood vessel; Artery; Vein; Vascular
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2210/00Indexing scheme for image generation or computer graphics
    • G06T2210/41Medical

Definitions

  • the present disclosure generally relates to anatomical modeling, and in particular, to a system and associated methods for three-dimensional anatomical reconstruction using a monoscopic photogrammetry technique.
  • Immersive anatomical environments offer an alternative when anatomical laboratory access is limited, but current three-dimensional (3D) renderings are not able to simulate the anatomical detail and surgical perspectives needed for microsurgical education.
  • FIG. 1A is a simplified block diagram showing a system for generating 3D anatomical models from 2-D images using monoscopic photogrammetry;
  • FIG. 2 is a simplified diagram showing an exemplary computing system for implementation of the system of FIG. 1A;
  • FIG. 4 is a simplified diagram showing a training sequence for training the neural network
  • FIGS. 6A-6D are a series of images showing an extended reality environment for display of a 3D anatomical model of the system of FIG. 1 A;
  • FIGS. 7A-7D are a series of images showing QR codes for the extended reality environment of the system of FIG. 1A;
  • FIGS. 8A-8I are a series of images showing photographs, depth maps and histograms of a Transsylvian exposure at varying magnification;
  • FIGS. 9A-9I are a series of images showing photographs, depth maps and histograms of a mastoidectomy dissection at varying magnification according to the system of FIG. 1A;
  • FIGS. 10A-10D are a series of images showing high- magnification photographs and their corresponding depth maps, with labeled anatomical ROIs according to the system of FIG. 1A;
  • FIGS. 11A-11 D are a series of images showing a 3D reconstruction procedure according to the system of FIG. 1A.
  • system 100 The present disclosure outlines a system and associated methods for computer-implemented 3D anatomical model reconstruction (hereinafter, system 100) that applies machine learning principles to transform high- resolution photographs into immersive anatomical models, offering a convenient and efficient resource for neurosurgical education.
  • the system 100 applies a 3D reconstruction process (e.g., monoscopic photogrammetry) that transforms two- dimensional (2D) high-resolution anatomical input images into volumetric surface reconstructions.
  • 3D reconstruction process e.g., monoscopic photogrammetry
  • 2D two- dimensional
  • the system 100 includes a device 102 (e.g., a computing device) that implements a neural network 120 operable for monoscopic depth estimation that receives a set of 2D anatomical image data 112 and outputs a set of depth estimation data 114 (e.g., a depth map) for the set of 2D anatomical image data 112 using a monoscopic depth estimation process.
  • the system 100 further implements a 3D reconstruction framework 130 that transforms the set of depth estimation data 114 into a 3D anatomical model 116.
  • the system 100 can include a server 104 operable for communication with an extended reality display device 150 such as a virtual reality display device 152 or a mobile device 154 (e.g., a smartphone), where the server 104 hosts an application programming interface (API) 140 for display of and interaction with the 3D anatomical model 116 at the extended reality display device 150.
  • an extended reality display device 150 such as a virtual reality display device 152 or a mobile device 154 (e.g., a smartphone)
  • API application programming interface
  • This technology allows accurate spatial representation of micro- anatomical structures that are usually not well represented in digital renderings or DICOM reconstructions.
  • the system 100 only requires one high resolution image, instead of multiple images as in traditional photogrammetry. Further, the system 100 performs a spatial correlation analysis using neuronavigation to ensure accuracy of depth estimations.
  • FIGS. 1B-1E illustrate sequential generation of a 3D anatomical model 116 from a set of 2D anatomical image data 112 by the system 100.
  • FIG. 1B shows a set of 2D anatomical image data 112, which is provided as input to the neural network 120.
  • FIG. 1C shows a resultant set of depth estimation data 114 which is an output of the neural network 120 represented visually using a gradient (e.g., in which “lighter” pixels represent “closer” portions of the 2D anatomical image data 112 and “darker” pixels represent “farther” portions of the 2D anatomical image data 112). Note that while FIG.
  • 1C shows one example representation of the set of depth estimation data 114 as an image with a color value of each respective pixel corresponding to a “depth” value
  • the set of depth estimation data 114 can be represented within a comma-separated-variable (.csv) file, a bitmap (.bmp) file, or another suitable file format where each respective pixel in the 2D anatomical image data 112 is associated with a “depth” value.
  • FIGS. 1D and 1E show generation of a 3D anatomical model 116 at the 3D reconstruction framework 130 of the system 100 using the set of depth estimation data 114.
  • FIG. 1 D shows a 3D mesh 115 of the 3D anatomical model 116 generated using the 2D anatomical image data 112 and the set of depth estimation data 114.
  • FIG. 1E shows a final render of the 3D anatomical model 116 including the 3D mesh 115 and also including color information for each pixel within the 2D anatomical image data 112.
  • the 3D anatomical model 116 can be viewed at a display device within an extended reality environment and/or at a standard computing device.
  • a method implemented by system 100 herein includes: accessing the set of 2D anatomical image data 112 and generating the set of depth estimation data 114 based on the set of 2D anatomical image data 112 by a monoscopic photogrammetry procedure.
  • the method can further include generating the 3D mesh 115 of the 3D anatomical model 116 using the set of depth estimation data 114. This step can encompass converting the 3D point cloud (represented by coordinate pairs with X, Y, and Z values) to the 3D mesh 115 of the 3D anatomical model 116. Finally, the method can include generating a final rendering of the 3D anatomical model 116 using the set of 2D anatomical image data 112 and the 3D mesh 115,
  • the method can include communicating the final rendering of the 3D anatomical model 116 to the server 104 hosting the API 140 operable for communication with the extended reality device 150.
  • the extended reality device 150 is the virtual reality display device 152
  • the AP1 140 provides instructions executable by a processor associated with the virtual reality display device 152 to display the final rendering of the 3D anatomical model 116 within a virtual reality environment.
  • the API 140 provides instructions executable by a processor of the mobile device 154 to display the final rendering of the 3D anatomical model 116 within an augmented-reality environment, These instructions can include superimposing the final rendering of the 3D anatomical model 116 over a live feed captured by a camera of the mobile device 154, where the augmented-reality environment is accessible by the mobile device 154 through a quick-response code readable by a camera of the mobile device 154.
  • FIG. 2 is a schematic block diagram of an example device 102 that may be used with one or more embodiments described herein, e.g., as a component of system 100.
  • Device 102 comprises one or more network interfaces 210 (e.g., wired, wireless, PLC, etc.), at least one processor 220, and a memory 240 interconnected by a system bus 250, as well as a power supply 260 (e.g., battery, plug-in, etc.).
  • Device 102 can further include display device 230 in communication with the processor 220 that displays information to a user.
  • Network interface(s) 210 include the mechanical, electrical, and signaling circuitry for communicating data over the communication links coupled to a communication network.
  • Network interfaces 210 are configured to transmit and/or receive data using a variety of different communication protocols. As illustrated, the box representing network interfaces 210 is shown for simplicity, and it is appreciated that such interfaces may represent different types of network connections such as wireless and wired (physical) connections.
  • Network interfaces 210 are shown separately from power supply 260, however it is appreciated that the interfaces that support PLC protocols may communicate through power supply 260 and/or may be an integral component coupled to power supply 260.
  • Memory 240 includes a plurality of storage locations that are addressable by processor 220 and network interfaces 210 for storing software programs and data structures associated with the embodiments described herein.
  • device 102 may have limited memory or no memory (e.g., no memory for storage other than for programs/processes operating on the device and associated caches).
  • Memory 240 can include instructions executable by the processor 220 that, when executed by the processor 220, cause the processor 220 to implement aspects of the system 100 outlined herein.
  • modules or engines may be interchangeable.
  • the term module or engine refers to model or an organization of interrelated software components/functions.
  • monoscopic photogrammetry processes/services 290 is shown as a standalone process, those skilled in the art will appreciate that this process may be executed as a routine or module within other processes.
  • FIG. 3 is a schematic block diagram of an example neural network architecture 300 that may be used with one or more embodiments described herein, e.g., as a component of system 100 shown in FIG. 1A, and particularly as a component of the neural network 120. Possible implementations of the neural network architecture 300 can be used by the system 100 to extract the set of depth estimation data 114 from the 2D anatomical image data 112.
  • Architecture 300 includes the neural network 120 defined by an example neural network description 301 in an engine model (neural controller) 330.
  • the neural network description 301 can include a full specification of the neural network 120, including the neural network architecture 300.
  • the neural network description 301 can include a description or specification of the architecture 300 of the neural network 120 (e.g., the layers, layer interconnections, number of nodes in each layer, etc.); an input and output description which indicates how the input and output are formed or processed; an indication of the activation functions in the neural network, the operations or filters in the neural network, etc.; neural network parameters such as weights, biases, etc.; and so forth.
  • the neural network 120 reflects the architecture 300 defined in the neural network description 301.
  • the neural network 120 includes an input layer 302, which receives input data including the 2D anatomical image data 112 corresponding to one or more nodes 308.
  • the input layer 302 can include data representing a portion of input media data such as a patch of data or pixels in the 2D anatomical image data 112.
  • the neural network 120 includes hidden layers 304A through 304/V (collectively “304” hereinafter).
  • the hidden layers 304 can include n number of hidden layers, where n is an integer greater than or equal to one.
  • the number of hidden layers can include as many layers as needed for a desired processing outcome and/or rendering intent.
  • the neural network 120 further includes an output layer 306 that provides an output (e.g., the set of depth estimation data 114) resulting from the processing performed by the hidden layers 304.
  • the output layer 306 can provide the set of depth estimation data 114 as an image with a color value of each respective pixel corresponding to a “depth” value.
  • the set of depth estimation data 114 can be represented within a comma-separated-variable (.csv) file, a bitmap (.bmp) file, or another suitable file format where each respective pixel of the 2D anatomical image data 112 provided to the input layer 302 is associated with a “depth” value.
  • the neural network 120 in this example is a multi-layer neural network of interconnected nodes. Each node can represent a piece of information. Information associated with the nodes is shared among the different layers and each layer retains information as information is processed.
  • the neural network 120 can include a feed-forward neural network, in which case there are no feedback connections where outputs of the neural network are fed back into itself.
  • the neural network 120 can include a recurrent neural network, which can have loops that allow information to be carried across nodes while reading in input.
  • Information can be exchanged between nodes through node-to- node interconnections between the various layers. Nodes of the input layer 302 can activate a set of nodes in the first hidden layer 304A.
  • each of the input nodes of the input layer 302 is connected to each of the nodes of the first hidden layer 304A.
  • the nodes of the hidden layer 304A can transform the information of each input node by applying activation functions to the information.
  • the information derived from the transformation can then be passed to and can activate the nodes of the next hidden layer (e.g., 304B), which can perform their own designated functions.
  • Example functions include convolutional, up-sampling, data transformation, pooling, and/or any other suitable functions.
  • the output of the hidden layer e.g., 304B
  • the output of the last hidden layer can activate one or more nodes of the output layer 306, at which point an output is provided.
  • nodes e.g., nodes 308A, 308B, 308C
  • a node has a single output and all lines shown as being output from a node represent the same output value.
  • each node or interconnection between nodes can have a weight that is a set of parameters derived from training the neural network 120.
  • an interconnection between nodes can represent a piece of information learned about the interconnected nodes.
  • the interconnection can have a numeric weight that can be tuned (e.g., based on a training dataset), allowing the neural network 120 to be adaptive to inputs and able to learn as more data is processed.
  • the neural network 120 can be pre-trained to process the features from the data in the input layer 302 using the different hidden layers 304 to provide the output through the output layer 306.
  • the neural network 120 can learn to estimate depths from the 2D anatomical image data 112 and can be trained using training data that includes example depth estimations from a training dataset. For instance, training data can be input into the neural network 120, which can be processed by the neural network 120 to generate outputs which can be used to tune one or more aspects of the neural network 120, such as weights, biases, etc.
  • the neural network 120 can adjust weights of nodes using a training process called backpropagation.
  • Backpropagation can include a forward pass, a loss function, a backward pass, and a weight update.
  • the forward pass, loss function, backward pass, and parameter update is performed for one training iteration.
  • the process can be repeated for a certain number of iterations for each set of training media data until the weights of the layers are accurately tuned.
  • the output can include values that do not give preference to any particular class due to the weights being randomly selected at initialization. For example, if the output is a vector with probabilities that the object includes different product(s) and/or different users, the probability value for each of the different product and/or user may be equal or at least very similar (e.g., for ten possible products or users, each class may have a probability value of 0.1). With the initial weights, the neural network 120 is unable to determine low level features and thus cannot make an accurate determination of what the classification of the object might be.
  • a loss function can be used to analyze errors in the output. Any suitable loss function definition can be used.
  • the loss can be high for the first training dataset (e.g., images) since the actual values will be different than the predicted output.
  • the goal of training is to minimize the amount of loss so that the predicted output comports with a target or ideal output.
  • the neural network 120 can perform a backward pass by determining which inputs (weights) most contributed to the loss of the neural network 120 and can adjust the weights so that the loss decreases and is eventually minimized.
  • a derivative of the loss with respect to the weights can be computed to determine the weights that contributed most to the loss of the neural network 120.
  • a weight update can be performed by updating the weights of the filters.
  • the weights can be updated so that they change in the opposite direction of the gradient.
  • a learning rate can be set to any suitable value, with a high learning rate including larger weight updates and a lower value indicating smaller weight updates.
  • the neural network 120 can include any suitable neural or deep learning network.
  • One example includes a convolutional neural network (CNN), which includes an input layer and an output layer, with multiple hidden layers between the input and out layers.
  • the hidden layers of a CNN include a series of convolutional, nonlinear, pooling (for downsampling), and fully connected layers.
  • the neural network 120 can represent any other neural or deep learning network, such as an autoencoder, a deep belief nets (DBNs), and recurrent neural networks (RNNs), etc.
  • DNNs deep belief nets
  • RNNs recurrent neural networks
  • a method includes applying a spatial correlation analysis to the set of (training) depth estimation data produced by the neural network 120 during training with respect to a set of neuronavigation data (e.g., ground truth information 410 that corresponds to training 2D anatomical image data 412 provided as input).
  • the neural network 120 can be trained in a multi-stage process involving different types of images and different types of ground truth data such that the neural network 120 can handle different types of data.
  • FIG. 4 shows a training workflow fortraining the neural network 120. In each stage, a method of training the neural network 120 can include providing training anatomical structure image data 412 to the neural network 120 as input.
  • the neural network 120 can generate a set of training depth estimation data 414 based on the training anatomical structure image data 412.
  • the neural network 120 can be refined, updated, or validated by correlating the set of training depth estimation data 414 with ground truth information 410 including actual empirical depth or linear measurement data.
  • the training anatomical structure image data 414 can include general anatomical structure images that can be taken from microscopy recordings, etc.
  • the correlation procedure for the first stage can use ground truth information 410 in the form of empirical depth or linear measurement data captured with surgical neuronavigation methods.
  • the first stage can allow the neural network 120 to develop a general framework for generating depth estimation data given an anatomical image.
  • the training anatomical structure image data 412 can include high-resolution cadaveric dissection images.
  • the correlation procedure for the second stage can use ground truth information 410 in the form of empirical depth or linear measurement data captured during cadaveric dissection.
  • the second stage can refine the neural network 120 to provide microscopically accurate depth estimations.
  • the training anatomical structure image data 412 can include different open-access microsurgical images taken from anatomy databases, and the correlation procedure for the third stage can use ground truth information 410 in the form of actual data that accompanies the open-access microsurgical images.
  • the third stage can train the neural network 120 to handle different types of data that might not be uniform or consistent, so that the neural network 120 may be more universal.
  • the neural network 120 can compute an inverse depth, allowing training on diverse data (because ground truth data are not always of the same type).
  • inverse depth can be used interchangeably as a surrogate for depth, although it is important to note that the neural network 120 may or may not compute absolute depth. Because every point of the set of depth estimation data 114 can be located in space using specific cartesian coordinates (X, Y, Z), thus creating a 3D point cloud, 3D volumetric reconstructions are possible.
  • the system 100 can include the 3D reconstruction framework 130 including various post-processing modules to generate a 3D anatomical model 116 for 2D anatomical image data 112 based on the set of depth estimation data 114.
  • the 3D reconstruction framework 130 includes Open3D library to create a point cloud from the set of depth estimation data 114, particularly from depth maps obtained by the neural network 120, and further applies additional 3D postprocessing using Meshlab software.
  • FIGS. 5A-5C show an example user interface 500 of the system 100 and images for generation and viewing of the 3D anatomical model 116.
  • FIG. 5A in particular shows 2D anatomical image data 112 and a menu 502 of the user interface 500 for display at the display device (shown as display device 230 in FIG. 2).
  • the menu 502 can include an option to select and upload a reference image (e.g., the 2D anatomical image data 112) and/or a depth image (e.g., a previously obtained set of depth estimation data 114 for 2D anatomical image data 112).
  • a reference image e.g., the 2D anatomical image data 112
  • a depth image e.g., a previously obtained set of depth estimation data 114 for 2D anatomical image data 112
  • the menu 502 can include a “create model” button that, when interacted with, instructs the system 100 to generate a 3D mesh 115 shown in FIG. 5B and consequently, a final rendering of the 3D anatomical model 116 shown in FIG. 5C.
  • the menu 502 can include a “download .obj” button that enables a computing device of the user to download an object file representative of the 3D anatomical model 116.
  • Other options provided at the menu 502 can include “render mode” settings that display the 3D anatomical model 116 in solid form, in point cloud form, and/or in wireframe form, and can also include other display configurations for illustration. Additional settings can include sliders for focal distance, a near plane, a far plane, a mesh smoothness, a quad size, a point size, and downsampling parameters.
  • the system 100 can include the server 104 that hosts the API 140 and is operable for communication with the extended reality device 150 such as virtual reality (VR) display device 152 or mobile device 154.
  • the API 140 can provide instructions executable by a processor associated with the extended reality device 150 to display the final rendering of the 3D anatomical model 116 within an extended reality environment.
  • FIGS. 6A extended reality device 150 - 7D show example environments that can be displayed at the extended reality device 150.
  • the server 104 can communicate with the VR display device 152 (FIG. 1A) such as an Oculus Quest headset (Meta Platforms, Inc., Menlo Park, CA).
  • the VR display device 152 can be connected to a computer station for screen casting and video recording.
  • the one or more 3D anatomical models 116 can be provided to the API 140 which can implement a VR collaboration platform such as Spatial (Spatial Dev, New York, NY) to display the final rendering of the 3D anatomical model 116 within a virtual reality environment at the VR display device 152.
  • a VR collaboration platform such as Spatial (Spatial Dev, New York, NY)
  • the one or more 3D anatomical models 116 can be provided to the API 140 which can implement an augmented-reality environment for display at the mobile device 154 having a camera such as a tablet (e.g., an iPad Pro 11-inch second generation, Apple, Inc., Cupertino, CA) to display the final rendering of the 3D anatomical model 116 within an augmented reality environment.
  • the AP1 140 can instruct the mobile device 154 to superimpose the final rendering of the 3D anatomical model 116 over a live feed captured by a camera of the mobile device 154.
  • the augmented-reality environment can be accessible by the mobile device 154 through a quick-response code readable by a camera of the mobile device 154.
  • the augmented-reality environment can be developed using an open-source JavaScript library such as AR.js, with a quick response (QR)-marker-based detection for deployment using the mobile device 154.
  • FIGS. 6A-7D show example images for representation of the one or more 3D anatomical models 116 within the extended reality environment.
  • FIGS. 6A and 6B illustrate an example display interface at the VR device 152 for display of the one or more 3D anatomical models 116.
  • FIGS. 6C and 6D show example images for display of the one or more 3D anatomical models 116 within the augmented-reality environment;
  • FIG. 6C shows a QR code for scanning by the camera of the mobile device 154 and FIG.
  • FIG. 6D shows resultant display of the one or more 3D anatomical models 116 at the mobile device 154 within the augmented- reality environment.
  • FIGS. 7A-7D further show example QR codes and resultant overlay of the one or more 3D anatomical models 116 at the mobile device 154 within the augmented-reality environment.
  • the 3D methodology of the system 100 produces detailed, realistic 3D anatomical models 116 for display at the VR device 152 (FIGS. 6A and 6B). Head tracking allowed a wearer-visualized perspective for improved recognition of anatomical spatial relations. Using the controllers, the 3D anatomical model 116 displayed at the VR device 152 could be manipulated, including 360° rotation and translation movements. Scale adjustments were possible for higher magnifications. Pial surfaces, arachnoid membranes, cranial nerves, and microvascular structures, including perforators, could be observed in a 3D virtual environment. Multiuser participation in the same VR environment allowed a collaborative educational experience. FIG. 6A shows a visualization of an orbitozygomatic transsylvian approach in virtual reality. FIG. 6B shows the model scale has been increased. The oculomotor nerve, Liliequist membrane, and basilar artery are observed deep in the surgical field.
  • FIG. 6C shows an augmented-reality QR marker is placed on a flat surface for camera detection.
  • FIG. 6D shows an immersive anatomical model of a right extradural approach to the cavernous sinus is displayed. After scanning the QR code, the model was displayed and could be manipulated and rotated in all directions, and the magnification could be increased. Depth perception of anatomical structures could be experienced, albeit in limited fashion, because of knowledge of spatial distribution of the field, even while viewing a screen presenting a 2D image.
  • Augmented reality allowed a side-by-side comparison of virtual models with real models of cranial dissections for improved anatomical learning.
  • the reader can visualize example anatomical models by scanning the QR codes presented in FIGS. 7B and 7D with a mobile device
  • FIG. 7A shows an anatomical model of the T ranssylvian approach incorporated into a QR code augmented-reality experience.
  • FIG. 7B shows an example QR code that can be scanned with mobile devices for visualization of the 3D model of FIG. 7A.
  • FIG. 7C shows an anatomical model of the perilabyrinthine dissection incorporated into a QR code augmented-reality experience.
  • FIG. 7D shows a QR code that can be scanned with mobile devices for visualization of the 3D model of FIG. 7C.
  • the neural network 120 is first trained to obtain and analyze a set of depth estimation data 114 for the 2D anatomical image data 112.
  • the set of depth estimation data 114 is analyzed through a correlation procedure for correlation with empirical depth or linear measurements obtained through surgical neuronavigation methods to ensure that the neural network 120 can make accurate predictions.
  • the correlation procedure is applied to high-resolution images of cadaveric brain dissections to correlate the depth estimation data 114 predicted by the neural network 120 with actual measurements.
  • the correlation procedure can also be applied to samples of 2D anatomical image data 112 from an open-access microsurgical anatomy database (e.g., the Rhoton collection).
  • the system 100 enables development of 3D anatomical models 116 using the set of depth estimation data 114 and following verification of correct correlation with respect to actual data. Finally, the 3D anatomical models 116 are exported for display within an extended reality environment for assessment of depth perception and user experience.
  • stepwise dissection of cadaveric specimens can be performed within a specialized neurosurgical laboratory facility using standard microsurgical equipment.
  • a Zeiss Pentero microscope Carl Zeiss Surgical, Inc., Oberkochen, Germany
  • a Medtronic StealthStation S7 Medtronic, Minneapolis, MN neuronavigation system was used to define coordinates of relevant anatomical targets.
  • acquisition of 2D anatomical image data 112 during stepwise dissection was performed using three different cameras: Zeiss Trenion 3D-HD platform with full HD capability, a Canon Eos 5DSr camera with a 50.6-megapixels sensor and a 100-mm lens, and Sony a6000 camera with a 24- megapixels sensor and 50-mm lens complemented with 26-mm rings.
  • depth allocations by the neural network 120 are examined through a correlation process to verify consistency of the set of depth estimation data 114 with measurements collected with a neuronavigation system (e.g., a Medtronic StealthStation S7 neuronavigation system).
  • a neuronavigation system e.g., a Medtronic StealthStation S7 neuronavigation system.
  • ROIs anatomical regions of interest
  • an inverse correlation is expected between increasing depth of anatomical structures and mean pixel intensity values measured in the depth map obtained by the neural network 120, where every anatomical structure is allocated a value from a range of 0 (purple) to 255 (yellow), corresponding to the deepest and the most superficial point of the image, respectively.
  • Stata software version 13.0 (Stata Corp LLC, College Station, TX) is used for statistical analysis. Correlational analysis is performed using Spearman’s rho. An alpha value of ⁇ 0.05 was considered statistically significant.
  • FIGS. 8A-8I show Transsylvian exposure, right side, at different magnifications.
  • FIG. 9A shows a low-magnification photograph
  • FIG. 8B shows a depth map of the photograph of FIG. 8A
  • FIG. 8C shows a histogram of the photograph of FIG. 8A.
  • FIG. 8D shows an intermediate-magnification photograph
  • FIG. 8B shows a depth map of the photograph of FIG. 8D
  • FIG. 8C shows a histogram of the photograph of FIG. 8D.
  • FIG. 8G shows a high-magnification photograph
  • FIG. 8H shows a depth map of the photograph of FIG. 8G
  • FIG. 81 shows a histogram of the photograph of FIG. 8G, respectively.
  • FIGS. 9A-9I show a mastoidectomy dissection, the center of the image corresponding to an intermediate region in terms of depth (semicircular canals).
  • the model assigned pixels in the range of 100 to 200 with increasing magnification, meaning the model allocated to an intermediate depth from the most superficial ROI (spine of Henle) and the deepest point Qugular bulb).
  • FIGS. 9A-9I show Mastoidectomy, left side, at different magnifications.
  • FIG. 9A shows a low-magnification photograph
  • FIG. 9B shows a depth map of the photograph of FIG. 9A
  • FIG. 9C shows a histogram of the photograph of FIG. 9A.
  • FIG. 9D shows an intermediate-magnification photograph
  • FIG. 9B shows a depth map of the photograph of FIG. 9D
  • FIG. 9C shows a histogram of the photograph of FIG. 9D.
  • FIG. 9G shows a high-magnification photograph
  • FIG. 9H shows a depth map of the photograph of FIG. 9G
  • FIG. 9I shows a histogram of the photograph of FIG. 9G, respectively.
  • FIGS. 10A-10D show high-magnification images with corresponding depth maps.
  • FIG. 10A shows a transsylvian exposure photograph and
  • FIG. 10C shows a mastoidectomy exposure photograph and FIG.
  • 10D shows a depth map with overlay of anatomical regions of interest used to calculate mean pixel intensity values.
  • Quantitative spatial discrimination between the optic nerve (mean ⁇ SD pixel intensity: 77.71 ⁇ 0.65), supraclinoid internal carotid artery (69.91 ⁇ 0.33) and the basilar apex (58.98 ⁇ 0.47) is observed in the transsylvian exposure (Table 1 , FIGS. 10A and 10B).
  • the incus (128.1 ⁇ 1 .0), lateral semicircular canal (131 .4 ⁇ 0.7), and vertical segment of the facial nerve (130.6 ⁇ 0.4) were assigned similar mean pixel intensity values with a mastoidectomy exposure, consistent with their known anatomical relations and positions (FIGS. 10C and 10D, Table 2).
  • ICA internal carotid artery
  • MCA middle cerebral artery
  • REZ root entry zone
  • ROI region of interest.
  • Facial nerve vertical 29.0 77.5 + 0.5 75.7 + 0.7 130.6 + 0.4 segment
  • FIGS. 11A-11D show a 3D reconstruction procedure.
  • FIG. 11 A shows a microsurgical image provided as input to the neural network 120
  • FIG. 11B shows a depth map generated by the neural network 120
  • FIG. 11C shows a mesh surface reconstruction generated by the 3D reconstruction framework 130
  • FIG. 11D shows a final immersive anatomical model generated by the 3D reconstruction framework 130 from the mesh surface reconstruction.
  • Extended reality simulators for neurosurgery training have been in development for more than 20 years, and their use has increased in many centers worldwide.
  • DICOM data from neuroimaging studies 3D rendering of anatomical models is possible with accurate spatial localization of structures.
  • these 3D renderings may be cartoonish or artificial and often fail to emulate essential features such as textures, colors, and lighting of the exposed neurovascular structures that must be identified when performing complex cranial approaches.
  • the present disclosure establishes that machine learning can deliver plausible anatomical depth estimations for specific sets of ROIs using only one image as input. This does not mean that depth allocations will be always accurate; it is necessary to perform a cross-check against neuronavigation or DICOM (Digital Imaging and Communications in Medicine) data to verify the accuracy of estimations.
  • DICOM Digital Imaging and Communications in Medicine
  • the system 100 can be applied to existing surgical image compendiums to enhance education.
  • the Rhoton Collection with its lectures and photographs of anatomical dissections using 3D stereoscopy that facilitate understanding of spatial relationships between complex structures, was used as a prime example of a compendium of images to which the system 100 could offer added value, allowing conversion of classic anatomical dissections into conveniently accessible, viewable, immersive, 3D, realistic environments.
  • Further development of anatomy simulators based on other surgical dissection repositories and real-time surgery is an exciting possibility with this technology.
  • Collaborative VR anatomical training with digital performance metrics and artificial intelligence integration for surgical performance classification could augment cadaver dissection in the future, even remotely.
  • Anatomical depth estimations calculated by pretrained models are relative estimations and are subject to differences with absolute spatial measurement. Further training and fine-tuning of the neural network 120 (and potential inclusion of additional or alternative neural networks) with a large data set of microsurgical images with “ground truth” information should increase the accuracy of estimations.
  • individual coordinates co-registered from the neuronavigation system might not reflect in vivo reality due to the nature of cadaveric specimens and the natural phenomenon of brain shifting.
  • the system 100 implements machine learning principles for depth estimation of anatomical structures in microsurgical approaches using only one image as input.
  • the methodologies discussed herein are inherently different from other forms of photogrammetry, which require multiple images, different angles, and prior knowledge of the structures.
  • the system 100 and associated methodologies allow creation of immersive, realistic, anatomical models useful for augmenting neuroanatomy education. Such a process could be especially convenient where either the availability of cadaveric specimens is limited or microsurgical dissection laboratories are absent, or merely when enhanced comprehension of the spatial relation of anatomical structures is desired.

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Public Health (AREA)
  • Medical Informatics (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Primary Health Care (AREA)
  • Epidemiology (AREA)
  • Data Mining & Analysis (AREA)
  • Biomedical Technology (AREA)
  • Software Systems (AREA)
  • Human Computer Interaction (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Databases & Information Systems (AREA)
  • Pathology (AREA)
  • Urology & Nephrology (AREA)
  • Radiology & Medical Imaging (AREA)
  • Surgery (AREA)
  • Computer Graphics (AREA)
  • Geometry (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Image Processing (AREA)

Abstract

A system applies a monoscopic photogrammetry process that transforms 2D high resolution anatomical images into volumetric surface reconstructions. The system includes a neural network capable of monoscopic depth estimation to generate a set of depth estimation data from a 2D anatomical image. The system then transforms the depth map into a navigable 3D model for display within an extended reality environment.

Description

SYSTEMS AND METHODS FOR 3-DIMENSIONAL ANATOMICAL RECONSTRUCTION USING MONOSCOPIC PHOTOGRAMMETRY
CROSS-REFERENCE TO RELATED APPLICATIONS
[0001] This is a PCT application that claims benefit to U.S. Provisional Application Serial no. 63/477,185 filed 24 December 2022, which is incorporated by reference in its entirety.
FIELD
[0002] The present disclosure generally relates to anatomical modeling, and in particular, to a system and associated methods for three-dimensional anatomical reconstruction using a monoscopic photogrammetry technique.
BACKGROUND
[0003] Immersive anatomical environments offer an alternative when anatomical laboratory access is limited, but current three-dimensional (3D) renderings are not able to simulate the anatomical detail and surgical perspectives needed for microsurgical education.
[0004] It is with these observations in mind, among others, that various aspects of the present disclosure were conceived and developed.
BRIEF DESCRIPTION OF THE DRAWINGS
[0005] FIG. 1A is a simplified block diagram showing a system for generating 3D anatomical models from 2-D images using monoscopic photogrammetry;
[0006] FIGS. 1B-1 E are a series of images showing sequential generation of a 3D anatomical model from a 2-D image by the system of FIG. 1 A;
[0007] FIG. 2 is a simplified diagram showing an exemplary computing system for implementation of the system of FIG. 1A;
[0008] FIG. 3 is a simplified diagram showing an example neural network architecture model for implementation of the system of FIG. 1 A;
[0009] FIG. 4 is a simplified diagram showing a training sequence for training the neural network
[0010] FIGS. 5A-5C are a series of images showing an example user interface of the system of FIG. 1 A and sequential generation of a 3D anatomical model;
[0011] FIGS. 6A-6D are a series of images showing an extended reality environment for display of a 3D anatomical model of the system of FIG. 1 A;
[0012] FIGS. 7A-7D are a series of images showing QR codes for the extended reality environment of the system of FIG. 1A;
[0013] FIGS. 8A-8I are a series of images showing photographs, depth maps and histograms of a Transsylvian exposure at varying magnification;
[0014] FIGS. 9A-9I are a series of images showing photographs, depth maps and histograms of a mastoidectomy dissection at varying magnification according to the system of FIG. 1A;
[0015] FIGS. 10A-10D are a series of images showing high- magnification photographs and their corresponding depth maps, with labeled anatomical ROIs according to the system of FIG. 1A; and
[0016] FIGS. 11A-11 D are a series of images showing a 3D reconstruction procedure according to the system of FIG. 1A.
[0017] Corresponding reference characters indicate corresponding elements among the view of the drawings. The headings used in the figures do not limit the scope of the claims. DETAILED DESCRIPTION
7. Introduction
[0018] Advances in artificial intelligence-based object detection and three-dimensional (3D) reconstruction have become fundamental for modern-day applications, such as self-driving vehicles, military aircraft and vehicle training, and face recognition software. Ongoing research has expanded the applications of this technology in neurosurgery, where new strategies have led to an evolution in the neurosurgical education paradigm. Trainees can now supplement their anatomical exposure with online platforms, and microsurgical libraries that contain high- resolution photographs of dissected anatomy.
[0019] For microsurgical anatomy, machine learning could offer new insights into surgical exposure analysis through a computer science perspective that augments the learning of anatomy with novel 3D reconstruction techniques. Traditional photogrammetry allows conversion of still pictures into 3D objects through the acquisition of multiple images from different angles of the subject to create a point cloud for reconstruction. Photogrammetry, traditionally performed with multiple images of an object obtained at different angles, is one such computer-aided image reconstruction procedure in which 3D information is extracted from photographs.
2. System Overview
[0020] The present disclosure outlines a system and associated methods for computer-implemented 3D anatomical model reconstruction (hereinafter, system 100) that applies machine learning principles to transform high- resolution photographs into immersive anatomical models, offering a convenient and efficient resource for neurosurgical education. The system 100 applies a 3D reconstruction process (e.g., monoscopic photogrammetry) that transforms two- dimensional (2D) high-resolution anatomical input images into volumetric surface reconstructions. First, the system 100 applies an anatomical image as input to a neural network capable of monoscopic depth estimation to create a depth map. The system 100 can then apply a 3D reconstruction process to the depth map to transform the anatomical image into navigable 3D models. In one implementation, with reference to FIG. 1A, the system 100 includes a device 102 (e.g., a computing device) that implements a neural network 120 operable for monoscopic depth estimation that receives a set of 2D anatomical image data 112 and outputs a set of depth estimation data 114 (e.g., a depth map) for the set of 2D anatomical image data 112 using a monoscopic depth estimation process. The system 100 further implements a 3D reconstruction framework 130 that transforms the set of depth estimation data 114 into a 3D anatomical model 116. In a further aspect, the system 100 can include a server 104 operable for communication with an extended reality display device 150 such as a virtual reality display device 152 or a mobile device 154 (e.g., a smartphone), where the server 104 hosts an application programming interface (API) 140 for display of and interaction with the 3D anatomical model 116 at the extended reality display device 150.
[0021] This technology allows accurate spatial representation of micro- anatomical structures that are usually not well represented in digital renderings or DICOM reconstructions. The system 100 only requires one high resolution image, instead of multiple images as in traditional photogrammetry. Further, the system 100 performs a spatial correlation analysis using neuronavigation to ensure accuracy of depth estimations.
[0022] FIGS. 1B-1E illustrate sequential generation of a 3D anatomical model 116 from a set of 2D anatomical image data 112 by the system 100. FIG. 1B shows a set of 2D anatomical image data 112, which is provided as input to the neural network 120. FIG. 1C shows a resultant set of depth estimation data 114 which is an output of the neural network 120 represented visually using a gradient (e.g., in which “lighter” pixels represent “closer” portions of the 2D anatomical image data 112 and “darker” pixels represent “farther” portions of the 2D anatomical image data 112). Note that while FIG. 1C shows one example representation of the set of depth estimation data 114 as an image with a color value of each respective pixel corresponding to a “depth” value, the set of depth estimation data 114 can be represented within a comma-separated-variable (.csv) file, a bitmap (.bmp) file, or another suitable file format where each respective pixel in the 2D anatomical image data 112 is associated with a “depth” value.
[0023] FIGS. 1D and 1E show generation of a 3D anatomical model 116 at the 3D reconstruction framework 130 of the system 100 using the set of depth estimation data 114. In particular, FIG. 1 D shows a 3D mesh 115 of the 3D anatomical model 116 generated using the 2D anatomical image data 112 and the set of depth estimation data 114. Finally, FIG. 1E shows a final render of the 3D anatomical model 116 including the 3D mesh 115 and also including color information for each pixel within the 2D anatomical image data 112. The 3D anatomical model 116 can be viewed at a display device within an extended reality environment and/or at a standard computing device.
[0024] A method implemented by system 100 herein includes: accessing the set of 2D anatomical image data 112 and generating the set of depth estimation data 114 based on the set of 2D anatomical image data 112 by a monoscopic photogrammetry procedure.
[0025] Generating the set of depth estimation data 114 can be performed by the neural network 120, the neural network 120 accepting the set of 2D anatomical image data 112 as input and being trained to output the set of depth estimation data 114 that correlates with the set of 2D anatomical image data 112. In some examples, the neural network 120 can compute an inverse depth value (e.g., depth or “Z” values) for each 2D coordinate pair of a plurality of 2D coordinate pairs (e.g., horizontal and vertical or “X and Y” values) of the set of 2D anatomical image data 112, and can generate a 3D point cloud of the set of depth estimation data 114 by combination of each 2D coordinate pair and the inverse depth value for the plurality of 2D coordinate pairs of the set of 2D anatomical image data 112.
[0026] The method can further include generating the 3D mesh 115 of the 3D anatomical model 116 using the set of depth estimation data 114. This step can encompass converting the 3D point cloud (represented by coordinate pairs with X, Y, and Z values) to the 3D mesh 115 of the 3D anatomical model 116. Finally, the method can include generating a final rendering of the 3D anatomical model 116 using the set of 2D anatomical image data 112 and the 3D mesh 115,
[0027] In a further aspect, the method can include communicating the final rendering of the 3D anatomical model 116 to the server 104 hosting the API 140 operable for communication with the extended reality device 150. When the extended reality device 150 is the virtual reality display device 152, the AP1 140 provides instructions executable by a processor associated with the virtual reality display device 152 to display the final rendering of the 3D anatomical model 116 within a virtual reality environment. When the extended reality device is the mobile device 154, the API 140 provides instructions executable by a processor of the mobile device 154 to display the final rendering of the 3D anatomical model 116 within an augmented-reality environment, These instructions can include superimposing the final rendering of the 3D anatomical model 116 over a live feed captured by a camera of the mobile device 154, where the augmented-reality environment is accessible by the mobile device 154 through a quick-response code readable by a camera of the mobile device 154.
2.1 Computing Device
[0028] FIG. 2 is a schematic block diagram of an example device 102 that may be used with one or more embodiments described herein, e.g., as a component of system 100.
[0029] Device 102 comprises one or more network interfaces 210 (e.g., wired, wireless, PLC, etc.), at least one processor 220, and a memory 240 interconnected by a system bus 250, as well as a power supply 260 (e.g., battery, plug-in, etc.). Device 102 can further include display device 230 in communication with the processor 220 that displays information to a user.
[0030] Network interface(s) 210 include the mechanical, electrical, and signaling circuitry for communicating data over the communication links coupled to a communication network. Network interfaces 210 are configured to transmit and/or receive data using a variety of different communication protocols. As illustrated, the box representing network interfaces 210 is shown for simplicity, and it is appreciated that such interfaces may represent different types of network connections such as wireless and wired (physical) connections. Network interfaces 210 are shown separately from power supply 260, however it is appreciated that the interfaces that support PLC protocols may communicate through power supply 260 and/or may be an integral component coupled to power supply 260.
[0031] Memory 240 includes a plurality of storage locations that are addressable by processor 220 and network interfaces 210 for storing software programs and data structures associated with the embodiments described herein. In some embodiments, device 102 may have limited memory or no memory (e.g., no memory for storage other than for programs/processes operating on the device and associated caches). Memory 240 can include instructions executable by the processor 220 that, when executed by the processor 220, cause the processor 220 to implement aspects of the system 100 outlined herein.
[0032] Processor 220 comprises hardware elements or logic adapted to execute the software programs (e.g., instructions) and manipulate data structures 245. An operating system 242, portions of which are typically resident in memory 240 and executed by the processor, functionally organizes device 102 by, inter alia, invoking operations in support of software processes and/or services executing on the device. These software processes and/or services may include monoscopic photogrammetry processes/services 290 that implement aspects of the system 100 and associated methods described herein. Note that while monoscopic photogrammetry processes/services 290 is illustrated in centralized memory 240, alternative embodiments provide for the process to be operated within the network interfaces 210, such as a component of a MAC layer, and/or as part of a distributed computing network environment.
[0033] It will be apparent to those skilled in the art that other processor and memory types, including various computer-readable media, may be used to store and execute program instructions pertaining to the techniques described herein. Also, while the description illustrates various processes, it is expressly contemplated that various processes may be embodied as modules or engines configured to operate in accordance with the techniques herein (e.g., according to the functionality of a similar process). In this context, the term module and engine may be interchangeable. In general, the term module or engine refers to model or an organization of interrelated software components/functions. Further, while the monoscopic photogrammetry processes/services 290 is shown as a standalone process, those skilled in the art will appreciate that this process may be executed as a routine or module within other processes.
2.2 Neural Network
[0034] FIG. 3 is a schematic block diagram of an example neural network architecture 300 that may be used with one or more embodiments described herein, e.g., as a component of system 100 shown in FIG. 1A, and particularly as a component of the neural network 120. Possible implementations of the neural network architecture 300 can be used by the system 100 to extract the set of depth estimation data 114 from the 2D anatomical image data 112.
[0035] Architecture 300 includes the neural network 120 defined by an example neural network description 301 in an engine model (neural controller) 330. The neural network description 301 can include a full specification of the neural network 120, including the neural network architecture 300. For example, the neural network description 301 can include a description or specification of the architecture 300 of the neural network 120 (e.g., the layers, layer interconnections, number of nodes in each layer, etc.); an input and output description which indicates how the input and output are formed or processed; an indication of the activation functions in the neural network, the operations or filters in the neural network, etc.; neural network parameters such as weights, biases, etc.; and so forth.
[0036] The neural network 120 reflects the architecture 300 defined in the neural network description 301. The neural network 120 includes an input layer 302, which receives input data including the 2D anatomical image data 112 corresponding to one or more nodes 308. In one illustrative example, the input layer 302 can include data representing a portion of input media data such as a patch of data or pixels in the 2D anatomical image data 112.
[0037] The neural network 120 includes hidden layers 304A through 304/V (collectively “304” hereinafter). The hidden layers 304 can include n number of hidden layers, where n is an integer greater than or equal to one. The number of hidden layers can include as many layers as needed for a desired processing outcome and/or rendering intent. The neural network 120 further includes an output layer 306 that provides an output (e.g., the set of depth estimation data 114) resulting from the processing performed by the hidden layers 304. In an illustrative example, the output layer 306 can provide the set of depth estimation data 114 as an image with a color value of each respective pixel corresponding to a “depth” value. As discussed above, the set of depth estimation data 114 can be represented within a comma-separated-variable (.csv) file, a bitmap (.bmp) file, or another suitable file format where each respective pixel of the 2D anatomical image data 112 provided to the input layer 302 is associated with a “depth” value.
[0038] The neural network 120 in this example is a multi-layer neural network of interconnected nodes. Each node can represent a piece of information. Information associated with the nodes is shared among the different layers and each layer retains information as information is processed. In some cases, the neural network 120 can include a feed-forward neural network, in which case there are no feedback connections where outputs of the neural network are fed back into itself. In other cases, the neural network 120 can include a recurrent neural network, which can have loops that allow information to be carried across nodes while reading in input. [0039] Information can be exchanged between nodes through node-to- node interconnections between the various layers. Nodes of the input layer 302 can activate a set of nodes in the first hidden layer 304A. For example, as shown, each of the input nodes of the input layer 302 is connected to each of the nodes of the first hidden layer 304A. The nodes of the hidden layer 304A can transform the information of each input node by applying activation functions to the information. The information derived from the transformation can then be passed to and can activate the nodes of the next hidden layer (e.g., 304B), which can perform their own designated functions. Example functions include convolutional, up-sampling, data transformation, pooling, and/or any other suitable functions. The output of the hidden layer (e.g., 304B) can then activate nodes of the next hidden layer (e.g., 304N), and so on. The output of the last hidden layer can activate one or more nodes of the output layer 306, at which point an output is provided. In some cases, while nodes (e.g., nodes 308A, 308B, 308C) in the neural network 120 are shown as having multiple output lines, a node has a single output and all lines shown as being output from a node represent the same output value.
[0040] In some cases, each node or interconnection between nodes can have a weight that is a set of parameters derived from training the neural network 120. For example, an interconnection between nodes can represent a piece of information learned about the interconnected nodes. The interconnection can have a numeric weight that can be tuned (e.g., based on a training dataset), allowing the neural network 120 to be adaptive to inputs and able to learn as more data is processed.
[0041] The neural network 120 can be pre-trained to process the features from the data in the input layer 302 using the different hidden layers 304 to provide the output through the output layer 306. The neural network 120 can learn to estimate depths from the 2D anatomical image data 112 and can be trained using training data that includes example depth estimations from a training dataset. For instance, training data can be input into the neural network 120, which can be processed by the neural network 120 to generate outputs which can be used to tune one or more aspects of the neural network 120, such as weights, biases, etc.
[0042] In some cases, the neural network 120 can adjust weights of nodes using a training process called backpropagation. Backpropagation can include a forward pass, a loss function, a backward pass, and a weight update. The forward pass, loss function, backward pass, and parameter update is performed for one training iteration. The process can be repeated for a certain number of iterations for each set of training media data until the weights of the layers are accurately tuned.
[0043] For a first training iteration for the neural network 120, the output can include values that do not give preference to any particular class due to the weights being randomly selected at initialization. For example, if the output is a vector with probabilities that the object includes different product(s) and/or different users, the probability value for each of the different product and/or user may be equal or at least very similar (e.g., for ten possible products or users, each class may have a probability value of 0.1). With the initial weights, the neural network 120 is unable to determine low level features and thus cannot make an accurate determination of what the classification of the object might be. A loss function can be used to analyze errors in the output. Any suitable loss function definition can be used.
[0044] The loss (or error) can be high for the first training dataset (e.g., images) since the actual values will be different than the predicted output. The goal of training is to minimize the amount of loss so that the predicted output comports with a target or ideal output. The neural network 120 can perform a backward pass by determining which inputs (weights) most contributed to the loss of the neural network 120 and can adjust the weights so that the loss decreases and is eventually minimized.
[0045] A derivative of the loss with respect to the weights can be computed to determine the weights that contributed most to the loss of the neural network 120. After the derivative is computed, a weight update can be performed by updating the weights of the filters. For example, the weights can be updated so that they change in the opposite direction of the gradient. A learning rate can be set to any suitable value, with a high learning rate including larger weight updates and a lower value indicating smaller weight updates.
[0046] The neural network 120 can include any suitable neural or deep learning network. One example includes a convolutional neural network (CNN), which includes an input layer and an output layer, with multiple hidden layers between the input and out layers. The hidden layers of a CNN include a series of convolutional, nonlinear, pooling (for downsampling), and fully connected layers. In other examples, the neural network 120 can represent any other neural or deep learning network, such as an autoencoder, a deep belief nets (DBNs), and recurrent neural networks (RNNs), etc.
2.3 Training the Neural Network
[0047] To train the neural network 120, a method includes applying a spatial correlation analysis to the set of (training) depth estimation data produced by the neural network 120 during training with respect to a set of neuronavigation data (e.g., ground truth information 410 that corresponds to training 2D anatomical image data 412 provided as input). In some examples, the neural network 120 can be trained in a multi-stage process involving different types of images and different types of ground truth data such that the neural network 120 can handle different types of data. FIG. 4 shows a training workflow fortraining the neural network 120. In each stage, a method of training the neural network 120 can include providing training anatomical structure image data 412 to the neural network 120 as input. The neural network 120 can generate a set of training depth estimation data 414 based on the training anatomical structure image data 412. The neural network 120 can be refined, updated, or validated by correlating the set of training depth estimation data 414 with ground truth information 410 including actual empirical depth or linear measurement data.
[0048] In the first stage, the training anatomical structure image data 414 can include general anatomical structure images that can be taken from microscopy recordings, etc. The correlation procedure for the first stage can use ground truth information 410 in the form of empirical depth or linear measurement data captured with surgical neuronavigation methods. The first stage can allow the neural network 120 to develop a general framework for generating depth estimation data given an anatomical image. In the second stage, the training anatomical structure image data 412 can include high-resolution cadaveric dissection images. The correlation procedure for the second stage can use ground truth information 410 in the form of empirical depth or linear measurement data captured during cadaveric dissection. The second stage can refine the neural network 120 to provide microscopically accurate depth estimations. In the third stage, the training anatomical structure image data 412 can include different open-access microsurgical images taken from anatomy databases, and the correlation procedure for the third stage can use ground truth information 410 in the form of actual data that accompanies the open-access microsurgical images. The third stage can train the neural network 120 to handle different types of data that might not be uniform or consistent, so that the neural network 120 may be more universal.
2.4 Example Implementation
[0049] The neural network 120 can be pretrained for depth estimation using images and video data sets, including 3D movies. One such example is the Intel ISL MiDaS (Intel, Santa Clara, CA) model, which has zero cross-data set transfer, meaning performance in data sets not used during the training process. Inputs to the neural network 120 can include 2D anatomical image data 112, and the output can include the set of depth estimation data 114 for the 2D anatomical image data 112. In one aspect, the neural network 120 exports a depth map expressive of the set of depth estimation data 114 for the 2D anatomical image data 112. A depth map is an image where every pixel of the original image is assigned an intensity value according to its location on the Z-axis. From a technical perspective, the neural network 120 (e.g., MiDaS) can compute an inverse depth, allowing training on diverse data (because ground truth data are not always of the same type). For the purposes of this disclosure, inverse depth can be used interchangeably as a surrogate for depth, although it is important to note that the neural network 120 may or may not compute absolute depth. Because every point of the set of depth estimation data 114 can be located in space using specific cartesian coordinates (X, Y, Z), thus creating a 3D point cloud, 3D volumetric reconstructions are possible. Further, the system 100 can include the 3D reconstruction framework 130 including various post-processing modules to generate a 3D anatomical model 116 for 2D anatomical image data 112 based on the set of depth estimation data 114. In one non-limiting example, the 3D reconstruction framework 130 includes Open3D library to create a point cloud from the set of depth estimation data 114, particularly from depth maps obtained by the neural network 120, and further applies additional 3D postprocessing using Meshlab software.
2.5 User Interface
[0050] FIGS. 5A-5C show an example user interface 500 of the system 100 and images for generation and viewing of the 3D anatomical model 116. FIG. 5A in particular shows 2D anatomical image data 112 and a menu 502 of the user interface 500 for display at the display device (shown as display device 230 in FIG. 2). In the example shown, the menu 502 can include an option to select and upload a reference image (e.g., the 2D anatomical image data 112) and/or a depth image (e.g., a previously obtained set of depth estimation data 114 for 2D anatomical image data 112). The menu 502 can include a “create model” button that, when interacted with, instructs the system 100 to generate a 3D mesh 115 shown in FIG. 5B and consequently, a final rendering of the 3D anatomical model 116 shown in FIG. 5C. The menu 502 can include a “download .obj” button that enables a computing device of the user to download an object file representative of the 3D anatomical model 116. Other options provided at the menu 502 can include “render mode” settings that display the 3D anatomical model 116 in solid form, in point cloud form, and/or in wireframe form, and can also include other display configurations for illustration. Additional settings can include sliders for focal distance, a near plane, a far plane, a mesh smoothness, a quad size, a point size, and downsampling parameters.
2.6 Extended Reality
[0051] In a further aspect, the system 100 can include the server 104 that hosts the API 140 and is operable for communication with the extended reality device 150 such as virtual reality (VR) display device 152 or mobile device 154. The API 140 can provide instructions executable by a processor associated with the extended reality device 150 to display the final rendering of the 3D anatomical model 116 within an extended reality environment. FIGS. 6A extended reality device 150 - 7D show example environments that can be displayed at the extended reality device 150.
[0052] In one example implementation, the server 104 can communicate with the VR display device 152 (FIG. 1A) such as an Oculus Quest headset (Meta Platforms, Inc., Menlo Park, CA). The VR display device 152 can be connected to a computer station for screen casting and video recording. For this implementation, the one or more 3D anatomical models 116 can be provided to the API 140 which can implement a VR collaboration platform such as Spatial (Spatial Dev, New York, NY) to display the final rendering of the 3D anatomical model 116 within a virtual reality environment at the VR display device 152. [0053] In another example, the one or more 3D anatomical models 116 can be provided to the API 140 which can implement an augmented-reality environment for display at the mobile device 154 having a camera such as a tablet (e.g., an iPad Pro 11-inch second generation, Apple, Inc., Cupertino, CA) to display the final rendering of the 3D anatomical model 116 within an augmented reality environment. The AP1 140 can instruct the mobile device 154 to superimpose the final rendering of the 3D anatomical model 116 over a live feed captured by a camera of the mobile device 154. The augmented-reality environment can be accessible by the mobile device 154 through a quick-response code readable by a camera of the mobile device 154.
[0054] In one such implementation, the augmented-reality environment can be developed using an open-source JavaScript library such as AR.js, with a quick response (QR)-marker-based detection for deployment using the mobile device 154. FIGS. 6A-7D show example images for representation of the one or more 3D anatomical models 116 within the extended reality environment. In particular, FIGS. 6A and 6B illustrate an example display interface at the VR device 152 for display of the one or more 3D anatomical models 116. FIGS. 6C and 6D show example images for display of the one or more 3D anatomical models 116 within the augmented-reality environment; FIG. 6C shows a QR code for scanning by the camera of the mobile device 154 and FIG. 6D shows resultant display of the one or more 3D anatomical models 116 at the mobile device 154 within the augmented- reality environment. FIGS. 7A-7D further show example QR codes and resultant overlay of the one or more 3D anatomical models 116 at the mobile device 154 within the augmented-reality environment.
[0055] The 3D methodology of the system 100 produces detailed, realistic 3D anatomical models 116 for display at the VR device 152 (FIGS. 6A and 6B). Head tracking allowed a wearer-visualized perspective for improved recognition of anatomical spatial relations. Using the controllers, the 3D anatomical model 116 displayed at the VR device 152 could be manipulated, including 360° rotation and translation movements. Scale adjustments were possible for higher magnifications. Pial surfaces, arachnoid membranes, cranial nerves, and microvascular structures, including perforators, could be observed in a 3D virtual environment. Multiuser participation in the same VR environment allowed a collaborative educational experience. FIG. 6A shows a visualization of an orbitozygomatic transsylvian approach in virtual reality. FIG. 6B shows the model scale has been increased. The oculomotor nerve, Liliequist membrane, and basilar artery are observed deep in the surgical field.
[0056] Augmented reality was assessed with the mobile device 154 (FIGS. 6C and 6D, see also FIGS. 7A-7D). FIG. 6C shows an augmented-reality QR marker is placed on a flat surface for camera detection. FIG. 6D shows an immersive anatomical model of a right extradural approach to the cavernous sinus is displayed. After scanning the QR code, the model was displayed and could be manipulated and rotated in all directions, and the magnification could be increased. Depth perception of anatomical structures could be experienced, albeit in limited fashion, because of knowledge of spatial distribution of the field, even while viewing a screen presenting a 2D image. Augmented reality allowed a side-by-side comparison of virtual models with real models of cranial dissections for improved anatomical learning. The reader can visualize example anatomical models by scanning the QR codes presented in FIGS. 7B and 7D with a mobile device
[0057] FIG. 7A shows an anatomical model of the T ranssylvian approach incorporated into a QR code augmented-reality experience. FIG. 7B shows an example QR code that can be scanned with mobile devices for visualization of the 3D model of FIG. 7A. FIG. 7C shows an anatomical model of the perilabyrinthine dissection incorporated into a QR code augmented-reality experience. FIG. 7D shows a QR code that can be scanned with mobile devices for visualization of the 3D model of FIG. 7C.
3. Materials and Methods
[0058] In a first stage of development of one implementation of the system 100, the neural network 120 is first trained to obtain and analyze a set of depth estimation data 114 for the 2D anatomical image data 112. The set of depth estimation data 114 is analyzed through a correlation procedure for correlation with empirical depth or linear measurements obtained through surgical neuronavigation methods to ensure that the neural network 120 can make accurate predictions. In a second stage of development and testing of the system 100, the correlation procedure is applied to high-resolution images of cadaveric brain dissections to correlate the depth estimation data 114 predicted by the neural network 120 with actual measurements. The correlation procedure can also be applied to samples of 2D anatomical image data 112 from an open-access microsurgical anatomy database (e.g., the Rhoton collection). The system 100 enables development of 3D anatomical models 116 using the set of depth estimation data 114 and following verification of correct correlation with respect to actual data. Finally, the 3D anatomical models 116 are exported for display within an extended reality environment for assessment of depth perception and user experience.
3.1 Anatomical Study and Photograph Acquisition
[0059] To obtain the 2D anatomical image data 112, stepwise dissection of cadaveric specimens can be performed within a specialized neurosurgical laboratory facility using standard microsurgical equipment. In one specific implementation, a Zeiss Pentero microscope (Carl Zeiss Surgical, Inc., Oberkochen, Germany) was used for localized magnification of anatomical structures for fine dissection, and a Medtronic StealthStation S7 (Medtronic, Minneapolis, MN) neuronavigation system was used to define coordinates of relevant anatomical targets.
[0060] In one implementation, acquisition of 2D anatomical image data 112 during stepwise dissection was performed using three different cameras: Zeiss Trenion 3D-HD platform with full HD capability, a Canon Eos 5DSr camera with a 50.6-megapixels sensor and a 100-mm lens, and Sony a6000 camera with a 24- megapixels sensor and 50-mm lens complemented with 26-mm rings.
3.2 Surgical Technique
[0061] Formalin-preserved cadaver heads with artery and vein systems injected with silicone were used for dissection. Using an orbitozygomatic approach, pretemporal transcavernous and the transsylvian corridors were dissected, with exposure of the basilar artery as the deepest anatomical structure within the surgical field. A mastoidectomy with skeletonization of the semicircular canals and fallopian canal was also performed. The jugular bulb was exposed at the infralabyrinthine compartment as the deepest structure of this corridor.
3.3 High-Definition Photogrammetry via Monocular Depth Estimation
[0062] A query was performed on the Rhoton Collection to obtain 2D anatomical image data 112 for study that meet the following inclusion criteria: clear visualization of a superficial and deep plane, blurring in only one of these planes (blurring in both planes can, in theory, lower accuracy of the depth estimation process), sufficient lighting and brightness, and no images that contain surgical instruments other than retractors (because of potential artifacts).
3.4 Correlation Analysis of Depth Estimations and Neuronavigation
Measured Depth
[0063] To reduce inaccuracy of the set of depth estimation data 114, depth allocations by the neural network 120 are examined through a correlation process to verify consistency of the set of depth estimation data 114 with measurements collected with a neuronavigation system (e.g., a Medtronic StealthStation S7 neuronavigation system). For this process, anatomical regions of interest (ROIs) are selected according to their depth (see Tables 1 and 2). The neuronavigation system is used to measure actual spatial coordinates, and depth is determined for each ROI using Heron’s formula, considering the most superficial ROI as the reference value (depth = 0 mm). To consider the set of depth estimation data 114 as accurate, an inverse correlation is expected between increasing depth of anatomical structures and mean pixel intensity values measured in the depth map obtained by the neural network 120, where every anatomical structure is allocated a value from a range of 0 (purple) to 255 (yellow), corresponding to the deepest and the most superficial point of the image, respectively.
3.5 Statistical Analysis
[0064] Stata software version 13.0 (Stata Corp LLC, College Station, TX) is used for statistical analysis. Correlational analysis is performed using Spearman’s rho. An alpha value of <0.05 was considered statistically significant.
4. Results
[0065] One implementation of the system 100 is examined in this section for validation. The depth of each anatomical ROI and corresponding depth estimation (mean pixel intensity value and SD) relative to references points for a pretemporal approach (surface of retractor on frontal lobe) and mastoidectomy (spine of Henle) dissection are presented in Tables 1 and 2, respectively. [0066] Depth estimations are presented graphically as depth maps and histograms (FIGS. 8A-9I). Analysis of the histograms revealed that pixel distribution changed with different magnifications. In FIGS. 8A-8I (pretemporal dissection), the center of the image corresponded to the deepest region of the surgical field (basilar artery); the histogram revealed an increased pixel allocation toward zero with increasing magnification, a finding consistent with increasing depth allocation by the model.
[0067] FIGS. 8A-8I show Transsylvian exposure, right side, at different magnifications. FIG. 9A shows a low-magnification photograph, FIG. 8B shows a depth map of the photograph of FIG. 8A, and FIG. 8C shows a histogram of the photograph of FIG. 8A. FIG. 8D shows an intermediate-magnification photograph, FIG. 8B shows a depth map of the photograph of FIG. 8D, and FIG. 8C shows a histogram of the photograph of FIG. 8D. FIG. 8G shows a high-magnification photograph, FIG. 8H shows a depth map of the photograph of FIG. 8G, and FIG. 81 shows a histogram of the photograph of FIG. 8G, respectively.
[0068] FIGS. 9A-9I show a mastoidectomy dissection, the center of the image corresponding to an intermediate region in terms of depth (semicircular canals). The model assigned pixels in the range of 100 to 200 with increasing magnification, meaning the model allocated to an intermediate depth from the most superficial ROI (spine of Henle) and the deepest point Qugular bulb).
[0069] FIGS. 9A-9I show Mastoidectomy, left side, at different magnifications. FIG. 9A shows a low-magnification photograph, FIG. 9B shows a depth map of the photograph of FIG. 9A, and FIG. 9C shows a histogram of the photograph of FIG. 9A. FIG. 9D shows an intermediate-magnification photograph, FIG. 9B shows a depth map of the photograph of FIG. 9D, and FIG. 9C shows a histogram of the photograph of FIG. 9D. FIG. 9G shows a high-magnification photograph, FIG. 9H shows a depth map of the photograph of FIG. 9G, and FIG. 9I shows a histogram of the photograph of FIG. 9G, respectively.
[0070] FIGS. 10A-10D show high-magnification images with corresponding depth maps. FIG. 10A shows a transsylvian exposure photograph and FIG. 10B shows a depth map with overlay of anatomical regions of interest used to calculate mean pixel intensity values. Regions of interest: 1 = inferior frontal gyrus, 2 = middle cerebral artery bifurcation, 3 = optic nerve, 4 = supraclinoid internal carotid artery, 5 = oculomotor porus, 6 = oculomotor root entry zone, 7 = posterior clinoid process, 8 = basilar apex. FIG. 10C shows a mastoidectomy exposure photograph and FIG. 10D shows a depth map with overlay of anatomical regions of interest used to calculate mean pixel intensity values. Regions of interest: 1 = spine of Henle, 2 = incus, 3 = lateral semicircular canal, 4 = facial nerve vertical segment, 5 = jugular bulb. Quantitative spatial discrimination between the optic nerve (mean ±SD pixel intensity: 77.71 ±0.65), supraclinoid internal carotid artery (69.91 ±0.33) and the basilar apex (58.98±0.47) is observed in the transsylvian exposure (Table 1 , FIGS. 10A and 10B). The incus (128.1 ±1 .0), lateral semicircular canal (131 .4±0.7), and vertical segment of the facial nerve (130.6±0.4) were assigned similar mean pixel intensity values with a mastoidectomy exposure, consistent with their known anatomical relations and positions (FIGS. 10C and 10D, Table 2).
Table 1. Depth relative to the reference point (surface of retractor on frontal lobe) for each ROI and the corresponding mean pixel value at all three levels of magnification
Region of interest Measured Low Intermediate High depth magnification magnification magnification
(mm)
Inferior frontal gyrus 0 159.05 ± 0.21 156.56 ± 0.49 149.16 ± 0.82
Superficial sylvian 10.95 140.06 ± 0.81 134.57 ± 0.84 vein
MCA bifurcation 41.91 106.04 ± 1.61 120.17 ± 0.91 114.96 ± 1.31
ICA bifurcation 48.12 - 75.57 ± 0.71
Optic nerve 50.78 86.78 ± 0.93 65.43 ± 0.49 77.71 ± 0.65
Supraclinoid ICA 52.47 53.19 ± 0.39 63.48 ± 0.50 69.91 ± 0.33
Oculomotor nerve 53.59 - 58.71 ± 0.73 75.04 ± 0.45
REZ
Oculomotor porus 55.51 - 63.09 ± 4.42 68.12 ± 0.59
Posterior clinoid 57.76 - 61.10 ± 0.53 65.89 ± 0.55
Basilar apex 61.39 - 53.05 ± 0.21 58.98 ± 0.47
[0071] With respect to Table 1 , data are shown as mean pixel intensity ± standard deviation unless otherwise noted. Abbreviations: ICA, internal carotid artery; MCA, middle cerebral artery; REZ, root entry zone; ROI, region of interest.
Table 2. Depth relative to the reference point (spine of Henle) for each ROI and the corresponding mean pixel value at all three levels of magnification Region of interest Measured Low Intermediate High depth magnification magnification magnification (mm)
Spine of Henle 0 139.8 + 1.1 154.5 + 0.9 155.1 + 0.7
Incus 26.2 99.7 + 1.1 115.9 + 1.0 128.1 + 1.0
Lateral semicircular 27.9 114.5 + 0.7 118.5 + 0.5 131.4 + 0.7 canal
Facial nerve, vertical 29.0 77.5 + 0.5 75.7 + 0.7 130.6 + 0.4 segment
Jugular bulb 36.0 77.1 + 0.32 69.2 + 0.4 67.5 + 2.2
[0072] With respect to Table 2, data are shown as mean pixel intensity ± standard deviation unless otherwise noted.
[0073] Correlation analysis between pixel intensity values and measured depth revealed statistically significant inverse correlations. For the pretemporal approach dissection, the Spearman’s rho was -1.0 for the low- magnification image (p<0.0001 ), -0.98 for intermediate magnification (p<0.0001), and -0.97 for high magnification (p<0.0001). For the posterior temporal bone dissection, the Spearman’s rho was -0.90 (p=0.04) for each magnification studied.
4.1 Anatomical 3D Construction
[0074] The 3D methodology process applied by the system 100 is demonstrated in FIGS. 11A-11 D.
[0075] FIGS. 11A-11D show a 3D reconstruction procedure. FIG. 11 A shows a microsurgical image provided as input to the neural network 120, FIG. 11B shows a depth map generated by the neural network 120, FIG. 11C shows a mesh surface reconstruction generated by the 3D reconstruction framework 130, and FIG. 11D shows a final immersive anatomical model generated by the 3D reconstruction framework 130 from the mesh surface reconstruction.
[0076] Access to example 3D models is available online: orbytozigomatic pretemporal approach, perilabyrinthine approach, and extradural approach to the cavernous sinus.
4.2 Discussion
[0077] The principle of “a picture is worth a thousand words” is true in anatomy education: a high-quality microneurosurgical image can provide greater understanding than long paragraphs of descriptive anatomical observations and measurements. In this study, monoscopic photogrammetry using machine learning produced 3D reconstructions of microsurgical images. This allowed creation of navigable immersive anatomy experiences that were accurate in terms of spatial representation even with high magnification, enhancing the understanding of complex anatomical relationships.
4.33D Extended Reality and Neurosurgery Education
[0078] Mastery of anatomical orientation and visuospatial skills are necessary for a neurosurgeon to perform microsurgery with comfort and confidence and should be a fundamental part of any neurosurgery training program. Stereopsis, the optical-neurological phenomenon producing depth perception, is fundamental to a neurosurgeon’s understanding of anatomical relationships in small spaces. The cadaver dissection laboratory is the ideal environment for gaining this type of mastery, yet the availability of this training has many well-known limitations. The COVID-19 pandemic has further accelerated the trend toward extended reality, which can be used remotely. In fact, remote learning has become the mainstay of education during the pandemic.
[0079] Extended reality simulators for neurosurgery training have been in development for more than 20 years, and their use has increased in many centers worldwide. Using DICOM data from neuroimaging studies, 3D rendering of anatomical models is possible with accurate spatial localization of structures. However, from a microsurgical training perspective, these 3D renderings may be cartoonish or artificial and often fail to emulate essential features such as textures, colors, and lighting of the exposed neurovascular structures that must be identified when performing complex cranial approaches.
4.4 Photogrammetry: Traditional Versus Present System
[0080] Traditional photogrammetry requires acquisition of multiple images for 3D reconstruction. A pioneer experience using a robotic microscope for stereoscopic photogrammetry performed over 20 years ago was one of the most realistic simulators of neurosurgical approaches at the time but required specialized and expensive equipment. Currently, photogrammetry is commonly performed using software that superimposes multiple images of a subject and, through a triangulation process, creates a volumetric surface reconstruction. Immersive anatomy experiences of cranial approaches developed with photogrammetry have been reported by Rodriguez Rubio et al. and others. Similarly, Qlone, an app that performs 360° photogrammetry using mobile devices, has been used to create macroscopic and sectional 3D models of the brain and brainstem.
[0081] However, a caveat of photogrammetry is that it is difficult to create 3D models that accurately represent a microsurgical perspective. Some important neuroanatomical structures are only visible from very restricted angles, and mobile devices do not provide the same field of view as a microscope. Because our goal is to faithfully preserve the approach perspective, the use of only one image as input for 3D reconstruction is consistent. By using a neural network, we captured the surgical perspective of these narrow corridors and developed 3D models using an image from the microscope.
[0082] The present disclosure establishes that machine learning can deliver plausible anatomical depth estimations for specific sets of ROIs using only one image as input. This does not mean that depth allocations will be always accurate; it is necessary to perform a cross-check against neuronavigation or DICOM (Digital Imaging and Communications in Medicine) data to verify the accuracy of estimations. Through observation and validation of the system 100, however, that in the case of Rhoton Collection models, even without ground truth measurements available, overall 3D reconstruction by the system 100 is possible and anatomically consistent. The relationship of one anatomical structure to another made intuitive sense and was realistic overall.
4.5 Example of Augmenting the Rhoton Collection and Beyond
[0083] The system 100 can be applied to existing surgical image compendiums to enhance education. In one example, the Rhoton Collection, with its lectures and photographs of anatomical dissections using 3D stereoscopy that facilitate understanding of spatial relationships between complex structures, was used as a prime example of a compendium of images to which the system 100 could offer added value, allowing conversion of classic anatomical dissections into conveniently accessible, viewable, immersive, 3D, realistic environments. Further development of anatomy simulators based on other surgical dissection repositories and real-time surgery is an exciting possibility with this technology. Collaborative VR anatomical training with digital performance metrics and artificial intelligence integration for surgical performance classification could augment cadaver dissection in the future, even remotely.
4.6 Conclusions
[0084] Anatomical depth estimations calculated by pretrained models are relative estimations and are subject to differences with absolute spatial measurement. Further training and fine-tuning of the neural network 120 (and potential inclusion of additional or alternative neural networks) with a large data set of microsurgical images with “ground truth” information should increase the accuracy of estimations. In addition, individual coordinates co-registered from the neuronavigation system might not reflect in vivo reality due to the nature of cadaveric specimens and the natural phenomenon of brain shifting.
[0085] The system 100 implements machine learning principles for depth estimation of anatomical structures in microsurgical approaches using only one image as input. The methodologies discussed herein are inherently different from other forms of photogrammetry, which require multiple images, different angles, and prior knowledge of the structures. There appears to be no preceding studies that introduce 3D anatomical rendering of a single photo (monoscopic photogrammetry) and apply the 3D anatomical rendering to the field of microneurosurgery anatomy. The system 100 and associated methodologies allow creation of immersive, realistic, anatomical models useful for augmenting neuroanatomy education. Such a process could be especially convenient where either the availability of cadaveric specimens is limited or microsurgical dissection laboratories are absent, or merely when enhanced comprehension of the spatial relation of anatomical structures is desired.
[0086] It should be understood from the foregoing that, while particular embodiments have been illustrated and described, various modifications can be made thereto without departing from the spirit and scope of the invention as will be apparent to those skilled in the art. Such changes and modifications are within the scope and teachings of this invention as defined in the claims appended hereto.

Claims

CLAIMS What is claimed is:
1. A system, comprising: a processor in communication with a memory, the memory including instructions executable by the processor to: access a set of two-dimensional anatomical image data; generate a set of depth estimation data based on the set of two-dimensional anatomical image data by a monoscopic photogrammetry procedure; generate a three-dimensional mesh of a three- dimensional anatomical model using the set of depth estimation data; and generate a final rendering of the three-dimensional anatomical model using the set of two-dimensional anatomical image data and the three-dimensional mesh.
2. The system of claim 1 , the memory including instructions further executable by the processor to: generate the set of depth estimation data by a neural network, the neural network accepting the set of two-dimensional anatomical image data as input and being trained to output the set of depth estimation data that correlates with the set of two-dimensional anatomical image data.
3. The system of claim 2, the memory including instructions further executable by the processor to: compute an inverse depth value for each two-dimensional coordinate pair of a plurality of two-dimensional coordinate pairs of the set of two- dimensional anatomical image data; and generate a three-dimensional point cloud of the set of depth estimation data by combination of each two-dimensional coordinate pair and the inverse depth value for the plurality of two-dimensional coordinate pairs of the set of two-dimensional anatomical image data.
4. The system of claim 3, the memory including instructions further executable by the processor to: convert the three-dimensional point cloud to the three-dimensional mesh of the three-dimensional anatomical model.
5. The system of claim 1 , the memory further including instructions executable by the processor to: apply a spatial correlation analysis to the set of depth estimation data with respect to a set of neuronavigation data.
6. The system of claim 5, the set of depth estimation data being generated by a neural network, the neural network being trained by correlating a set of training depth estimation data generated by the neural network with empirical depth and/or linear measurements obtained through surgical neuronavigation and/or cadaveric dissection.
7. The system of claim 5, the set of depth estimation data being generated by a neural network, the neural network being trained by correlating a set of training two-dimensional anatomical image data provided as input to the neural network with empirical depth and/or linear measurements obtained through surgical neuronavigation and/or cadaveric dissection.
8. The system of claim 1 , further comprising: an application programming interface hosted at a server operable for communication with a virtual reality display device, the application programming interface providing instructions executable by a processor associated with the virtual reality display device to display the final rendering of the three-dimensional anatomical model within a virtual reality environment.
9. The system of claim 1 , further comprising: an application programming interface hosted at a server operable for communication with a mobile device, the application programming interface providing instructions executable by a processor of the mobile device to display the final rendering of the three-dimensional anatomical model within an augmented-reality environment.
10. The system of claim 9, the application programming interface providing instructions executable by the processor of the mobile device to: superimpose the final rendering of the three-dimensional anatomical model over a live feed captured by a camera of the mobile device.
11 . The system of claim 9, the augmented-reality environment being accessible by the mobile device through a quick-response code readable by a camera of the mobile device.
12. A method, comprising: accessing a set of two-dimensional anatomical image data; generating a set of depth estimation data based on the set of two- dimensional anatomical image data by a monoscopic photogrammetry procedure; generating a three-dimensional mesh of a three-dimensional anatomical model using the set of depth estimation data; and generating a final rendering of the three-dimensional anatomical model using the set of two-dimensional anatomical image data and the three- dimensional mesh.
13. The method of claim 12, further comprising: generating the set of depth estimation data by a neural network, the neural network accepting the set of two-dimensional anatomical image data as input and being trained to output the set of depth estimation data that correlates with the set of two-dimensional anatomical image data.
14. The method of claim 13, further comprising: computing an inverse depth value for each two-dimensional coordinate pair of a plurality of two-dimensional coordinate pairs of the set of two- dimensional anatomical image data; and generating a three-dimensional point cloud of the set of depth estimation data by combination of each two-dimensional coordinate pair and the inverse depth value for the plurality of two-dimensional coordinate pairs of the set of two-dimensional anatomical image data.
15. The method of claim 14, further comprising: converting the three-dimensional point cloud to the three-dimensional mesh of the three-dimensional anatomical model.
16. The method of claim 12, further comprising: applying a spatial correlation analysis to the set of depth estimation data with respect to a set of neuronavigation data.
17. The method of claim 16, the set of depth estimation data being generated by a neural network, the neural network being trained by correlating a set of training depth estimation data generated by the neural network with empirical depth and/or linear measurements obtained through surgical neuronavigation and/or cadaveric dissection.
18. The method of claim 16, the set of depth estimation data being generated by a neural network, the neural network being trained by correlating a set of training two-dimensional anatomical image data provided as input to the neural network with empirical depth and/or linear measurements obtained through surgical neuronavigation and/or cadaveric dissection.
19. The method of claim 12, further comprising: communicating the final rendering of the three-dimensional anatomical model to a server hosting an application programming interface operable for communication with a mobile device, the application programming interface providing instructions executable by a processor of the mobile device to display the final rendering of the three-dimensional anatomical model within an augmented-reality environment.
20. The method of claim 12, further comprising: communicating the final rendering of the three-dimensional anatomical model to a server hosting an application programming interface operable for communication with a virtual reality display device, the application programming interface providing instructions executable by a processor associated with the virtual reality display device to display the final rendering of the three-dimensional anatomical model within a virtual reality environment.
PCT/US2023/085928 2022-12-24 2023-12-26 Systems and methods for 3-dimensional anatomical reconstruction using monoscopic photogrammetry WO2024138224A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US202263477185P 2022-12-24 2022-12-24
US63/477,185 2022-12-24

Publications (1)

Publication Number Publication Date
WO2024138224A1 true WO2024138224A1 (en) 2024-06-27

Family

ID=91590217

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2023/085928 WO2024138224A1 (en) 2022-12-24 2023-12-26 Systems and methods for 3-dimensional anatomical reconstruction using monoscopic photogrammetry

Country Status (1)

Country Link
WO (1) WO2024138224A1 (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170284940A1 (en) * 2016-04-01 2017-10-05 Black Light Surgical, Inc. Systems, devices, and methods for time-resolved fluorescent spectroscopy
US20190371272A1 (en) * 2018-05-29 2019-12-05 SentiAR, Inc. Disposable sticker within augmented reality environment
US20210012568A1 (en) * 2018-02-23 2021-01-14 Sony Corporation Methods, devices and computer program products for gradient based depth reconstructions with robust statistics
US20220230303A1 (en) * 2021-01-15 2022-07-21 Boston Scientific Scimed Inc. Methods and apparatuses for generating anatomical models using diagnostic images

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170284940A1 (en) * 2016-04-01 2017-10-05 Black Light Surgical, Inc. Systems, devices, and methods for time-resolved fluorescent spectroscopy
US20210012568A1 (en) * 2018-02-23 2021-01-14 Sony Corporation Methods, devices and computer program products for gradient based depth reconstructions with robust statistics
US20190371272A1 (en) * 2018-05-29 2019-12-05 SentiAR, Inc. Disposable sticker within augmented reality environment
US20220230303A1 (en) * 2021-01-15 2022-07-21 Boston Scientific Scimed Inc. Methods and apparatuses for generating anatomical models using diagnostic images

Similar Documents

Publication Publication Date Title
CN110599605B (en) Image processing method and device, electronic equipment and computer readable storage medium
US20190238621A1 (en) Method and system for simulating surgical procedures
CN111784821B (en) Three-dimensional model generation method and device, computer equipment and storage medium
CN107822690B (en) Hybrid image/scene renderer with hands-free control
JP2024522287A (en) 3D human body reconstruction method, apparatus, device and storage medium
Gonzalez-Romo et al. Anatomic depth estimation and 3-dimensional reconstruction of microsurgical anatomy using monoscopic high-definition photogrammetry and machine learning
Lasserre et al. A neuron membrane mesh representation for visualization of electrophysiological simulations
CN112950769A (en) Three-dimensional human body reconstruction method, device, equipment and storage medium
EP4118619A1 (en) Pose estimation method and apparatus
CN113538682B (en) Model training method, head reconstruction method, electronic device, and storage medium
JP2023545189A (en) Image processing methods, devices, and electronic equipment
DE112018007236T5 (en) METHOD AND DEVICE FOR GENERATING A THREE-DIMENSIONAL (3D) MODEL FOR THE RECONSTRUCTION OF A 3D SCENE
DE102023125923A1 (en) GENERATIVE MACHINE LEARNING MODELS FOR PRIVACY PRESERVING SYNTHETIC DATA GENERATION USING DIFFUSION
US20190057167A1 (en) Method and module for transforming a shape of an object
WO2024138224A1 (en) Systems and methods for 3-dimensional anatomical reconstruction using monoscopic photogrammetry
Lee et al. Multi‐layer structural wound synthesis on 3D face
Voronov et al. Designing a subsystem for creating a three-dimensional model of an orthopedic insole based on data from a laser 3d scanning of the patient's feet
Allioui et al. A cooperative approach for 3D image segmentation
CN108596900B (en) Thyroid-associated ophthalmopathy medical image data processing device and method, computer-readable storage medium and terminal equipment
Weld et al. Regularising disparity estimation via multi task learning with structured light reconstruction
Gîrbacia Evaluation of cognitive effort in the perception of engineering drawings as 3D models
CN116012449A (en) Image rendering method and device based on depth information
DE112022002137T5 (en) Hybrid differentiable rendering for light transport simulation systems and applications
CN115393527A (en) Anatomical navigation construction method and device based on multimode image and interactive equipment
Abdolhoseini et al. Neuron image synthesizer via Gaussian mixture model and Perlin noise

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23908679

Country of ref document: EP

Kind code of ref document: A1