EP3452988B1 - Patientenvorpositionierung in rahmenloser kranialer radiochirurgie mithilfe von thermischer bildgebung - Google Patents

Patientenvorpositionierung in rahmenloser kranialer radiochirurgie mithilfe von thermischer bildgebung Download PDF

Info

Publication number
EP3452988B1
EP3452988B1 EP16721414.7A EP16721414A EP3452988B1 EP 3452988 B1 EP3452988 B1 EP 3452988B1 EP 16721414 A EP16721414 A EP 16721414A EP 3452988 B1 EP3452988 B1 EP 3452988B1
Authority
EP
European Patent Office
Prior art keywords
data
computer
thermal
reference structure
patient
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
EP16721414.7A
Other languages
English (en)
French (fr)
Other versions
EP3452988A1 (de
Inventor
Kajetan Berlinger
Hagen KAISER
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Brainlab AG
Original Assignee
Brainlab AG
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Brainlab AG filed Critical Brainlab AG
Publication of EP3452988A1 publication Critical patent/EP3452988A1/de
Application granted granted Critical
Publication of EP3452988B1 publication Critical patent/EP3452988B1/de
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61NELECTROTHERAPY; MAGNETOTHERAPY; RADIATION THERAPY; ULTRASOUND THERAPY
    • A61N5/00Radiation therapy
    • A61N5/10X-ray therapy; Gamma-ray therapy; Particle-irradiation therapy
    • A61N5/103Treatment planning systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/02Arrangements for diagnosis sequentially in different planes; Stereoscopic radiation diagnosis
    • A61B6/03Computed tomography [CT]
    • A61B6/032Transmission computed tomography [CT]
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/04Positioning of patients; Tiltable beds or the like
    • A61B6/0487Motor-assisted positioning
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/04Positioning of patients; Tiltable beds or the like
    • A61B6/0492Positioning of patients; Tiltable beds or the like using markers or indicia for aiding patient positioning
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61NELECTROTHERAPY; MAGNETOTHERAPY; RADIATION THERAPY; ULTRASOUND THERAPY
    • A61N5/00Radiation therapy
    • A61N5/10X-ray therapy; Gamma-ray therapy; Particle-irradiation therapy
    • A61N5/1048Monitoring, verifying, controlling systems and methods
    • A61N5/1064Monitoring, verifying, controlling systems and methods for adjusting radiation treatment in response to monitoring
    • A61N5/1069Target adjustment, e.g. moving the patient support
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/13Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/33Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
    • G06T7/337Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods involving reference images or patches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/37Determination of transform parameters for the alignment of images, i.e. image registration using transform domain methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/04Positioning of patients; Tiltable beds or the like
    • A61B6/0407Supports, e.g. tables or beds, for the body or parts of the body
    • A61B6/0421Supports, e.g. tables or beds, for the body or parts of the body with immobilising means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10048Infrared image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • G06T2207/30201Face
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30204Marker

Definitions

  • the present invention relates to a computer-implemented method for supporting positioning a patient for treatment by at least one of radiotherapy or radiosurgery, a corresponding computer program, a non-transitory program storage medium storing such a program and a computer for executing the program, as well as a system for supporting positioning a patient for treatment by at least one of radiotherapy or radiosurgery, the system comprising an electronic data storage device and the aforementioned computer.
  • WO 2015/120906 A1 discloses a pre-positioning for a patient during radiotherapy or radiosurgery which involves infrared-based tracking of an array of reflective markers attached to a supporting frame having a known position relative to the patient's body.
  • KAMATH R ET AL "Initial clinical experience with frameless radiosurgery for patients with intracranial metastases", INTERNATIONAL JOURNAL OF RADIATION: ONCOLOGY BIOLOGY PHYSICS, PERGAMON PRESS, USA, vol. 61, no. 5, 1 April 2005 (2005-04-01), pages 1467-1472 discloses using a biteplate containing optical markers for patient positioning in radiosurgery. The marker spheres are detected, and a thermal image is generated which shows the position of the marker spheres and of the patient.
  • TIMOTHY C RYKEN ET AL "INITIAL CLINICAL EXPERIENCE WITH FRAMELESS STEREOTACTIC RADIOSUGERY: ANALYSIS OF ACCURACY AND FEASIBILITY", INT. J. RADIATION ONCOLOGY BIOL. PHYS, vol. 51, no. 4, (2001-11-15) pages 1152-1158 also discloses use of biteplates for patient positioning in stereotactic radiotherapy.
  • the present invention is designed render a pre-positioning procedure for radiotherapy or radiosurgery more efficient.
  • the present invention can be used in connection with a system for image-guided radiotherapy such as ExacTrac®, a product of Brainlab AG.
  • the disclosed method encompasses comparing a live thermal image (infrared image) of a reference structure having a known position relative to an anatomical body part of a patient to a predetermined medical image of the reference structure associated with a known position relative to a reference position such as a radiotherapy isocentre. On that basis, it is determined whether the reference structure has moved relative to the reference position.
  • the aim of the method is to check whether the reference structure and therefore the patient has been correctly positioned and/or kept his desired position relative to a treatment device, and to compensate any possible positional deviation by moving e.g. the patient.
  • the invention reaches the aforementioned object by providing, in a first aspect, a computer-implemented method for supporting positioning (for example, pre-positioning or monitoring the position of) a patient for treatment by at least one of radiotherapy or radiosurgery.
  • the method comprises executing, on at least one processor of at least one computer, the following exemplary steps which are executed by the at least one processor.
  • planning image data is acquired which describes (for example, defines or represents) a digital planning image of a reference structure.
  • the reference structure comprises (for example, is) at least one of a medical device (such as at least one of a part of a patient support unit on which the patient is placed for the at least one of radiotherapy or radiosurgery (for example, cranial radiotherapy or radiosurgery), or a face mask for attaching to the patient's face e.g. for localizing the patient in space) or an anatomical body part (such as a part of the patient's face, for example on the surface of the face, or a part of the patient's torso).
  • a medical device such as at least one of a part of a patient support unit on which the patient is placed for the at least one of radiotherapy or radiosurgery (for example, cranial radiotherapy or radiosurgery), or a face mask for attaching to the patient's face e.g. for localizing the patient in space
  • an anatomical body part
  • the planning image data has been generated with at least one of different medical imaging modalities such as computed x-ray tomography (CT), magnetic resonance tomography (MRT), ultrasound tomography or thermal (infrared) imaging.
  • CT computed x-ray tomography
  • MRT magnetic resonance tomography
  • ultrasound tomography
  • thermal (infrared) imaging the planning image data has been generated by applying a tomographic imaging modality (e.g. CT, MRT or ultrasound) to the reference structure, or by imaging the reference structure with an infrared-sensitive imaging device.
  • the step of generating the planning image data is not necessarily part of the disclosed method, but may under circumstances be part of the disclosed method.
  • the planning image data has been generated before execution of the disclosed method starts so that the planning image data is predetermined and merely used as an input to the disclosed method.
  • the planning image is taken before the patient is placed ready (pre-positioned) for the at least one of radiotherapy or radiosurgery.
  • reference structure position data is acquired which describes (for example defines or represents) a predetermined (for example at least one of known or fixed) relative position between the reference structure and an anatomical body part of the patient's body and describing a relative position between the reference structure and a predetermined (for example, known and specifically, fixed) reference position.
  • the predetermined reference position was for example known from the imaging geometry (i.e. at least the relative position between an imaging device used for taking the planning image and the imaged reference structure) at the time at which the planning image data was generated and/or the planning image was taken.
  • the anatomical body part may be any part of the patient's body, and if the reference structure is an anatomical body part, the anatomical body part and the reference structure may be at least substantially the same and/or identical.
  • the predetermined reference position is in one example the isocentre of a treatment device usable for treating the patient with the at least one of radiotherapy or radiosurgery.
  • thermal image data is acquired which describes (for example, defines or represents) a digital thermal image (i.e. an infrared image) of the reference structure.
  • the thermal image is taken after the patient has been placed ready (pre-positioned) for the at least one of radiotherapy or radiosurgery (for example on a patient support unit such as a couch of a treatment device for carrying out the at least one of radiotherapy or radiosurgery).
  • generation of the thermal image data is in one example part of the disclosed method.
  • imaging device position data is acquired which describes (for example, defines or represents) a relative position between the reference structure and a thermal imaging device (for example, an infrared camera, specifically, a stereo-camera sensitive in the infrared wavelength range) used for taking the digital thermal image and describing a predetermined (for example, at least one of known or fixed) relative position between the thermal imaging device and the predetermined reference position.
  • a predetermined relative position may describe (for example, define or represent) at least one of the distance and the (for example three-dimensional) orientation and/or perspective from the thermal imaging device (specifically, from the position of the thermal imaging device) to the predetermined reference position.
  • planning image pattern data is determined which describes (for example, defines or represents) an image pattern (such as a at least one of a contour or planar area) in the digital planning image assigned to (for example, being) the representation of the reference structure, hereinforth called planning image pattern.
  • the planning image data is determined for example based on (for example, from) the planning image data.
  • determining the planning image pattern data comprises segmentation of the representation of the reference structure in the digital planning image. Thereby, for example a contour (e.g. the circumference) of at least part of the reference structure or an area describing the reference structure in the planning image can be determined.
  • thermal image pattern data is determined which describes (for example, defines or represents) an image pattern (such as at least one of a contour or planar area) in the digital thermal image assigned to (for example, being) the representation of the reference structure, hereinforth called thermal image pattern.
  • the thermal image pattern data is determined for example based on (for example, from) the thermal image data.
  • determining the thermal image pattern data comprises segmentation of the representation of the reference structure in the thermal image.
  • a contour e.g. the circumference
  • the thermal image pattern is an infrared signature of the reference structure.
  • determining the planning image data and the thermal image pattern data comprises determining similarity measure data (for example, by comparing the planning image to the thermal image) describing a similarity measure (such as a normalized cross-correlation, or image unit-wise (e.g. pixel-wise or voxel-wise) difference image and/or sum of squared differences) defining a similarity between image features in the digital planning image and the digital thermal image.
  • a similarity measure such as a normalized cross-correlation, or image unit-wise (e.g. pixel-wise or voxel-wise) difference image and/or sum of squared differences
  • At least one of them may be stretched for example by interpolation of image unit values (pixel values or voxel values) in the images so as to match the image dimensions with one another for comparing them to determine the similarity measure.
  • image unit values pixel values or voxel values
  • positional difference data is determined which describes (for example defines or represents) a difference between a relative position between the anatomical body part and the predetermined reference position at the point in time at which planning image data was generated on the one hand, and a relative position between the anatomical body part and the predetermined reference position at the point in time at which the thermal image data was generated on the other hand.
  • the difference may be equal to zero, but may also differ from zero.
  • the positional difference data is determined for example based on (for example, from) the reference structure position data and the imaging device position data and the planning image pattern data and the thermal image pattern data.
  • the positional difference data describes a transformation between the relative position between the anatomical body part and the predetermined reference position at the point in time at which planning image data was generated on the one hand, and the relative position between the anatomical body part and the predetermined reference position at the point in time at which the thermal image data was generated on the other hand.
  • a transformation is understood to be a linear transformation such as a mapping which can be embodied by a mapping matrix (which encompasses a vector) which considers for example at least three spatial degrees of freedom.
  • a transformation can be determined by accordingly solving the associated linear mapping equation, using basic linear algebra.
  • determining the positional difference data comprises: determining, based on (for example, from) the similarity measure data and the reference structure position data and the imaging device position data, relative position data describing (for example defining or representing) the relative position between the anatomical body part and the predetermined reference position at the point in time at which planning image data was generated on the one hand, and the relative position between the anatomical body part and the predetermined reference position at the point in time at which the thermal image data was generated on the other hand.
  • the disclosed method comprises:
  • the thermal reference data (for example, the predetermined model) is predetermined, i.e. generated before execution of the disclosed method starts, and read as an input to the disclosed method.
  • the predetermined model is generated for example by a statistical analysis of a plurality of thermal images so as to generate for example an average infrared signature of the reference structure for example under standard external conditions. If the reference structure is an anatomical body part, the predetermined model is generated for example by a statistical analysis of thermal images of the corresponding anatomical body part of a suitable plurality of human bodies.
  • the predetermined model may therefore be considered to represent a thermal atlas (thermo-atlas) of the reference structure.
  • the thermal planning data is for example generated by applying an image fusion algorithm to the planning image data and the thermal reference data in order to establish a transformation describing (for example defining or representing) the mapping of the predetermined model onto the digital planning image, specifically of corresponding image features in the predetermined model and the digital planning image onto each other.
  • a transformation describing for example defining or representing
  • the representation of the reference structure in the predetermined model is mapped onto (for example matched with) the representation of the reference structure in the planning image.
  • a re-shaped predetermined model conforming to the shape of the representation of the reference structure in the planning image can be generated.
  • the re-shaped predetermined model can then be compared to the representation of the reference structure in the thermal image, for example by applying an image fusion algorithm to the thermal image data and re-shaped thermal reference data (or the thermal planning data and the thermal reference data) so as to determine the position of the representation of the reference structure in the thermal image.
  • the position of the representation of the reference structure in the thermal image may be determined by the ways of determining the similarity measure as described above in the context of the for example fourth and fifth exemplary steps, namely by determining a similarity measure between the re-shaped predetermined model and the thermal image.
  • the thermal reference data can have been generated for example by imaging the reference structure with a non-thermal imaging modality (e.g. with CT or MRT) and assigning current or predetermined temperature values to the reference structure.
  • a non-thermal imaging modality e.g. with CT or MRT
  • the reference structure can be imaged in a state in which it is attached to the patient.
  • the representation of the reference structure can be segmented (i.e. determined by contour analysis), and different temperature values can be assigned to the representation of the reference structure and the representation of an anatomical body part of the patient (e.g. parts of the face which are visible through the mask).
  • the reference structure can be assigned lower temperature values than the anatomical body part.
  • the positional difference data can then be determined by comparing the artificial thermal image to the thermal image described by the thermal image data, for example by image unit-wise comparison.
  • the disclosed method comprises a step of determining, based on (e.g. from) the positional difference data, treatment device control data describing (for example, defining or representing) positional control information for controlling a relative position between a patient support unit for supporting the patient's body and a treatment device usable for treating the patient with at least one of radiotherapy or radiosurgery (for example, by moving at least one of the patient support unit or another part of the treatment device such as a beam source using a motorized motion control unit).
  • control data describing (for example, defining or representing) a command for moving the patient support unit relative to the treatment device is determined based on the treatment device control data. The command can be issued to the motion control unit.
  • the relative position between the patient support unit and the treatment device is changed such that a target area being the target of the at least one of radiotherapy or radiosurgery is positioned at a predetermined position such as the reference position.
  • the control data may describer (for example, define or represent) a command for controlling a beam source of the treatment device, for example to switch on or switch of a treatment beam or generally change the intensity of a treatment device.
  • the invention is directed to a computer program which, when running on at least one processor (for example, a processor) of at least one computer (for example, a computer) or when loaded into at least one memory (for example, a memory) of at least one computer (for example, a computer), causes the at least one computer to perform the above-described method according to the first aspect.
  • processor for example, a processor
  • memory for example, a memory
  • the invention is directed to a non-transitory computer-readable program storage medium on which the program according to the second aspect is stored.
  • the invention is directed to at least one computer (for example, a computer), comprising at least one processor (for example, a processor) and at least one memory (for example, a memory), wherein the program according to the second aspect is running on the processor or is loaded into the memory, or wherein the at least one computer comprises the program storage medium according to the third aspect.
  • a computer for example, a computer
  • the program according to the second aspect is running on the processor or is loaded into the memory
  • the at least one computer comprises the program storage medium according to the third aspect.
  • the invention is directed to a system for supporting positioning a patient for treatment by at least one of radiotherapy or radiosurgery, the system comprising:
  • the at least one computer is operably coupled to the at least one electronic data storage device for acquiring, from the at least one data storage device, the planning image data, the reference structure position data and the imaging device position data. Furthermore, the computer is operably coupled to at least one of the treatment device or the patient support unit.
  • computer program elements can be embodied by hardware and/or software (this includes firmware, resident software, micro-code, etc.).
  • computer program elements can take the form of a computer program product which can be embodied by a computer-usable, for example computer-readable data storage medium comprising computer-usable, for example computer-readable program instructions, "code” or a "computer program” embodied in said data storage medium for use on or in connection with the instruction-executing system.
  • Such a system can be a computer; a computer can be a data processing device comprising means for executing the computer program elements and/or the program in accordance with the invention, for example a data processing device comprising a digital processor (central processing unit or CPU) which executes the computer program elements, and optionally a volatile memory (for example a random access memory or RAM) for storing data used for and/or produced by executing the computer program elements.
  • a computer-usable, for example computer-readable data storage medium can be any data storage medium which can include, store, communicate, propagate or transport the program for use on or in connection with the instruction-executing system, apparatus or device.
  • the computer-usable, for example computer-readable data storage medium can for example be, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared or semiconductor system, apparatus or device or a medium of propagation such as for example the Internet.
  • the computer-usable or computer-readable data storage medium could even for example be paper or another suitable medium onto which the program is printed, since the program could be electronically captured, for example by optically scanning the paper or other suitable medium, and then compiled, interpreted or otherwise processed in a suitable manner.
  • the data storage medium is preferably a non-volatile data storage medium.
  • the computer program product and any software and/or hardware described here form the various means for performing the functions of the invention in the example embodiments.
  • the computer and/or data processing device can for example include a guidance information device which includes means for outputting guidance information.
  • the guidance information can be outputted, for example to a user, visually by a visual indicating means (for example, a monitor and/or a lamp) and/or acoustically by an acoustic indicating means (for example, a loudspeaker and/or a digital speech output device) and/or tactilely by a tactile indicating means (for example, a vibrating element or a vibration element incorporated into an instrument).
  • a computer is a technical computer which for example comprises technical, for example tangible components, for example mechanical and/or electronic components. Any device mentioned as such in this document is a technical and for example tangible device.
  • the method in accordance with the invention is for example a computer implemented method.
  • all the steps or merely some of the steps (i.e. less than the total number of steps) of the method in accordance with the invention can be executed by a computer (for example, at least one computer).
  • An embodiment of the computer implemented method is a use of the computer for performing a data processing method.
  • An embodiment of the computer implemented method is a method concerning the operation of the computer such that the computer is operated to perform one, more or all steps of the method.
  • the computer for example comprises at least one processor and for example at least one memory in order to (technically) process the data, for example electronically and/or optically.
  • the processor being for example made of a substance or composition which is a semiconductor, for example at least partly n- and/or p-doped semiconductor, for example at least one of II-, III-, IV-, V-, VI-semiconductor material, for example (doped) silicon and/or gallium arsenide.
  • the calculating steps described are for example performed by a computer. Determining steps or calculating steps are for example steps of determining data within the framework of the technical method, for example within the framework of a program.
  • a computer is for example any kind of data processing device, for example electronic data processing device.
  • a computer can be a device which is generally thought of as such, for example desktop PCs, notebooks, netbooks, etc., but can also be any programmable apparatus, such as for example a mobile phone or an embedded processor.
  • a computer can for example comprise a system (network) of "sub-computers", wherein each sub-computer represents a computer in its own right.
  • the term "computer” includes a cloud computer, for example a cloud server.
  • the term "cloud computer” includes a cloud computer system which for example comprises a system of at least one cloud computer and for example a plurality of operatively interconnected cloud computers such as a server farm.
  • Such a cloud computer is preferably connected to a wide area network such as the world wide web (WWW) and located in a so-called cloud of computers which are all connected to the world wide web.
  • WWW world wide web
  • Such an infrastructure is used for "cloud computing", which describes computation, software, data access and storage services which do not require the end user to know the physical location and/or configuration of the computer delivering a specific service.
  • the term "cloud” is used in this respect as a metaphor for the Internet (world wide web).
  • the cloud provides computing infrastructure as a service (laaS).
  • the cloud computer can function as a virtual host for an operating system and/or data processing application which is used to execute the method of the invention.
  • the cloud computer is for example an elastic compute cloud (EC2) as provided by Amazon Web ServicesTM.
  • a computer for example comprises interfaces in order to receive or output data and/or perform an analogue-to-digital conversion.
  • the data are for example data which represent physical properties and/or which are generated from technical signals.
  • the technical signals are for example generated by means of (technical) detection devices (such as for example devices for detecting marker devices) and/or (technical) analytical devices (such as for example devices for performing imaging methods), wherein the technical signals are for example electrical or optical signals.
  • the technical signals for example represent the data received or outputted by the computer.
  • the computer is preferably operatively coupled to a display device which allows information outputted by the computer to be displayed, for example to a user.
  • a display device is an augmented reality device (also referred to as augmented reality glasses) which can be used as "goggles" for navigating.
  • augmented reality glasses also referred to as augmented reality glasses
  • Google Glass a trademark of Google, Inc.
  • An augmented reality device can be used both to input information into the computer by user interaction and to display information outputted by the computer.
  • a display device would be a standard computer monitor comprising for example a liquid crystal display operatively coupled to the computer for receiving display control data from the computer for generating signals used to display image information content on the display device.
  • a specific embodiment of such a computer monitor is a digital lightbox.
  • the monitor may also be the monitor of a portable, for example handheld, device such as a smart phone or personal digital assistant or digital media player.
  • acquiring data for example encompasses (within the framework of a computer implemented method) the scenario in which the data are determined by the computer implemented method or program.
  • Determining data for example encompasses measuring physical quantities and transforming the measured values into data, for example digital data, and/or computing the data by means of a computer and for example within the framework of the method in accordance with the invention.
  • the meaning of "acquiring data” also for example encompasses the scenario in which the data are received or retrieved by the computer implemented method or program, for example from another program, a previous method step or a data storage medium, for example for further processing by the computer implemented method or program. Generation of the data to be acquired may but need not be part of the method in accordance with the invention.
  • the expression "acquiring data” can therefore also for example mean waiting to receive data and/or receiving the data.
  • the received data can for example be inputted via an interface.
  • the expression "acquiring data” can also mean that the computer implemented method or program performs steps in order to (actively) receive or retrieve the data from a data source, for instance a data storage medium (such as for example a ROM, RAM, database, hard drive, etc.), or via the interface (for instance, from another computer or a network).
  • the data acquired by the disclosed method or device, respectively may be acquired from a database located in a data storage device which is operably to a computer for data transfer between the database and the computer, for example from the database to the computer.
  • the computer acquires the data for use as an input for steps of determining data.
  • the determined data can be output again to the same or another database to be stored for later use.
  • the database or database used for implementing the disclosed method can be located on network data storage device or a network server (for example, a cloud data storage device or a cloud server) or a local data storage device (such as a mass storage device operably connected to at least one computer executing the disclosed method).
  • the data can be made "ready for use” by performing an additional step before the acquiring step.
  • the data are generated in order to be acquired.
  • the data are for example detected or captured (for example by an analytical device).
  • the data are inputted in accordance with the additional step, for instance via interfaces.
  • the data generated can for example be inputted (for instance into the computer).
  • the data can also be provided by performing the additional step of storing the data in a data storage medium (such as for example a ROM, RAM, CD and/or hard drive), such that they are ready for use within the framework of the method or program in accordance with the invention.
  • a data storage medium such as for example a ROM, RAM, CD and/or hard drive
  • the step of "acquiring data” can therefore also involve commanding a device to obtain and/or provide the data to be acquired.
  • the acquiring step does not involve an invasive step which would represent a substantial physical interference with the body, requiring professional medical expertise to be carried out and entailing a substantial health risk even when carried out with the required professional care and expertise.
  • the step of acquiring data does not involve a surgical step and in particular does not involve a step of treating a human or animal body using surgery or therapy.
  • the data are denoted (i.e. referred to) as "XY data” and the like and are defined in terms of the information which they describe, which is then preferably referred to as "XY information" and the like.
  • imaging methods are used to generate image data (for example, two-dimensional or three-dimensional image data) of anatomical structures (such as soft tissues, bones, organs, etc.) of the human body.
  • image data for example, two-dimensional or three-dimensional image data
  • medical imaging methods is understood to mean (advantageously apparatus-based) imaging methods (so-called medical imaging modalities and/or radiological imaging methods) such as for instance computed tomography (CT) and cone beam computed tomography (CBCT, such as volumetric CBCT), x-ray tomography, magnetic resonance tomography (MRT or MRI), conventional x-ray, sonography and/or ultrasound examinations, and positron emission tomography.
  • CT computed tomography
  • CBCT cone beam computed tomography
  • MRT or MRI magnetic resonance tomography
  • sonography and/or ultrasound examinations
  • positron emission tomography positron emission tomography
  • the image data thus generated is also termed "medical imaging data".
  • Analytical devices for example are used to generate the image data in apparatus-based imaging methods.
  • the imaging methods are for example used for medical diagnostics, to analyse the anatomical body in order to generate images which are described by the image data.
  • the imaging methods are also for example used to detect pathological changes in the human body.
  • some of the changes in the anatomical structure such as the pathological changes in the structures (tissue) may not be detectable and for example may not be visible in the images generated by the imaging methods.
  • a tumour represents an example of a change in an anatomical structure. If the tumour grows, it may then be said to represent an expanded anatomical structure.
  • This expanded anatomical structure may not be detectable; for example, only a part of the expanded anatomical structure may be detectable.
  • Primary/high-grade brain tumours are for example usually visible on MRI scans when contrast agents are used to infiltrate the tumour.
  • MRI scans represent an example of an imaging method.
  • the signal enhancement in the MRI images due to the contrast agents infiltrating the tumour
  • the tumour is detectable and for example discernible in the image generated by the imaging method.
  • enhancing tumours it is thought that approximately 10% of brain tumours are not discernible on a scan and are for example not visible to a user looking at the images generated by the imaging method.
  • Image fusion can be elastic image fusion or rigid image fusion.
  • rigid image fusion the relative position between the pixels of a 2D image and/or voxels of a 3D image is fixed, while in the case of elastic image fusion, the relative positions are allowed to change.
  • image morphing is also used as an alternative to the term “elastic image fusion”, but with the same meaning.
  • Elastic fusion transformations are for example designed to enable a seamless transition from one dataset (for example a first dataset such as for example a first image) to another dataset (for example a second dataset such as for example a second image).
  • the transformation is for example designed such that one of the first and second datasets (images) is deformed, for example in such a way that corresponding structures (for example, corresponding image elements) are arranged at the same position as in the other of the first and second images.
  • the deformed (transformed) image which is transformed from one of the first and second images is for example as similar as possible to the other of the first and second images.
  • (numerical) optimisation algorithms are applied in order to find the transformation which results in an optimum degree of similarity.
  • the degree of similarity is preferably measured by way of a measure of similarity (also referred to in the following as a "similarity measure").
  • the parameters of the optimisation algorithm are for example vectors of a deformation field. These vectors are determined by the optimisation algorithm in such a way as to result in an optimum degree of similarity.
  • the optimum degree of similarity represents a condition, for example a constraint, for the optimisation algorithm.
  • the bases of the vectors lie for example at voxel positions of one of the first and second images which is to be transformed, and the tips of the vectors lie at the corresponding voxel positions in the transformed image.
  • a plurality of these vectors is preferably provided, for instance more than twenty or a hundred or a thousand or ten thousand, etc.
  • constraints include for example the constraint that the transformation is regular, which for example means that a Jacobian determinant calculated from a matrix of the deformation field (for example, the vector field) is larger than zero, and also the constraint that the transformed (deformed) image is not self-intersecting and for example that the transformed (deformed) image does not comprise faults and/or ruptures.
  • the constraints include for example the constraint that if a regular grid is transformed simultaneously with the image and in a corresponding manner, the grid is not allowed to interfold at any of its locations.
  • the optimising problem is for example solved iteratively, for example by means of an optimisation algorithm which is for example a first-order optimisation algorithm, such as a gradient descent algorithm.
  • Other examples of optimisation algorithms include optimisation algorithms which do not use derivations, such as the downhill simplex algorithm, or algorithms which use higher-order derivatives such as Newton-like algorithms.
  • the optimisation algorithm preferably performs a local optimisation. If there is a plurality of local optima, global algorithms such as simulated annealing or generic algorithms can be used. In the case of linear optimisation problems, the simplex method can for instance be used.
  • the voxels are for example shifted by a magnitude in a direction such that the degree of similarity is increased.
  • This magnitude is preferably less than a predefined limit, for instance less than one tenth or one hundredth or one thousandth of the diameter of the image, and for example about equal to or less than the distance between neighbouring voxels.
  • Large deformations can be implemented, for example due to a high number of (iteration) steps.
  • the determined elastic fusion transformation can for example be used to determine a degree of similarity (or similarity measure, see above) between the first and second datasets (first and second images).
  • the deviation between the elastic fusion transformation and an identity transformation is determined.
  • the degree of deviation can for instance be calculated by determining the difference between the determinant of the elastic fusion transformation and the identity transformation. The higher the deviation, the lower the similarity, hence the degree of deviation can be used to determine a measure of similarity.
  • a measure of similarity can for example be determined on the basis of a determined correlation between the first and second datasets.
  • the present invention relates to the field of controlling a treatment beam.
  • the treatment beam treats body parts which are to be treated and which are referred to in the following as "treatment body parts". These body parts are for example parts of a patient's body, i.e. anatomical body parts.
  • the present invention relates to the field of medicine and for example to the use of beams, such as radiation beams, to treat parts of a patient's body, which are therefore also referred to as treatment beams.
  • a treatment beam treats body parts which are to be treated and which are referred to in the following as "treatment body parts". These body parts are for example parts of a patient's body, i.e. anatomical body parts.
  • Ionising radiation is for example used for the purpose of treatment.
  • the treatment beam comprises or consists of ionising radiation.
  • the ionising radiation comprises or consists of particles (for example, sub-atomic particles or ions) or electromagnetic waves which are energetic enough to detach electrons from atoms or molecules and so ionise them.
  • ionising radiation examples include x-rays, high-energy particles (high-energy particle beams) and/or ionising radiation emitted from a radioactive element.
  • the treatment radiation for example the treatment beam, is for example used in radiation therapy or radiotherapy, such as in the field of oncology.
  • parts of the body comprising a pathological structure or tissue such as a tumour are treated using ionising radiation.
  • the tumour is then an example of a treatment body part.
  • the treatment beam is preferably controlled such that it passes through the treatment body part.
  • the treatment beam can have a negative effect on body parts outside the treatment body part. These body parts are referred to here as "outside body parts".
  • a treatment beam has to pass through outside body parts in order to reach and so pass through the treatment body part.
  • Fig. 1 is a flow diagram illustrating the basic steps of the disclosed method which in the illustrative example of Fig. 1 starts with a step S1 of acquiring the planning image data.
  • step S2 the reference structure position data is acquired, followed by step S3 which is directed to acquiring the thermal image data.
  • step S4 continues with acquiring the imaging device position data.
  • step S5 which encompasses determining the planning image pattern data
  • step S6 which encompasses determining the thermal image pattern data.
  • step S7 which is illustrated in Fig. 1 as the final method step, the positional difference data is determined.
  • Fig. 2a is a photograph of a patient wearing an open face mask, and a thermal image corresponding to this photograph is shown in Fig. 2b .
  • the temperature of the visible ("open") portion of the face is approximately 37 °C which is significantly higher than that of the mask which is approximately 20 °C (room temperature).
  • the patient is lying on the treatment couch, but as treatment couch and mask do have room temperature, only the open face becomes apparent in Fig. 2b .
  • the patient is wearing the mask already during the scanning process for generating the planning image data. Based on the planning image, the treatment is planned and the target position (isocentre) is determined.
  • Fig. 3a shows a segmentation of the mask in the CT data corresponding to the planning image data for determining which image constituent corresponds to the mask and which image constituent corresponds to the patient's face.
  • a thermal image of the patient wearing the mask can also be taken from a different perspective.
  • a simulated thermal image (which can be a binary thermal image representing the silhouette of the face) can be rendered from the thermal volume.
  • CT data volume data
  • an exact reference image for pre-positioning the patient can be rendered.
  • the treatment couch can then be iteratively driven into the isocentre position so that the resulting live thermal image matches the simulated thermal image.
  • the position of the thermal volume is virtually optimized in space so that the resulting simulated thermal image matches the thermal image (live thermal image) described by the thermal image data.
  • the transformation resulting therefrom is used for pre-positioning.
  • Fig. 4 shows an embodiment of the algorithm shown in Fig. 1 .
  • Initial step S11 encompasses acquisition of the planning CT embodying the planning image of the patient to whose face an open face mask is fixated.
  • the representation of the mask is segmented in the planning image (CT data).
  • a thermal volume (having the same dimensions as the planning image) is created in step S12 by assigning to voxels of the mask a low temperature, and to voxels of the face a high temperature value.
  • a simulated thermal image is rendered in step S13 using the known geometrical setup of the camera calibration to the machine isocentre for the thermal camera (represented by the imaging device position data) and information about the treatment isocentre already known from the planning image data.
  • the simulated thermal image is rendered in an appearance it would have if the thermal volume were at a desired (planned) position.
  • the simulated thermal image is used as a reference image to preposition the patient.
  • Fig. 5 illustrates an algorithm for generating the similarity measure if the thermal image described by the thermal image data is considered to be at a certain position, and the thermal volume is virtually transformed in space from the desired (planned) position to match that certain position.
  • the live thermal image i.e. the thermal image described by the thermal image data
  • step S26 the rendered simulated thermal image is acquired.
  • step S22 the live thermal image and the simulated thermal image are compared to determine the similarity measure. Then, step S23 continues determining whether the similarity measure fulfils a predetermined convergence criterion. If this is the case, step S24 terminates the procedure with the transformation thus determined describing the difference defined by the positional difference data.
  • step S23 determines that the convergence criterion is not fulfilled, the thermal volume is virtually transformed in space to render a new simulated thermal image (from the thermal volume) in step S26, and the procedure continues again with step S22.
  • Fig. 6 shows the case in which the simulated thermal image is considered to be associated with a fixed position, and the live thermal image is considered to be moving.
  • the simulated thermal image is acquired as a reference image in an appearance it would have if the thermal volume were at a desired (planned) position, and compared in step S32 with the live thermal image acquired in step S36 to determine the similarity measure between the two images.
  • step S33 continues with determining whether the similarity measure fulfils a predetermined convergence criterion. If it is determined that this is the case, step S34 terminates the procedure with the current position of the treatment couch. If step S33 determines that the convergence criterion is not fulfilled, the position of the couch (and the patient) relative to the machine isocentre is physically changed, and a new real live thermal image is acquired in step S36.
  • Fig. 7 shows an implementation of the second approach shown in Fig. 6 .
  • initial step S41 the open face silhouette of the patient with the face mask attached to his face is manually brought into the field of view of the thermal camera by couch movement.
  • approach to (the second approach) is executed in step S42 for a first (coarse) pre-fusion by applying large translational steps used within the optimizer applied in step S33.
  • This process is repeated in step S43 using smaller translational and rotational steps within the optimizer for a fine fusion.
  • the resulting transformation is applied to the treatment couch in step S44 to appropriately position the patient.
  • Fig. 8 shows the overall flow of a pre-positioning procedure using the method disclosed herein.
  • step S51 the known geometrical setup between the couch and the patient fixation system for fixing the patient on the couch is used to automatically drive the couch to a position at which the mask (mask system) is brought into the field of view (FoV) of the thermal camera.
  • step S52 which encompasses using the second approach (approach 2) for a first coarse pre-fusion (using large translational steps within the optimizer of step S33).
  • step S53 is repeated in step S53 in order to generate a fine fusion using small translational and rotational steps within the optimizer of step S33.
  • the resulting transformation is applied to the treatment couch in step S54 to appropriately position the patient.
  • step S55 the couch movement can be monitored by applying the first approach (approach 1) between execution of step S51 and S52.
  • Fig. 9 shows a system for executing the method disclosed herein.
  • the system comprises a computer having a processor 12 and a volatile memory (RAM) 13 which are operably coupled to a non-volatile memory 14 being an electronic data storage device (such as a hard disk) storing at least the planning image data, the reference structure position data and the imaging device position data.
  • the computer is also coupled to an input device (such as at least one of a keyboard or a pointing device like a mouse) 15 and a display device 16 which can be embodied by a monitor.
  • the computer is operably coupled to the thermal imaging device embodied by a thermal camera (an infrared-sensitive camera) 17 having two optical units 18 for taking the digital thermal image described by the thermal image data.
  • a thermal camera an infrared-sensitive camera
  • the computer is also operably coupled to at least one of a treatment device 9 being a linear accelerator having a beam source for emitting an ionizing treatment beam which has a specifiable position relative to a patient support unit embodied by a couch 5 on which the patient's body 1 is positioned.
  • a reference structure represented by an open face mask 3 is attached to the patient's face 2.
  • the treatment device further may comprise an x-ray tube 8 and an x-ray detector 7 for taking x-ray images for checking the patient's position during the treatment.
  • the computer may alternatively or additionally be operably coupled to a motion control unit embodied by a motor 6 for driving the couch 5 (treatment table) to a desired position relative to the beam source.

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Physics & Mathematics (AREA)
  • Medical Informatics (AREA)
  • Biomedical Technology (AREA)
  • Theoretical Computer Science (AREA)
  • Pathology (AREA)
  • Animal Behavior & Ethology (AREA)
  • Veterinary Medicine (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Public Health (AREA)
  • General Health & Medical Sciences (AREA)
  • Radiology & Medical Imaging (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Molecular Biology (AREA)
  • Surgery (AREA)
  • High Energy & Nuclear Physics (AREA)
  • Optics & Photonics (AREA)
  • Biophysics (AREA)
  • Pulmonology (AREA)
  • Radiation-Therapy Devices (AREA)

Claims (12)

  1. Computerimplementiertes medizinisches Datenverarbeitungsverfahren zur Unterstützung der Positionierung eines Patienten zur Behandlung durch eine Strahlentherapie und/oder Radiochirurgie, wobei das Verfahren das Ausführen der folgenden Schritte auf mindestens einem Prozessor (12) mindestens eines Computers umfasst:
    a) Erfassen (S1), an dem mindestens einen Prozessor (12), von Planungsbilddaten, die ein digitales Planungsbild eines anatomischen Körperteils des Patienten und einer Referenzstruktur (3) beschreiben, wobei die Referenzstruktur (3) eine am Patienten angebrachte Vorrichtung ist;
    b) Erfassen (S2), an dem mindestens einen Prozessor (12) und basierend auf den Planungsbilddaten, von Referenzstruktur-Positionsdaten, die eine vorbestimmte relative Position zwischen der Referenzstruktur (3) und dem anatomischen Körperteil beschreiben, und die eine relative Position zwischen der Referenzstruktur und einem Zielbereich beschreiben, der planmäßig an einer vorbestimmten Referenzposition positioniert werden soll, wobei die vorbestimmte Referenzposition das Isozentrum einer Behandlungsvorrichtung (9) ist, die zur Behandlung des Patienten mit der Strahlentherapie und/oder Radiochirurgie verwendet werden kann;
    c) Erfassen (S3), an dem mindestens einen Prozessor (12), von Wärmebilddaten, die ein digitales Wärmebild der Referenzstruktur (3) beschreiben;
    d) Erfassen (S4), an dem mindestens einen Prozessor (12), von Bildgebungsvorrichtungs-Positionsdaten, die eine vorbestimmte relative Position zwischen der Wärmebildgebungsvorrichtung (17) und der vorbestimmten Referenzposition beschreiben;
    e) Bestimmen (S5), durch den mindestens einen Prozessor (12) und basierend auf den Planungsbilddaten, von Planungsbild-Musterdaten, die ein Bildmuster in dem digitalen Planungsbild beschreiben, das der Darstellung der Referenzstruktur (3) zugeordnet ist, nachstehend Planungsbildmuster genannt;
    f) Bestimmen (S6), durch den mindestens einen Prozessor (12) und basierend auf den Wärmebilddaten, von Wärmebild-Musterdaten, die ein Bildmuster in dem digitalen Wärmebild beschreiben, das der Darstellung der Referenzstruktur (3) zugeordnet ist, nachstehend Wärmebildmuster genannt.
    g) Bestimmen (S7), durch den mindestens einen Prozessor (12) und basierend auf den Positionsdaten der Referenzstruktur und den Positionsdaten der Bildgebungsvorrichtung und den Planungsbild-Musterdaten und den Wärmebild-Musterdaten, von Positionsdifferenzdaten, die eine Differenz beschreiben, zwischen einerseits einer relativen Position zwischen dem anatomischen Körperteil und dem Zielbereich beschreiben, zu dem Zeitpunkt, an dem Referenzstruktur-Positionsdaten erzeugt wurden, und andererseits einer relativen Position zwischen dem anatomischen Körperteil und der vorbestimmten Referenzposition zu dem Zeitpunkt, an dem die Wärmebilddaten erzeugt wurden,
    gekennzeichnet durch
    das Erfassen, an dem mindestens einen Prozessor (12), von thermischen Referenzdaten, die ein vorbestimmtes Modell einer Wärmebild-Erscheinung der Referenzstruktur beschreiben; und
    das Bestimmen, durch den mindestens einen Prozessor (12) und basierend auf den thermischen Referenzdaten und den Planungsbilddaten, von thermischen Planungsdaten, die ein Mapping des vorbestimmten Modells auf das digitale Planungsbild beschreiben,
    wobei das Bestimmen (S7) der Positionsdifferenzdaten ferner auf den thermischen Planungsdaten basiert.
  2. Verfahren nach dem vorhergehenden Anspruch, wobei das Bestimmen der Planungsbild-Musterdaten (S5) das Segmentieren der Darstellung der Referenzstruktur (3) in dem digitalen Planungsbild umfasst.
  3. Verfahren nach einem der vorhergehenden Ansprüche, wobei das Bestimmen der Wärmebild-Musterdaten (S6) eine Segmentierung der Darstellung der Referenzstruktur (3) in dem digitalen Wärmebild umfasst.
  4. Verfahren nach einem der vorhergehenden Ansprüche, wobei das Planungsbildmuster eine Kontur oder ein Bereich in dem digitalen Planungsbild ist.
  5. Verfahren nach einem der vorhergehenden Ansprüche, wobei das Wärmebildmuster eine Kontur oder ein Bereich in dem Planungsbild ist.
  6. Verfahren nach einem der vorhergehenden Ansprüche, wobei die Positionsdifferenzdaten eine Transformation beschreiben, zwischen einerseits der relativen Position zwischen dem anatomischen Körperteil und dem Zielbereich zu dem Zeitpunkt, an dem Referenzstruktur-Positionsdaten erzeugt wurden, und andererseits der relativen Position zwischen dem anatomischen Körperteil und der vorbestimmten Referenzposition zu dem Zeitpunkt, an dem die Wärmebilddaten erzeugt wurden.
  7. Verfahren nach einem der vorhergehenden Ansprüche, wobei die Planungsbilddaten durch Anwenden einer tomographischen Bildgebungsmodalität auf die Referenzstruktur (3) und den anatomischen Körperteil erzeugt wurden, oder durch das Abbilden der Referenzstruktur (3) und des anatomischen Körperteils mit einem infrarotempfindlichen Bildgebungsgerät.
  8. Verfahren nach einem der vorhergehenden Ansprüche, welches umfasst:
    Bestimmen, durch den mindestens einen Prozessor (12) und basierend auf dem Positionsdifferenzdaten, von Behandlungsvorrichtungs-Steuerungsdaten, die Positionssteuerungsinformationen zum Steuern einer relativen Position zwischen einer Patiententrägereinheit (5) zum Tragen des Körpers des Patienten (1) und einer Behandlungsvorrichtung (9) beschreiben, die zum Behandeln des Patienten durch Strahlentherapie und/oder Radiochirurgie verwendet werden kann, und zum Beispiel zum Bestimmen, durch den mindestens einen Prozessor (12) und basierend auf den Behandlungsvorrichtungs-Steuerungsdaten, von Steuerungsdaten, die einen Befehl zum Bewegen der Patiententrägereinheit (5) relativ zu dem Behandlungsgerät (9) beschreiben, und zum Ausgeben des Befehls an eine Bewegungssteuerungseinheit (6), beispielsweise so, dass der Zielbereich, der das Ziel der Strahlentherapie und/oder Radiochirurgie ist, an der vorbestimmten Referenzposition positioniert wird.
  9. Computerprogramm, das, wenn es auf mindestens einem Prozessor (12) von mindestens einem Computer ausgeführt wird, oder wenn es in den Speicher (13) von mindestens einem Computer geladen wird, bewirkt, dass der mindestens eine Computer das Verfahren gemäß einem der vorhergehenden Ansprüche ausführt.
  10. Ein nicht-flüchtiges, computerlesbares Programmspeichermedium (14), auf dem das Programm gemäß dem vorhergehenden Anspruch gespeichert ist.
  11. Mindestens ein Computer mit mindestens einem Prozessor (12) und einem Speicher (13), wobei das Programm nach Anspruch 9 auf dem mindestens einen Prozessor (12) läuft oder in den Speicher (13) geladen ist, oder wobei der mindestens eine Computer das Programmspeichermedium gemäß dem vorhergehenden Anspruch umfasst.
  12. System zur Unterstützung der Positionierung eines Patienten zur Behandlung durch eine Strahlentherapie und/oder Radiochirurgie, wobei das System umfasst:
    a) den mindestens einen Computer gemäß dem vorhergehenden Anspruch;
    b) mindestens eine elektronische Datenspeichervorrichtung (14), die mindestens die Planungsbilddaten, die Positionsdaten der Referenzstruktur und die Positionsdaten der Bildgebungsvorrichtung speichert;
    c) eine Wärmebildgebungsvorrichtung (17) zum Aufnehmen des digitalen Wärmebildes, wobei die Wärmebildvorrichtung (17) operabel mit dem mindestens einen Computer gekoppelt ist, um ein Signal an den mindestens einen Computer zu übertragen, das den Wärmebilddaten entspricht;
    d) eine Behandlungsvorrichtung (9), die zur Behandlung des Patienten durch eine Strahlentherapie und/oder Radiochirurgie verwendet werden kann; und
    e) eine Patiententrägereinheit (5) zum Tragen des Körpers des Patienten,
    wobei der mindestens eine Computer operabel mit der mindestens einen elektronischen Datenspeichervorrichtung (14) gekoppelt ist, um von der mindestens einen Datenspeichervorrichtung die Planungsbilddaten, die Positionsdaten der Referenzstruktur und die Positionsdaten der Bildgebungsvorrichtung zu erfassen; und
    wobei der Computer operabel mit der Behandlungsvorrichtungen (9) und/oder der Patiententrägereinheit (5) gekoppelt ist.
EP16721414.7A 2016-05-04 2016-05-04 Patientenvorpositionierung in rahmenloser kranialer radiochirurgie mithilfe von thermischer bildgebung Active EP3452988B1 (de)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/EP2016/060048 WO2017190780A1 (en) 2016-05-04 2016-05-04 Patient pre-positioning in frameless cranial radiosurgery using thermal imaging

Publications (2)

Publication Number Publication Date
EP3452988A1 EP3452988A1 (de) 2019-03-13
EP3452988B1 true EP3452988B1 (de) 2021-06-23

Family

ID=55953148

Family Applications (1)

Application Number Title Priority Date Filing Date
EP16721414.7A Active EP3452988B1 (de) 2016-05-04 2016-05-04 Patientenvorpositionierung in rahmenloser kranialer radiochirurgie mithilfe von thermischer bildgebung

Country Status (3)

Country Link
US (1) US10926106B2 (de)
EP (1) EP3452988B1 (de)
WO (1) WO2017190780A1 (de)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11273326B2 (en) * 2017-06-29 2022-03-15 Canon Medical Systems Corporation Radiotherapy system and treatment support apparatus
JP6611833B2 (ja) * 2018-01-16 2019-11-27 キヤノン株式会社 放射線撮影システム、並びに、カメラ制御装置及びその制御方法
WO2019149400A1 (de) * 2018-01-31 2019-08-08 Siemens Healthcare Gmbh Verfahren zur positionsplanung eines aufnahmesystems eines medizinischen bildgebenden geräts und medizinisches bildgebendes gerät
CN108619621B (zh) * 2018-05-23 2020-08-21 郑向鹏 一种用于肿瘤患者放射治疗精准定位及摆位的系统
WO2020015822A1 (en) * 2018-07-17 2020-01-23 Brainlab Ag Partitioning a medical image

Family Cites Families (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5563925A (en) * 1995-07-20 1996-10-08 Siemens Medical Systems, Inc. Apparatus and method for adjusting radiation in a radiation-emitting device
US20040064890A1 (en) 2002-06-19 2004-04-08 Siyong Kim Interactive patient positioning system
GB2390792B (en) * 2002-07-08 2005-08-31 Vision Rt Ltd Image processing system for use with a patient positioning device
US7412029B2 (en) * 2003-06-25 2008-08-12 Varian Medical Systems Technologies, Inc. Treatment planning, simulation, and verification system
US6935984B2 (en) * 2003-07-31 2005-08-30 Zf Friedrichshafen Ag Multi-stage automatic transmission with three planetary gearwheel assemblies
EP1563799B2 (de) 2004-02-11 2012-11-28 BrainLAB AG Verstellbare Markeranordnung
GB2418495B (en) * 2004-09-24 2010-04-28 Vision Rt Ltd Image processing system for use with a patient positioning device
US7453983B2 (en) * 2005-01-20 2008-11-18 Carestream Health, Inc. Radiation therapy method with target detection
GB2455926B (en) * 2006-01-30 2010-09-01 Axellis Ltd Method of preparing a medical restraint
EP2012673A4 (de) * 2006-04-20 2011-05-18 Univ California Wärmebehandlungsverfahren für myolyse und zerstörung gutartiger gebärmuttertumoren
US7693257B2 (en) * 2006-06-29 2010-04-06 Accuray Incorporated Treatment delivery optimization
US8155416B2 (en) * 2008-02-04 2012-04-10 INTIO, Inc. Methods and apparatuses for planning, performing, monitoring and assessing thermal ablation
US8831706B2 (en) * 2006-11-03 2014-09-09 Accuray Incorporated Fiducial-less tracking of a volume of interest
JP5110356B2 (ja) * 2007-07-10 2012-12-26 オムロン株式会社 検出装置および方法、並びに、プログラム
EP2663363B1 (de) * 2011-01-12 2017-04-19 Brainlab AG Bestimmung der position eines körperteils während einer strahlentherapie
GB2506903A (en) 2012-10-12 2014-04-16 Vision Rt Ltd Positioning patient for radio-therapy using 3D models and reflective markers
US9367926B2 (en) * 2012-10-26 2016-06-14 Brainlab Ag Determining a four-dimensional CT image based on three-dimensional CT data and four-dimensional model data
EP3104781B1 (de) 2014-02-14 2018-10-03 Brainlab AG Rahmenlose vorpositionierung für radiochirurgie
US9486170B2 (en) * 2014-03-18 2016-11-08 Monteris Medical Corporation Image-guided therapy of a tissue
US10278616B2 (en) * 2015-05-12 2019-05-07 Navix International Limited Systems and methods for tracking an intrabody catheter
EP3285639B1 (de) * 2016-05-04 2019-02-27 Brainlab AG Überwachung der position eines patienten mit einem planungsbild und anschliessender thermischer bildgebung
EP3556433B1 (de) * 2016-07-29 2021-01-27 Brainlab AG System zur überwachung der position eines patienten, der eine 4-pi-strahlentherapie erhält
US11443441B2 (en) * 2017-02-24 2022-09-13 Brainlab Ag Deep inspiration breath-hold setup using x-ray imaging

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
KAMATH R ET AL: "Initial clinical experience with frameless radiosurgery for patients with intracranial metastases", INTERNATIONAL JOURNAL OF RADIATION: ONCOLOGY BIOLOGY PHYSICS, PERGAMON PRESS, USA, vol. 61, no. 5, 1 April 2005 (2005-04-01), pages 1467 - 1472, XP027750293, ISSN: 0360-3016, [retrieved on 20050401] *
MARK H PHILLIPS ET AL: "COMMISSIONING AN IMAGE-GUIDED LOCALIZATION SYSTEM FOR RADIOTHERAPY", INT. J. RADIATION ONCOLOGY BIOL. PHYS., vol. 48, no. 1, 1 August 2000 (2000-08-01), pages 267 - 276, XP055597163, DOI: https://doi.org/10.1016/S0360-3016(00)00581-2 *
TIMOTHY C RYKEN ET AL: "INITIAL CLINICAL EXPERIENCE WITH FRAMELESS STEREOTACTIC RADIOSURGERY: ANALYSIS OF ACCURACY AND FEASIBILITY", INT. J. RADIATION ONCOLOGY BIOL. PHYS, vol. 51, no. 4, 15 November 2001 (2001-11-15), pages 1152 - 1158, XP055676319, DOI: https://doi.org/10.1016/S0360-3016(01)01756-4 *

Also Published As

Publication number Publication date
US20190111279A1 (en) 2019-04-18
US10926106B2 (en) 2021-02-23
EP3452988A1 (de) 2019-03-13
WO2017190780A1 (en) 2017-11-09

Similar Documents

Publication Publication Date Title
US20230008051A1 (en) Utilization of a transportable ct-scanner for radiotherapy procedures
US11645768B2 (en) Multi image fusion based positioning verification
EP3479355B1 (de) Oberflächengeführte röntgenstrahlenregistrierung
US10543381B2 (en) System for monitoring the position of a patient receiving 4pi radiation therapy
EP3452988B1 (de) Patientenvorpositionierung in rahmenloser kranialer radiochirurgie mithilfe von thermischer bildgebung
US20180344265A1 (en) Frameless Pre-Positioning for Radiosurgery
US11669982B2 (en) Soft tissue stereo-tracking
EP3622479B1 (de) Unterteilung eines medizinischen bildes
WO2023110116A1 (en) Ct-less free breathing image contour for planning radiation treatment
EP3391336A1 (de) Bildgebung mit langer belichtungszeit zur bestimmung von periodisch bewegten strukturen

Legal Events

Date Code Title Description
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE

PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE

17P Request for examination filed

Effective date: 20181030

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

AX Request for extension of the european patent

Extension state: BA ME

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: EXAMINATION IS IN PROGRESS

17Q First examination report despatched

Effective date: 20190701

DAV Request for validation of the european patent (deleted)
DAX Request for extension of the european patent (deleted)
GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: GRANT OF PATENT IS INTENDED

INTG Intention to grant announced

Effective date: 20210104

INTG Intention to grant announced

Effective date: 20210113

GRAS Grant fee paid

Free format text: ORIGINAL CODE: EPIDOSNIGR3

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE PATENT HAS BEEN GRANTED

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

REG Reference to a national code

Ref country code: GB

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: CH

Ref legal event code: EP

REG Reference to a national code

Ref country code: DE

Ref legal event code: R096

Ref document number: 602016059632

Country of ref document: DE

Ref country code: AT

Ref legal event code: REF

Ref document number: 1404963

Country of ref document: AT

Kind code of ref document: T

Effective date: 20210715

REG Reference to a national code

Ref country code: IE

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: SE

Ref legal event code: TRGR

REG Reference to a national code

Ref country code: LT

Ref legal event code: MG9D

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623

Ref country code: FI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623

Ref country code: BG

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210923

Ref country code: HR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623

REG Reference to a national code

Ref country code: AT

Ref legal event code: MK05

Ref document number: 1404963

Country of ref document: AT

Kind code of ref document: T

Effective date: 20210623

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: GR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210924

Ref country code: LV

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623

Ref country code: NO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210923

Ref country code: RS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623

REG Reference to a national code

Ref country code: NL

Ref legal event code: MP

Effective date: 20210623

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: SM

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623

Ref country code: SK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623

Ref country code: CZ

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623

Ref country code: EE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623

Ref country code: AT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623

Ref country code: NL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623

Ref country code: PT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20211025

Ref country code: RO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623

Ref country code: ES

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: PL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623

REG Reference to a national code

Ref country code: DE

Ref legal event code: R097

Ref document number: 602016059632

Country of ref document: DE

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: DK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623

PLBE No opposition filed within time limit

Free format text: ORIGINAL CODE: 0009261

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

26N No opposition filed

Effective date: 20220324

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: AL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623

REG Reference to a national code

Ref country code: CH

Ref legal event code: PL

REG Reference to a national code

Ref country code: BE

Ref legal event code: MM

Effective date: 20220531

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MC

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623

Ref country code: LU

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20220504

Ref country code: LI

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20220531

Ref country code: CH

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20220531

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20220504

Ref country code: FR

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20220531

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: BE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20220531

P01 Opt-out of the competence of the unified patent court (upc) registered

Effective date: 20230512

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: HU

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT; INVALID AB INITIO

Effective date: 20160504

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623

Ref country code: CY

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: GB

Payment date: 20240521

Year of fee payment: 9

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: DE

Payment date: 20240521

Year of fee payment: 9

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: SE

Payment date: 20240521

Year of fee payment: 9

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20210623