EP1828808B1 - Truncation compensation algorithm for iterative reconstruction - Google Patents

Truncation compensation algorithm for iterative reconstruction Download PDF

Info

Publication number
EP1828808B1
EP1828808B1 EP05824462A EP05824462A EP1828808B1 EP 1828808 B1 EP1828808 B1 EP 1828808B1 EP 05824462 A EP05824462 A EP 05824462A EP 05824462 A EP05824462 A EP 05824462A EP 1828808 B1 EP1828808 B1 EP 1828808B1
Authority
EP
European Patent Office
Prior art keywords
image
detector
field
radiation
view
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
EP05824462A
Other languages
German (de)
French (fr)
Other versions
EP1828808A2 (en
Inventor
Jinghan Ye
Mary K. Durbin
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Koninklijke Philips NV
Original Assignee
Koninklijke Philips Electronics NV
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Koninklijke Philips Electronics NV filed Critical Koninklijke Philips Electronics NV
Publication of EP1828808A2 publication Critical patent/EP1828808A2/en
Application granted granted Critical
Publication of EP1828808B1 publication Critical patent/EP1828808B1/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01TMEASUREMENT OF NUCLEAR OR X-RADIATION
    • G01T1/00Measuring X-radiation, gamma radiation, corpuscular radiation, or cosmic radiation
    • G01T1/16Measuring radiation intensity
    • G01T1/161Applications in the field of nuclear medicine, e.g. in vivo counting
    • G01T1/1615Applications in the field of nuclear medicine, e.g. in vivo counting using both transmission and emission sources simultaneously

Definitions

  • the present invention relates to the diagnostic imaging systems and methods. It finds particular application in conjunction with the Single Photon Emission Tomography (SPECT) systems with attenuation compensation and shall be described with particular reference thereto. It will be appreciated that the invention is also applicable to other imaging systems such as Positron Emission Tomography systems (PET), Computed Tomography systems (CT), and the like.
  • PET Positron Emission Tomography
  • CT Computed Tomography systems
  • Nuclear medicine imaging employs a source of radioactivity to image a patient.
  • a radiopharmaceutical is injected into the patient.
  • Radiopharmaceutical compounds contain a radioisotope that undergoes gamma-ray decay at a predictable rate and characteristic energy.
  • One or more radiation detectors are placed adjacent to the patient to monitor and record emitted radiation. Sometimes, the detector is rotated or indexed around the patient to monitor the emitted radiation from a plurality of directions. Based on information such as detected position and energy, the radiopharmaceutical distribution in the body is determined and an image of the distribution is reconstructed to study the circulatory system, radiopharmaceutical uptake in selected organs or tissue, and the like.
  • the iterative reconstruction technique each time a new volume estimate of the reconstructed data is generated, the previously reconstructed volume of image data is forward projected onto the plane of the detector. The forward projected data is compared to the actual projection data. If the reconstructed image were perfect, these two projections of data would match and there would be no difference. However, as the image is being built, there typically is a difference or error. The error or its inverse is then backprojected into the image volume to correct the volumetric image.
  • WO 00/25268A discloses a method of CT image reconstruction in which different values for at least one reconstruction parameter are used for at least two different portions of the image.
  • the at least one reconstruction parameter comprises pixel size.
  • US6339652B1 discloses a method of ML-EM image reconstruction provided for use in connection with a diagnostic imaging apparatus that generates projection data.
  • the present invention provides a new and improved imaging apparatus and method which overcomes the above-referenced problems and others.
  • an imaging system according to claim 1.
  • a method of imaging detecting and measuring at least one of emission and transmission radiation from a subject at a plurality of projection angles; generating measured projection data at the plurality of projection angles; and iteratively reconstructing the radiation detected into image representations in an image memory, characterized by the step of determining a plurality of pixels which belong to a field of view within which the radiation is detected by a radiation detector at each projection angle and that only the radiation detected in the field of view is iteratively reconstructed into image representations in an image memory.
  • an imaging system is disclosed. At least one radiation detector is disposed adjacent a subject receiving aperture to detect and measure at least one of emission and transmission radiation from a subject, the detector being movable around the subject to receive the radiation and generating measured projection data at a plurality of projection angles.
  • a field-of-view (FOV) means determines a plurality of pixels which belongs to a field of view of the radiation detector at each projection angle.
  • An image processor iteratively reconstructs the radiation detected only in the field of view into image representations. The image representations are iteratively reconstructed in an image memory.
  • a method of imaging is disclosed. At least one of emission and transmission radiation from a subject at a plurality of projection angles is detected and measured. Measured projection data is generated at the plurality of projection angles. A plurality of pixels which belongs to a field of view is determined, within which the radiation is detected by a radiation detector at each projection angle. The radiation detected in the field of view is iteratively reconstructed into image representations in an image memory.
  • One advantage of the present invention resides in reduced artifacts.
  • Another advantage resides in reduced clipping in reconstructed images.
  • Another advantage resides in reduced blurring in reconstructed images.
  • the invention may take form in various components and arrangements of components, and in various steps and arrangements of steps.
  • the drawings are only for purposes of illustrating the preferred embodiments and are not to be construed as limiting the invention.
  • a nuclear imaging system 10 typically includes a stationary gantry 12 that supports a rotatable gantry 14.
  • One or more detection heads 16 are carried by the rotatable gantry 14 to detect radiation events emanating from a region of interest or examination region 18.
  • Each detection head includes two-dimensional arrays of detector elements or detector 20 such as a scintillator and light sensitive elements, e.g. photomultiplier tubes, photodiodes, and the like. Direct x-ray to electrical converters, such as CZT elements, are also contemplated.
  • Each head 16 includes circuitry 22 for converting each radiation response into a digital signal indicative of its location (x, y) on the detector face and its energy (z).
  • the location of an event on the detector 20 is resolved and/or determined in a two dimensional (2D) Cartesian coordinate system with nominally termed x and y coordinates.
  • 2D two dimensional
  • a collimator 24 controls the direction and angular spread, from which each element of the detector 20 can receive radiation, i.e., the detector 20 can receive radiation only along known rays.
  • the determined location on the detector 20 at which radiation is detected and the angular position of the camera 16 define the nominal ray along which each radiation event occurred.
  • an object to be imaged is injected with one or more radiopharmaceuticals or radioisotopes and placed in the examination region 18 supported by a couch 26.
  • Radiopharmaceuticals or radioisotopes are Tc-99m, Ga-67, and In-111.
  • the presence of the radiopharmaceuticals within the object produces emission radiation from the object.
  • Radiation is detected by the detection heads 16 which are able to be angularly indexed or rotated around the examination region 18 to collect the projection emission data at one or more selected projection directions.
  • the projection emission data e.g. the location (x, y), energy (z), and an angular position ( ⁇ ) of each detection head 16 around the examination region 18 (e.g., obtained from an angular position resolver 28) are stored in an emission data memory 30.
  • the transmission radiation source is utilized to provide additional attenuation information to correct the emission data.
  • one or more radiation sources 50 are mounted across the examination region 18 from the detection heads 16.
  • the radiation sources 50 are mounted between the detection heads 16 or to the detection heads 16 such that transmission radiation from the radiation sources 50 is directed toward and received by corresponding detection head(s) 16 on an opposite side of the examination region 18 to complement the emission data.
  • the collimators 24 collimate the transmission radiation.
  • the collimators 24 restrict the scintillator 20 from receiving those portions of transmission radiation not traveling along rays normal (for parallel beam configurations) to the detector 20 or other direct path between the source and the detector. Alternately, other collimation geometries are employed and/or the collimation may take place at the source.
  • the radiation source(s) 50 are line sources each extending the axial length of the respective detection heads 16 to which they correspond.
  • the line sources take the form of thin steel tubes filled with radionuclides and sealed at their ends.
  • the radiation sources 50 can also be bar sources, point sources, flat rectangular sources, disk sources, flood sources, a tube or vessel filled with radionuclides, or active radiation generators such as x-ray tubes.
  • a transmission scan is performed such that transmission radiation from the transmission radiation source(s) 50 is also received by the detection head(s) 16, and transmission projection data is generated.
  • emission projection data is collected concurrently, but at different energy.
  • the transmission and emission data are often truncated as the size of an object 52 is typically bigger than the field of view of the detector 20, resulting in truncated projections P1, P2.
  • Truncated parts 54 of the object 52 which are not included in the field of view of the detector 20 are clipped off. As will be discussed in a greater detail below, the truncated parts 54 are compensated for from the projections taken at other angles at which the truncated parts 54 are not truncated.
  • a sorter 60 sorts the emission projection data and transmission projection data on the basis of their relative energies.
  • the emission data is stored in the emission data memory 30 and the transmission data is stored in a transmission memory 62.
  • a FOV determining means 64 determines which data are collected in the field of view of a corresponding detector 20 by one of the methods well known in the art as will be discussed in a greater detail below.
  • a data processor 70 iteratively reconstructs a 3D transmission radiation image or attenuation map 74 from the FOV transmission data while an image processor 72 iteratively reconstructs a 3D emission image from the FOV emission data.
  • the data processor 70 executes an iterative Bayesian Iterative Transmission Gradient Algorithm (BITGA), while the image processor 72 executes a Maximum Likelihood Expectation Maximization Algorithm (MLEM).
  • BITGA Bayesian Iterative Transmission Gradient Algorithm
  • MLEM Maximum Likelihood Expectation Maximization Algorithm
  • an attenuation map memory 74 and an image memory 76 are initialized by loading the memories 74, 76 with assumed or first estimates of the attenuation map and the emission maps.
  • the first estimate for the attenuation map is optionally characterized by a uniform attenuation value inside a predetermined contour which contains the subject and zero outside the contour.
  • the first emission maps estimate is optionally characterized by a uniform value inside the contour and zero outside.
  • the availability of additional a priori information allows for more accurate first estimates.
  • a first forward projector 80 creates projection transmission data from the transmission attenuation map 74.
  • a first comparing means 82 compares the measured transmission data with the forward projected data to determine correction factors.
  • the FOV determining means 64 determines if any part of the object is bigger than the field of view by one of the methods known in the art. E.g., the FOV determining means 64 determines whether some pixels were forced to zero values. Preferably, a series of reference scans is generated, from which the field of view is determined. Alternatively, the FOV determining means 64 searches each line in the normalized transmission projection data from an edge to a center to determine a sharp change in values between adjacent pixels.
  • a filter means 84 excludes all pixels beyond determined point of truncation from the correction factors matrix by filtering the out of the field of view data. E.g. the next iteration of reconstructed data is not corrected erroneously with values from outside the FOV.
  • the correction factors that correspond to the field of view are stored in a first correction memory 86.
  • first truncated parts 90 are not detected by a projection P4, but the first parts 90 is detected by a projection P3 which is taken at a second angular position ⁇ 2.
  • second truncated parts 92 are not detected by the projection P3
  • the second truncated parts 92 are detected by the projection P4 which is taken at the first angular position ⁇ 1.
  • Parts 94 are truncated in both illustrated projections P3 and P4.
  • a truncated data compensating means 100 supplies data of untruncated projections taken at different angular positions to compensate for the truncated parts 90, 92 outside of the field of view FOV.
  • a first backprojector 102 backprojects the correction factors into the attenuation map memory 74.
  • the back projection (of non zero values) is limited to the region for which actual data P1, P2 was collected.
  • the elements of the attenuation map 74 represent the attenuation factors of the respective voxels that are stored in an attenuation factors memory 104.
  • the image processor 72 iteratively reconstructs 3D image representation in the image memory 76 with each iteration including a forward projection operation and a backprojection operation.
  • a second forward projector 110 creates projection emission data from the reconstructed emission image stored in the image memory 76. More specifically, for each ray along which emission data is received, the second forward projector 110 calculates the projection of a corresponding ray through the transmission attenuation factors stored in the attenuation factors memory 104.
  • An attenuation correction means 112 adjusts the received emission data to compensate for the attenuation factor projection along the same ray.
  • a second comparing means 114 compares the attenuation corrected measured emission data along the ray with the forward projected data along the same ray to determine correction factors.
  • a filter means 84' excludes all pixels beyond determined point of truncation from the reconstruction by filtering the out of the field of view data.
  • the correction factors which correspond to the field of view are stored in a second correction memory 116.
  • a second backprojector 118 backprojects the correction factors into the reconstructed image, that is the reconstructed image is updated with the correction factors.
  • the truncated data compensating means 100 supplies data of untruncated projections taken at different angular positions to compensate for the truncated parts 90, 92 outside of the field of view FOV. Any determined corrections for areas outside the field of view are discarded.
  • the truncated data compensating means 100 applies a priori knowledge to the algorithm to smooth boundaries at the truncated side in the reconstructed image. Some pixels in the reconstructed image may "see" the detector FOV in only a few projection angles if the projection is highly truncated. In this case, some fuzziness is created at the truncated side in the reconstructed image. Applying a priori knowledge to the algorithm helps to alleviate the fuzziness at the truncated side by forcing to zero values those pixels in the reconstructed object that are confirmed to be outside of the field of view.
  • the data processor 70 and the image processor 72 execute other iterative algorithms including both transmission and emission iterative algorithms as well as a variety of other gradient and non gradient iterative algorithms. Each successive iteration is performed with the most recently updated emission image. When the reconstructed emission data and the measured emission data converge, i.e. the corrections fall below a preselected level, the iterative reconstruction process ends.
  • a population of a priori images (i.e., a "knowledge set") is collected.
  • the a priori images may not be from the patient currently under examination but rather be cross-sections of a similar structure such as images of other subjects obtained from CT scans, PET transmission scans, other SPECT transmission scans, and the like.
  • the initial a priori image may be of the same subject but taken with a different modality.
  • a video processor 130 retrieves slices, projections, 3D renderings, and other image information from the 3D image memory 76 and appropriately formats an image representation for display on one or more human viewable displays, such as a video monitor 132, printer, storage media, or the like. If the video processor repeatedly retrieves the selected image formation during reconstruction, the display will become clearer with each iteration as the reconstructed image converges on a final image.
  • ⁇ n j is the attenuation coefficient at the j th pixel and the nth iteration
  • f i is the reference scan value at the i th detector
  • l i,j is the length of the segment of the ray extending from the i th detector within the jth pixel
  • K i,j is the set of pixels from the jth detector to the i th detector
  • Y i is the transmission counts at the i th detector.
  • the detector matrix covers a larger area than the actual detector field-of-view.
  • the pixels which are located in the truncated parts 54 outside the FOV in the detector matrix are forced to zero values and are clipped off by the FOV determining means 64 as it is discussed above. All pixels beyond determined point of truncation are not included in the reconstruction, e.g. the projection data is truncated to exclude erroneous zero values from the reconstruction.
  • ⁇ j n + 1 ⁇ j n ⁇ ⁇ i ⁇ FOV f i exp - ⁇ k ⁇ K i , j ⁇ k n l i , k l i , j ⁇ i ⁇ FOV Y i l i , j + 1 - a j * r j
  • the projection data is not truncated if the object is small enough so that no events fall outside the detector FOV.

Abstract

An imaging system (10) includes at least one radiation detector (20) disposed adjacent a subject receiving aperture (18) to detect and measure at least one of emission and transmission radiation from a subject, the detector (20) at a plurality of projection angles. A processor (64) determines which radiation data belong to a field of view of the radiation detector (20) at each projection angle. An image processor (70, 72) iteratively reconstructs the radiation detected only in the determined field of view into image representations. Truncated data is compensated by supplying the untruncated data from the projections taken at different angles at which the truncated data is untruncated.

Description

  • The present invention relates to the diagnostic imaging systems and methods. It finds particular application in conjunction with the Single Photon Emission Tomography (SPECT) systems with attenuation compensation and shall be described with particular reference thereto. It will be appreciated that the invention is also applicable to other imaging systems such as Positron Emission Tomography systems (PET), Computed Tomography systems (CT), and the like.
  • Nuclear medicine imaging employs a source of radioactivity to image a patient. Typically, a radiopharmaceutical is injected into the patient. Radiopharmaceutical compounds contain a radioisotope that undergoes gamma-ray decay at a predictable rate and characteristic energy. One or more radiation detectors are placed adjacent to the patient to monitor and record emitted radiation. Sometimes, the detector is rotated or indexed around the patient to monitor the emitted radiation from a plurality of directions. Based on information such as detected position and energy, the radiopharmaceutical distribution in the body is determined and an image of the distribution is reconstructed to study the circulatory system, radiopharmaceutical uptake in selected organs or tissue, and the like.
  • Typically, in the iterative reconstruction technique, each time a new volume estimate of the reconstructed data is generated, the previously reconstructed volume of image data is forward projected onto the plane of the detector. The forward projected data is compared to the actual projection data. If the reconstructed image were perfect, these two projections of data would match and there would be no difference. However, as the image is being built, there typically is a difference or error. The error or its inverse is then backprojected into the image volume to correct the volumetric image.
  • Although these techniques work well, they are prone to truncation errors. That is, when the object is not seen completely by the detector in all detector positions, data for reconstructing some of the voxels appears only in some of the views. Thus, the data in the primary region of interest is typically fully sampled; whereas, surrounding tissue is less densely sampled or sampled in only a fraction of the views. In each iterative cycle, the backprojected error values span the entire image volume, including the fully sampled and under sampled regions. Allowing backprojection of error values from a comparison in which no data was present in the detector for the comparison leads to errors and artifacts, such as object clipping, in the less densely sampled regions.
  • Relevant examples of prior art include WO 00/25268A , US6339652B1 , Riddell C et al Institute of Electrical and Electronics Engineers: "Iterative Reprojection Reconstruction Algorithm with Attenuation Correction Applied to Truncated Projections in Spect" Proceedings of the Annual International Conference of the Engineering in Medicine and Biology Society. Paris, Oct.29-Nov.1 1992, New York, IEEE, US, vol. VOL. 5 Conf. 14, 29 October 1992 (1992-10-29), pages 1818-1820 and Hutton, Brian F.: "An introduction to iterative reconstruction" Alasbimn, vol. 5, no. 18, October 2002 (2002-10).
  • WO 00/25268A discloses a method of CT image reconstruction in which different values for at least one reconstruction parameter are used for at least two different portions of the image. Preferably the at least one reconstruction parameter comprises pixel size.
  • US6339652B1 discloses a method of ML-EM image reconstruction provided for use in connection with a diagnostic imaging apparatus that generates projection data.
  • The present invention provides a new and improved imaging apparatus and method which overcomes the above-referenced problems and others.
  • In accordance with an aspect of the present invention there is provided an imaging system according to claim 1.
  • In accordance with another aspect of the present invention there is provided a method of imaging according to claim 10.
    detecting and measuring at least one of emission and transmission radiation from a subject at a plurality of projection angles;
    generating measured projection data at the plurality of projection angles; and
    iteratively reconstructing the radiation detected into image representations in an image memory, characterized by the step of
    determining a plurality of pixels which belong to a field of view within which the radiation is detected by a radiation detector at each projection angle and that only the radiation detected in the field of view is iteratively reconstructed into image representations in an image memory.
  • In accordance with one aspect of the present invention, an imaging system is disclosed. At least one radiation detector is disposed adjacent a subject receiving aperture to detect and measure at least one of emission and transmission radiation from a subject, the detector being movable around the subject to receive the radiation and generating measured projection data at a plurality of projection angles. A field-of-view (FOV) means determines a plurality of pixels which belongs to a field of view of the radiation detector at each projection angle. An image processor iteratively reconstructs the radiation detected only in the field of view into image representations. The image representations are iteratively reconstructed in an image memory.
  • In accordance with another aspect of the present invention, a method of imaging is disclosed. At least one of emission and transmission radiation from a subject at a plurality of projection angles is detected and measured. Measured projection data is generated at the plurality of projection angles. A plurality of pixels which belongs to a field of view is determined, within which the radiation is detected by a radiation detector at each projection angle. The radiation detected in the field of view is iteratively reconstructed into image representations in an image memory.
  • One advantage of the present invention resides in reduced artifacts.
  • Another advantage resides in reduced clipping in reconstructed images.
  • Another advantage resides in reduced blurring in reconstructed images.
  • Still further advantages and benefits of the present invention will become apparent to those of ordinary skill in the art upon reading and understanding the following detailed description of the preferred embodiments.
  • The invention may take form in various components and arrangements of components, and in various steps and arrangements of steps. The drawings are only for purposes of illustrating the preferred embodiments and are not to be construed as limiting the invention.
    • FIGURE 1 is a diagrammatic illustration of an imaging system;
    • FIGURE 2 is a diagrammatic illustration of a portion of the imaging system in detail;
    • FIGURE 3 diagrammatically shows a truncation at one side of a subject with parallel bore collimation;
    • FIGURE 4 diagrammatically shows a truncation at one side of a subject with magnifying or fanbeam collimation; and
    • FIGURE 5 diagrammatically shows truncation parts that are outside of the field of view of a first angular position of a detector, but within the field of view of a second angular position of the detector.
  • With reference to FIGURE 1, a nuclear imaging system 10 typically includes a stationary gantry 12 that supports a rotatable gantry 14. One or more detection heads 16 are carried by the rotatable gantry 14 to detect radiation events emanating from a region of interest or examination region 18. Each detection head includes two-dimensional arrays of detector elements or detector 20 such as a scintillator and light sensitive elements, e.g. photomultiplier tubes, photodiodes, and the like. Direct x-ray to electrical converters, such as CZT elements, are also contemplated. Each head 16 includes circuitry 22 for converting each radiation response into a digital signal indicative of its location (x, y) on the detector face and its energy (z). The location of an event on the detector 20 is resolved and/or determined in a two dimensional (2D) Cartesian coordinate system with nominally termed x and y coordinates. However, other coordinate systems are contemplated. A collimator 24 controls the direction and angular spread, from which each element of the detector 20 can receive radiation, i.e., the detector 20 can receive radiation only along known rays. Thus, the determined location on the detector 20 at which radiation is detected and the angular position of the camera 16 define the nominal ray along which each radiation event occurred.
  • Typically, an object to be imaged is injected with one or more radiopharmaceuticals or radioisotopes and placed in the examination region 18 supported by a couch 26. Few examples of such isotopes are Tc-99m, Ga-67, and In-111. The presence of the radiopharmaceuticals within the object produces emission radiation from the object. Radiation is detected by the detection heads 16 which are able to be angularly indexed or rotated around the examination region 18 to collect the projection emission data at one or more selected projection directions. The projection emission data, e.g. the location (x, y), energy (z), and an angular position (θ) of each detection head 16 around the examination region 18 (e.g., obtained from an angular position resolver 28) are stored in an emission data memory 30.
  • As the emission data normally contains inaccuracies caused by varying absorption characteristics of the patient's anatomy, often the transmission radiation source is utilized to provide additional attenuation information to correct the emission data. In one embodiment, one or more radiation sources 50 are mounted across the examination region 18 from the detection heads 16. Optionally, the radiation sources 50 are mounted between the detection heads 16 or to the detection heads 16 such that transmission radiation from the radiation sources 50 is directed toward and received by corresponding detection head(s) 16 on an opposite side of the examination region 18 to complement the emission data. Preferably, the collimators 24 collimate the transmission radiation. E.g., the collimators 24 restrict the scintillator 20 from receiving those portions of transmission radiation not traveling along rays normal (for parallel beam configurations) to the detector 20 or other direct path between the source and the detector. Alternately, other collimation geometries are employed and/or the collimation may take place at the source.
  • Preferably, the radiation source(s) 50 are line sources each extending the axial length of the respective detection heads 16 to which they correspond. Preferably, the line sources take the form of thin steel tubes filled with radionuclides and sealed at their ends. Alternately, the radiation sources 50 can also be bar sources, point sources, flat rectangular sources, disk sources, flood sources, a tube or vessel filled with radionuclides, or active radiation generators such as x-ray tubes.
  • With continuing reference to FIGURE 1 and further reference to FIGURES 3-4, a transmission scan is performed such that transmission radiation from the transmission radiation source(s) 50 is also received by the detection head(s) 16, and transmission projection data is generated. Typically, emission projection data is collected concurrently, but at different energy. The transmission and emission data are often truncated as the size of an object 52 is typically bigger than the field of view of the detector 20, resulting in truncated projections P1, P2. Truncated parts 54 of the object 52 which are not included in the field of view of the detector 20 are clipped off. As will be discussed in a greater detail below, the truncated parts 54 are compensated for from the projections taken at other angles at which the truncated parts 54 are not truncated.
  • A sorter 60 sorts the emission projection data and transmission projection data on the basis of their relative energies. The emission data is stored in the emission data memory 30 and the transmission data is stored in a transmission memory 62. A FOV determining means 64 determines which data are collected in the field of view of a corresponding detector 20 by one of the methods well known in the art as will be discussed in a greater detail below.
  • With continuing reference to FIGURE 1 and further reference to FIGURE 2, a data processor 70 iteratively reconstructs a 3D transmission radiation image or attenuation map 74 from the FOV transmission data while an image processor 72 iteratively reconstructs a 3D emission image from the FOV emission data.
  • Preferably, the data processor 70 executes an iterative Bayesian Iterative Transmission Gradient Algorithm (BITGA), while the image processor 72 executes a Maximum Likelihood Expectation Maximization Algorithm (MLEM). In preparation for the first iteration of the reconstruction process, an attenuation map memory 74 and an image memory 76 are initialized by loading the memories 74, 76 with assumed or first estimates of the attenuation map and the emission maps. The first estimate for the attenuation map is optionally characterized by a uniform attenuation value inside a predetermined contour which contains the subject and zero outside the contour. Likewise, the first emission maps estimate is optionally characterized by a uniform value inside the contour and zero outside. Alternately, the availability of additional a priori information allows for more accurate first estimates.
  • A first forward projector 80 creates projection transmission data from the transmission attenuation map 74. A first comparing means 82 compares the measured transmission data with the forward projected data to determine correction factors. The FOV determining means 64 determines if any part of the object is bigger than the field of view by one of the methods known in the art. E.g., the FOV determining means 64 determines whether some pixels were forced to zero values. Preferably, a series of reference scans is generated, from which the field of view is determined. Alternatively, the FOV determining means 64 searches each line in the normalized transmission projection data from an edge to a center to determine a sharp change in values between adjacent pixels. If the difference between a current pixel and a previous pixel is greater than a prespecified threshold, the FOV determining means 64 concludes that there exists a truncation. A filter means 84 excludes all pixels beyond determined point of truncation from the correction factors matrix by filtering the out of the field of view data. E.g. the next iteration of reconstructed data is not corrected erroneously with values from outside the FOV. The correction factors that correspond to the field of view are stored in a first correction memory 86.
  • With continuing reference to FIGURES 1 and 2 and further reference to FIGURE 5, while taking projections at a first angular position α1, first truncated parts 90 are not detected by a projection P4, but the first parts 90 is detected by a projection P3 which is taken at a second angular position α2. Likewise, while the projections are taken at the second angular position α2, second truncated parts 92 are not detected by the projection P3, the second truncated parts 92 are detected by the projection P4 which is taken at the first angular position α1. Parts 94 are truncated in both illustrated projections P3 and P4. A truncated data compensating means 100 supplies data of untruncated projections taken at different angular positions to compensate for the truncated parts 90, 92 outside of the field of view FOV. A first backprojector 102 backprojects the correction factors into the attenuation map memory 74. The back projection (of non zero values) is limited to the region for which actual data P1, P2 was collected. The elements of the attenuation map 74 represent the attenuation factors of the respective voxels that are stored in an attenuation factors memory 104.
  • With continuing reference to FIGURES 1 and 2, the image processor 72 iteratively reconstructs 3D image representation in the image memory 76 with each iteration including a forward projection operation and a backprojection operation. A second forward projector 110 creates projection emission data from the reconstructed emission image stored in the image memory 76. More specifically, for each ray along which emission data is received, the second forward projector 110 calculates the projection of a corresponding ray through the transmission attenuation factors stored in the attenuation factors memory 104. An attenuation correction means 112 adjusts the received emission data to compensate for the attenuation factor projection along the same ray. A second comparing means 114 compares the attenuation corrected measured emission data along the ray with the forward projected data along the same ray to determine correction factors. A filter means 84' excludes all pixels beyond determined point of truncation from the reconstruction by filtering the out of the field of view data. The correction factors which correspond to the field of view are stored in a second correction memory 116. A second backprojector 118 backprojects the correction factors into the reconstructed image, that is the reconstructed image is updated with the correction factors. The truncated data compensating means 100 supplies data of untruncated projections taken at different angular positions to compensate for the truncated parts 90, 92 outside of the field of view FOV. Any determined corrections for areas outside the field of view are discarded.
  • In one embodiment, the truncated data compensating means 100 applies a priori knowledge to the algorithm to smooth boundaries at the truncated side in the reconstructed image. Some pixels in the reconstructed image may "see" the detector FOV in only a few projection angles if the projection is highly truncated. In this case, some fuzziness is created at the truncated side in the reconstructed image. Applying a priori knowledge to the algorithm helps to alleviate the fuzziness at the truncated side by forcing to zero values those pixels in the reconstructed object that are confirmed to be outside of the field of view.
  • Of course, it is also contemplated that the data processor 70 and the image processor 72 execute other iterative algorithms including both transmission and emission iterative algorithms as well as a variety of other gradient and non gradient iterative algorithms. Each successive iteration is performed with the most recently updated emission image. When the reconstructed emission data and the measured emission data converge, i.e. the corrections fall below a preselected level, the iterative reconstruction process ends.
  • In one embodiment, a population of a priori images (i.e., a "knowledge set") is collected. The a priori images may not be from the patient currently under examination but rather be cross-sections of a similar structure such as images of other subjects obtained from CT scans, PET transmission scans, other SPECT transmission scans, and the like. As another option, the initial a priori image may be of the same subject but taken with a different modality.
  • A video processor 130 retrieves slices, projections, 3D renderings, and other image information from the 3D image memory 76 and appropriately formats an image representation for display on one or more human viewable displays, such as a video monitor 132, printer, storage media, or the like. If the video processor repeatedly retrieves the selected image formation during reconstruction, the display will become clearer with each iteration as the reconstructed image converges on a final image.
  • With continuing reference to FIGURES 1 and 2, the Iterative Transmission Gradient Algorithm can be expressed as: µ j n + 1 = µ j n i f i exp - k K i , j µ k n l i , k l i , j i Y i l i , j
    Figure imgb0001

    where
    µ n j is the attenuation coefficient at the jth pixel and the nth iteration,
    fi is the reference scan value at the ith detector,
    li,j is the length of the segment of the ray extending from the ith detector within the jth pixel,
    Ki,j is the set of pixels from the jth detector to the ith detector, and
    Yi is the transmission counts at the ith detector.
  • The Bayesian Iterative Transmission Gradient Algorithm (BITGA) which includes a spatially varying gamma prior can be expressed as: µ j n + 1 = a j µ j n i f i exp - k K i , j µ k n l i , k l i , j i Y i l i , j + 1 - a j r j
    Figure imgb0002

    where rj is value "preferred" pixel value, and aj is its weight.
  • Preferably, the Bayesian Iterative Transmission Gradient Algorithm includes a modified prior block which uses a priori knowledge to encourage the each pixel's value to converge either to the attenuation coefficient of water, e.g. tissue, or air, e.g. air in the lungs: r j = µ ST
    Figure imgb0003
    a j = a 0 exp - µ j - µ ST δ 2
    Figure imgb0004

    where δ defines the extent of the prior.
  • Typically, the detector matrix covers a larger area than the actual detector field-of-view. The pixels which are located in the truncated parts 54 outside the FOV in the detector matrix are forced to zero values and are clipped off by the FOV determining means 64 as it is discussed above. All pixels beyond determined point of truncation are not included in the reconstruction, e.g. the projection data is truncated to exclude erroneous zero values from the reconstruction.
  • The modified Bayesian Iterative Transmission Gradient Algorithm which reconstructs pixels only from the field of view can be expressed as: µ j n + 1 = µ j n i FOV f i exp - k K i , j µ k n l i , k l i , j i FOV Y i l i , j + 1 - a j * r j
    Figure imgb0005
  • Of course, the projection data is not truncated if the object is small enough so that no events fall outside the detector FOV.

Claims (17)

  1. An imaging system (10) comprising:
    at least one radiation detector (20) disposed adjacent a subject receiving aperture (18) to detect and measure at least one of emission and transmission radiation from a subject, the detector (20) being movable around the subject to receive the radiation and generating measured projection data at a plurality of projection angles;
    an image processor (70, 72) which iteratively reconstructs the radiation detected into image representations; and
    an image memory (74, 76) in which the image representations are iteratively reconstructed; characterised in that the imaging system further comprises:
    a FOV means (64) for determining a plurality of pixels in the reconstructed image representation which belong to a field of view of the radiation detector (20) at each projection angle, and that the image processor only iteratively reconstructs the radiation detected in the field of view into image representations.
  2. The system as set forth in claim 1, wherein the image processor (70, 72) includes:
    a forward projector (80, 110) which projects the image representations from the corresponding image memory (74, 76);
    a means (82, 114) for comparing the forward projected image representations with the measured projection data and, based on the comparison, determining a set of correction factors; and
    a backprojector (102, 118) for backprojecting the correction factors, which correspond only to the pixels within the determined field of view, into the image representations.
  3. The system as set forth in claim 2, wherein the image processor (70, 72) further includes:
    a filter means (84, 84') for removing the out of the field of view data prior to back projecting correction factors.
  4. The system as set forth in claim 3, wherein the backprojector (102, 118) does not back project the outside of the field of view pixels.
  5. The system as set forth in claim 1, further including:
    a means (100) for compensating for truncated data by at least one of:
    supplying the untruncated data from the projections taken at different angles at which the truncated data is untruncated, and
    using a priori knowledge.
  6. The system as set forth in claim 1, wherein the image processor (70) executes at least one of:
    an iterative transmission reconstruction gradient algorithm;
    an iterative transmission reconstruction non gradient algorithm;
    an iterative emission reconstruction gradient algorithm; and
    an iterative emission reconstruction non gradient algorithm.
  7. The system as set forth in claim 6, wherein the image processor (70) executes an Iterative Transmission Gradient Algorithm, wherein the out of the field of view pixels are excluded from the reconstruction: µ j n + 1 = µ j n i FOV f i exp - k K i , j µ k n l i , k l i , j i FOV Y i l i , j
    Figure imgb0006

    where
    µ n j is the attenuation coefficient at a jth pixel and the nth iteration,
    fi is a reference scan value at an ith detector,
    li,j is a length of a segment of a ray extending from the ith detector within the jth pixel,
    Ki,j is a set of pixels from the jth detector to the ith detector, and
    Yi is transmission counts at the ith detector.
  8. The system as set forth in claim 7, further including:
    a means (100) for compensating for truncated data which compensating means (100) uses a priori knowledge to reconfirm the out of the field of view pixels and force the reconfirmed out of the field of view pixels to zero values in the reconstructed image.
  9. The system as set forth in claim 1, further including:
    at least one of a PET scanner and a SPECT scanner to detect emission projections from the subject.
  10. A method of imaging comprising:
    detecting and measuring at least one of emission and transmission radiation from a subject at a plurality of projection angles;
    generating measured projection data at the plurality of projection angles;
    iteratively reconstructing the radiation detected into image representations in an image memory, characterised by the step of:
    determining a plurality of pixels in the reconstructed image representation which belong to a field of view within which the radiation is detected by a radiation detector at each projection angle and that only the radiation detected in the field of view is iteratively reconstructed into image representations in an image memory.
  11. The method as set forth in claim 10, wherein the step of reconstruction includes:
    forward projecting the image representations;
    comparing the forward projected image representations with the measured projection data;
    based on the comparison, determining a set of correction factors; and
    backprojecting the correction factors, which correspond only to the pixels within the determined field of view, into the image representations.
  12. The method as set forth in claim 11, wherein the step of reconstruction further includes:
    removing the out of the field of view data prior to back projecting.
  13. The method as set forth in claim 10, wherein the step of determining the FOV includes:
    generating reference scans.
  14. The method as set forth in claim 10, further including:
    compensating for truncated data by at least one of:
    supplying an untruncated data from other projections, and
    using a priori knowledge.
  15. The method as set forth in claim 10, wherein the step of reconstruction includes executing at least one of:
    an iterative transmission reconstruction gradient algorithm;
    an iterative transmission reconstruction non gradient algorithm;
    an iterative emission reconstruction gradient algorithm; and
    an iterative emission reconstruction non gradient algorithm
  16. The method as set forth in claim 15, wherein the step of reconstruction further includes:
    executing an Iterative Transmission Gradient Algorithm, wherein the out of the field of view pixels are excluded from the reconstruction: µ j n + 1 = µ j n i FOV f i exp - k K i , j µ k n l i , k l i , j i FOV Y i l i , j
    Figure imgb0007

    where .
    µ n j the attenuation coefficient at a jth pixel and the nth iteration,
    fi is a reference scan value at an ith detector,
    li,j is a length of a segment of a ray extending from the ith detector within the jth pixel,
    Ki,j is a set of pixels from the jth detector to the ith detector, and
    Yi is transmission counts at the ith detector.
  17. The method as set forth in claim 16, further including:
    reconfirming the out of the field of view pixels by using a priori knowledge,
    and
    forcing the reconfirmed out of the field of view pixels to zero values in the reconstructed image.
EP05824462A 2004-12-17 2005-12-05 Truncation compensation algorithm for iterative reconstruction Active EP1828808B1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US63718604P 2004-12-17 2004-12-17
PCT/IB2005/054070 WO2006064404A2 (en) 2004-12-17 2005-12-05 Truncation compensation algorithm for iterative reconstruction

Publications (2)

Publication Number Publication Date
EP1828808A2 EP1828808A2 (en) 2007-09-05
EP1828808B1 true EP1828808B1 (en) 2010-04-21

Family

ID=36168839

Family Applications (1)

Application Number Title Priority Date Filing Date
EP05824462A Active EP1828808B1 (en) 2004-12-17 2005-12-05 Truncation compensation algorithm for iterative reconstruction

Country Status (7)

Country Link
US (1) US8013307B2 (en)
EP (1) EP1828808B1 (en)
JP (1) JP4731571B2 (en)
CN (1) CN101080651B (en)
AT (1) ATE465425T1 (en)
DE (1) DE602005020844D1 (en)
WO (1) WO2006064404A2 (en)

Families Citing this family (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5180181B2 (en) * 2006-03-16 2013-04-10 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Computer tomography data collection apparatus and method
US8335358B2 (en) * 2007-03-29 2012-12-18 Palodex Group Oy Method and system for reconstructing a medical image of an object
US8600136B2 (en) * 2008-09-19 2013-12-03 Koninklijke Philips N.V. Method for generation of attenuation map in PET-MR
US8299438B2 (en) * 2009-07-16 2012-10-30 Siemens Medical Solutions Usa, Inc. Model based estimation of a complete or partial positron emission tomography attenuation map using maximum likelihood expectation maximization
EP2467830B1 (en) * 2009-08-20 2014-10-29 Koninklijke Philips N.V. Reconstruction of a region-of-interest image
US9053569B2 (en) * 2010-11-04 2015-06-09 Siemens Medical Solutions Usa, Inc. Generating attenuation correction maps for combined modality imaging studies and improving generated attenuation correction maps using MLAA and DCC algorithms
EP2668639B1 (en) * 2011-01-27 2016-11-02 Koninklijke Philips N.V. Truncation compensation for iterative cone-beam ct reconstruction for spect/ct systems
DE102011075912A1 (en) * 2011-05-16 2012-11-22 Siemens Ag Method for providing three dimensional image data set of element e.g. screw implanted into biological body of patient, involves changing gray values of subsequent three-dimensional image data set by registering initial image data sets
US9012856B2 (en) * 2011-11-22 2015-04-21 Koninklijke Philips N.V. Gantry-free spect system
US9135695B2 (en) * 2012-04-04 2015-09-15 Siemens Aktiengesellschaft Method for creating attenuation correction maps for PET image reconstruction
CN103961123B (en) * 2013-01-31 2018-11-06 Ge医疗系统环球技术有限公司 Computer tomography(CT)Method and CT system
US11838851B1 (en) 2014-07-15 2023-12-05 F5, Inc. Methods for managing L7 traffic classification and devices thereof
WO2016063211A1 (en) 2014-10-20 2016-04-28 Koninklijke Philips N.V. Classified truncation compensation
US10182013B1 (en) 2014-12-01 2019-01-15 F5 Networks, Inc. Methods for managing progressive image delivery and devices thereof
US9713450B2 (en) * 2014-12-15 2017-07-25 General Electric Company Iterative reconstruction of projection data
US11895138B1 (en) 2015-02-02 2024-02-06 F5, Inc. Methods for improving web scanner accuracy and devices thereof
CN109961491B (en) 2019-04-12 2023-05-26 上海联影医疗科技股份有限公司 Multi-mode image truncation compensation method, device, computer equipment and medium
CN113269733B (en) * 2021-05-14 2024-04-16 成都真实维度科技有限公司 Artifact detection method for radioactive particles in tomographic image

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5210421A (en) * 1991-06-10 1993-05-11 Picker International, Inc. Simultaneous transmission and emission converging tomography
FR2736163B1 (en) * 1995-06-29 1997-08-22 Sopha Medical METHOD FOR OBTAINING, IN NUCLEAR MEDICINE, AN IMAGE OF THE BODY OF A PATIENT CORRECTED IN THE CROWNS
US6539103B1 (en) * 1997-11-12 2003-03-25 The University Of Utah Method and apparatus for image reconstruction using a knowledge set
JPH11298380A (en) * 1998-04-08 1999-10-29 Nec Saitama Ltd Clock generation circuit
IL126761A0 (en) * 1998-10-26 1999-08-17 Romidot Ltd Computerized tomography for non-destructive testing
US6310968B1 (en) * 1998-11-24 2001-10-30 Picker International, Inc. Source-assisted attenuation correction for emission computed tomography
US6490476B1 (en) * 1999-10-14 2002-12-03 Cti Pet Systems, Inc. Combined PET and X-ray CT tomograph and method for using same
GB0128361D0 (en) * 2001-11-27 2002-01-16 British Nuclear Fuels Plc Improvements in and relating to instruments
US20030190065A1 (en) * 2002-03-26 2003-10-09 Cti Pet Systems, Inc. Fast iterative image reconstruction from linograms
US6856666B2 (en) * 2002-10-04 2005-02-15 Ge Medical Systems Global Technology Company, Llc Multi modality imaging methods and apparatus
US6768782B1 (en) * 2002-12-16 2004-07-27 University Of Notre Dame Du Lac Iterative method for region-of-interest reconstruction
US7173248B2 (en) * 2004-10-20 2007-02-06 General Electric Company Methods and systems for positron emission tomography data correction

Also Published As

Publication number Publication date
WO2006064404A3 (en) 2006-08-31
JP4731571B2 (en) 2011-07-27
CN101080651B (en) 2010-06-16
US20090310746A1 (en) 2009-12-17
DE602005020844D1 (en) 2010-06-02
ATE465425T1 (en) 2010-05-15
EP1828808A2 (en) 2007-09-05
JP2008524575A (en) 2008-07-10
CN101080651A (en) 2007-11-28
WO2006064404A2 (en) 2006-06-22
US8013307B2 (en) 2011-09-06

Similar Documents

Publication Publication Date Title
EP1828808B1 (en) Truncation compensation algorithm for iterative reconstruction
US6310968B1 (en) Source-assisted attenuation correction for emission computed tomography
EP1828977B1 (en) Restoration of the nuclear medicine 2d planar image by iterative constrained deconvolution
EP1934942B1 (en) Iterative reconstruction with enhanced noise control filtering
US5565684A (en) Three-dimensional SPECT reconstruction of combined cone-beam and fan-beam data
US6381349B1 (en) Projector/backprojector with slice-to-slice blurring for efficient 3D scatter modeling
EP2210238B1 (en) Apparatus and method for generation of attenuation map
US10126439B2 (en) Reconstruction with multiple photopeaks in quantitative single photon emission computed tomography
US6539103B1 (en) Method and apparatus for image reconstruction using a knowledge set
US7675038B2 (en) Truncation compensation in transmission reconstructions for a small FOV cardiac gamma camera
US6921902B2 (en) Scatter correction device for radiative tomographic scanner
JP4298297B2 (en) Diagnostic imaging apparatus and image processing method
US7890282B2 (en) Estimation of crystal efficiency with axially compressed sinogram
Han Image reconstruction in quantitaive cardiac SPECT with varying focal-length fan-beam collimators
Kalki et al. Myocardial perfusion imaging with a correlated X-ray CT and SPECT system: An animal study
Wong et al. Development of quantitative imaging methods for the GE Hawkeye CT/SPECT system
Liu A comprehensive method for quantitative cardiac SPECT
Laurette et al. Simultaneous 3D correction for attenuation, scatter, and detector response in SPECT
Mao Segmented parallel and slant-hole stationary cardiac single photon emission computed tomography

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20070717

AK Designated contracting states

Kind code of ref document: A2

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LI LT LU LV MC NL PL PT RO SE SI SK TR

DAX Request for extension of the european patent (deleted)
17Q First examination report despatched

Effective date: 20080310

GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

GRAS Grant fee paid

Free format text: ORIGINAL CODE: EPIDOSNIGR3

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LI LT LU LV MC NL PL PT RO SE SI SK TR

REG Reference to a national code

Ref country code: GB

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: CH

Ref legal event code: EP

REG Reference to a national code

Ref country code: IE

Ref legal event code: FG4D

REF Corresponds to:

Ref document number: 602005020844

Country of ref document: DE

Date of ref document: 20100602

Kind code of ref document: P

REG Reference to a national code

Ref country code: NL

Ref legal event code: VDEP

Effective date: 20100421

LTIE Lt: invalidation of european patent or patent extension

Effective date: 20100421

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: SE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100421

Ref country code: NL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100421

Ref country code: ES

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100801

Ref country code: LT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100421

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LV

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100421

Ref country code: AT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100421

Ref country code: FI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100421

Ref country code: IS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100821

Ref country code: SI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100421

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: PL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100421

Ref country code: GR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100722

Ref country code: CY

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100512

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: PT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100823

Ref country code: DK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100421

Ref country code: EE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100421

PLBE No opposition filed within time limit

Free format text: ORIGINAL CODE: 0009261

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CZ

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100421

Ref country code: RO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100421

Ref country code: SK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100421

Ref country code: BE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100421

26N No opposition filed

Effective date: 20110124

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100421

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MC

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20101231

REG Reference to a national code

Ref country code: CH

Ref legal event code: PL

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CH

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20101231

Ref country code: IE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20101205

Ref country code: LI

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20101231

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LU

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20101205

Ref country code: BG

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100421

Ref country code: HU

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20101022

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: TR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100421

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: BG

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20100721

REG Reference to a national code

Ref country code: DE

Ref legal event code: R082

Ref document number: 602005020844

Country of ref document: DE

Representative=s name: MEISSNER, BOLTE & PARTNER GBR, DE

REG Reference to a national code

Ref country code: DE

Ref legal event code: R082

Ref document number: 602005020844

Country of ref document: DE

Representative=s name: MEISSNER, BOLTE & PARTNER GBR, DE

Effective date: 20140328

Ref country code: DE

Ref legal event code: R082

Ref document number: 602005020844

Country of ref document: DE

Representative=s name: MEISSNER BOLTE PATENTANWAELTE RECHTSANWAELTE P, DE

Effective date: 20140328

Ref country code: DE

Ref legal event code: R081

Ref document number: 602005020844

Country of ref document: DE

Owner name: KONINKLIJKE PHILIPS N.V., NL

Free format text: FORMER OWNER: KONINKLIJKE PHILIPS ELECTRONICS N.V., EINDHOVEN, NL

Effective date: 20140328

REG Reference to a national code

Ref country code: FR

Ref legal event code: CD

Owner name: KONINKLIJKE PHILIPS ELECTRONICS N.V., NL

Effective date: 20141126

Ref country code: FR

Ref legal event code: CA

Effective date: 20141126

REG Reference to a national code

Ref country code: FR

Ref legal event code: PLFP

Year of fee payment: 11

REG Reference to a national code

Ref country code: FR

Ref legal event code: PLFP

Year of fee payment: 12

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: FR

Payment date: 20161229

Year of fee payment: 12

REG Reference to a national code

Ref country code: FR

Ref legal event code: ST

Effective date: 20180831

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: FR

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20180102

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: GB

Payment date: 20211221

Year of fee payment: 17

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: DE

Payment date: 20220628

Year of fee payment: 18

GBPC Gb: european patent ceased through non-payment of renewal fee

Effective date: 20221205

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: GB

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20221205

REG Reference to a national code

Ref country code: DE

Ref legal event code: R084

Ref document number: 602005020844

Country of ref document: DE