WO2010035033A1 - Improvements in the field of imaging - Google Patents

Improvements in the field of imaging Download PDF

Info

Publication number
WO2010035033A1
WO2010035033A1 PCT/GB2009/051243 GB2009051243W WO2010035033A1 WO 2010035033 A1 WO2010035033 A1 WO 2010035033A1 GB 2009051243 W GB2009051243 W GB 2009051243W WO 2010035033 A1 WO2010035033 A1 WO 2010035033A1
Authority
WO
WIPO (PCT)
Prior art keywords
region
function
target object
sub
functions
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/GB2009/051243
Other languages
English (en)
French (fr)
Inventor
Paul Roberts
Andrew Maiden
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Phase Focus Ltd
Original Assignee
Phase Focus Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Phase Focus Ltd filed Critical Phase Focus Ltd
Priority to JP2011528427A priority Critical patent/JP5575776B2/ja
Priority to EP09737121.5A priority patent/EP2338144B1/en
Priority to CN200980147396.XA priority patent/CN102227751B/zh
Priority to US13/121,064 priority patent/US9202295B2/en
Publication of WO2010035033A1 publication Critical patent/WO2010035033A1/en
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/003Reconstruction from projections, e.g. tomography
    • G06T11/006Inverse problem, transformation from projection-space into object-space, e.g. transform methods, back-projection, algebraic methods
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/52Devices using data or image processing specially adapted for radiation diagnosis
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N23/00Investigating or analysing materials by the use of wave or particle radiation, e.g. X-rays or neutrons, not covered by groups G01N3/00 – G01N17/00, G01N21/00 or G01N22/00
    • G01N23/20Investigating or analysing materials by the use of wave or particle radiation, e.g. X-rays or neutrons, not covered by groups G01N3/00 – G01N17/00, G01N21/00 or G01N22/00 by using diffraction of the radiation by the materials, e.g. for investigating crystal structure; by using scattering of the radiation by the materials, e.g. for investigating non-crystalline materials; by using reflection of the radiation by the materials
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/12Arrangements for detecting or locating foreign bodies
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T1/00General purpose image data processing
    • G06T1/20Processor architectures; Processor configuration, e.g. pipelining

Definitions

  • the present invention relates to an improved method and apparatus for providing image data from which an image of a target object may be generated.
  • WO 2005/106531 discloses a method and apparatus for producing high resolution images of a region of a target object.
  • incident radiation is caused to fall upon the target object and is scattered as it passes through and beyond the target object, causing a wave function of the incident radiation to be modified in both amplitude and phase with respect to the wave function of the incident radiation at a pre-target side of the object.
  • An array of detectors such as a CCD detector, is arranged a suitable distance from the target object and a diffraction pattern is formed by propagation to a diffraction plane.
  • an aperture is located post- object to select an area of the target object for investigation.
  • a weakly focussing lens is used to illuminate an area of the object.
  • an algorithm is utilised to determine an object function in the form of a two dimensional complex function for the target object.
  • the algorithm is explained in WO 2005/106531 particularly with reference to Figure 7 of that document.
  • the algorithm requires repeated iteration until a predetermined event occurs. For example, the iteration may be repeated 1000 times or until a sum squared error (SSE) is sufficiently small. Repetition of the algorithm represents a significant processing requirement which even on a relatively fast processor takes an appreciable processing time.
  • An aim of embodiments of the present invention is to allow the production of image data in a shorter period of time, from which images, including high resolution images may optionally be produced.
  • a computer-based method for providing image data of a region of a target object comprising the steps of: estimating a first intermediate object function indicating at least one characteristic of a first sub-region of said region of the target object; estimating a second intermediate object function indicating at least one characteristic of a second sub-region of said region of the target object; and combining the first and second intermediate object functions to form a combined object function indicating at least one characteristic of the region of the target object.
  • a method of providing image data of a region of a target object without high resolution positioning of incident radiation or a post target object aperture relative to the target object comprising the steps of: providing incident radiation, from a radiation source, at the target object; via at least one detector, detecting the intensity of radiation scattered by said target object; responsive to the detected intensity, estimating first and second intermediate object functions indicating at least one characteristic of first and second sub-regions, respectively, of said region of the target object; and combining the first and second object functions to form a combined object function indicating at least one characteristic of the region of the target object.
  • an apparatus for providing image data relating to a region of a target object comprising: a first computational unit arranged to estimate a first intermediate object function indicating at least one characteristic of a first sub-region of said region of the target object; and a second computational unit estimating a second intermediate object function indicating at least one characteristic of a second sub-region of said region of the target object; wherein a computational unit is arranged to combine the first and second object functions to form a combined object function indicating at least one characteristic of the region of the target object.
  • an apparatus for providing image data for generating an image of a region of a target object comprising: locating means for locating a target object at a predetermined location; a radiation source for providing incident radiation at a target object located by said locating means; at least one detector device for detecting an intensity of radiation scattered by said target object; radiation locating means for locating incident radiation or an aperture, located post target object, at one or more locations with respect to said target object; first processing means for determining a first intermediate object function responsive to a detected intensity of scattered radiation from a sub-region of the target object comprising one or more areas selected by the radiation locating means; and second processing means for determining a second intermediate object function responsive to a detected intensity of scattered radiation from a sub-region of the target object comprising one or more areas selected by the radiation locating means; wherein one of the first or second processing means, or a third processing means, is arranged to combine the first and second object functions to form a combined object function indicating at least one
  • Embodiments of the present invention reduce an object function computation time by employing parallel processing.
  • Two or more computational units are each utilised to compute an intermediate object function.
  • a computational unit After computation of the intermediate object functions a computational unit combines the intermediate object functions to form an estimate of the object function for that iteration.
  • the two or more processing units utilise the previous estimate of the object function to again form intermediate object functions, which are then combined by a computational unit to form the following iteration's estimate of the object function.
  • parallel processing of intermediate object functions is utilised.
  • a disparity between object functions estimated for individual regions does not occur, or at least is minimised, by frequent combination of intermediate object function estimates.
  • Figure 1 shows a moving aperture arrangement known from the prior art
  • Figure 2 shows an illustration of moving a post-target aperture known from the prior art
  • FIG. 3 shows illustrates a prior art phase retrieval algorithm
  • Figure 4 shows an illustration of a first embodiment of a method according to the present invention
  • Figure 5 schematically illustrates four aperture positions
  • Figure 6 schematically illustrates an arrangement of computational units of an embodiment of the present invention.
  • Figure 7 shows an illustration of a second embodiment of a method according to the present invention.
  • WO 2005/106531 Before discussing embodiments of the present invention, a brief introduction to a prior art apparatus, method and algorithm as disclosed in WO 2005/106531, which is incorporated herein in its entirety by reference for all purposes, will be provided.
  • the embodiment of the prior art discussed is a moving aperture arrangement as disclosed in WO 2005/106531.
  • a prior art method of moving a weakly focussing lens is also known and that embodiments of the present invention may also be used in conjunction with such an arrangement of weakly focussing lens.
  • embodiments of the present invention may also be used with a pre-target aperture to illuminate an area of the object. Referring to Figure 1, a prior art arrangement is shown in which incident radiation 30 is caused to fall upon a target object 31.
  • target object refers to any specimen or item placed in the path of incident radiation which causes scattering of that radiation. It will be understood that the target object should be at least partially transparent to incident radiation. The target object may or may not have some repetitive structure.
  • radiation is to be broadly construed as energy from a radiation source. This will include electro magnetic radiation including
  • X-rays emitted particles such as electrons and/or acoustic waves.
  • Such radiation may be represented by a wave function ⁇ (r).
  • This wave function includes a real part and an imaginary part as will be understood by those skilled in the art. This may be represented by the wave functions modulus and phase.
  • the incident radiation 30 is scattered as it passes through and beyond the target object 31.
  • the wave function of the incident radiation as it exits the target object 31 will be modified in both amplitude and phase with respect to the wave function of the incident radiation at the pre-target side of the target object 31.
  • the scattering which occurs may include Fourier diffraction, refraction and/or Fresnel diffraction and any other form of scattering in which characteristics of the incident radiation are modified as a result of propagating after the target object 31. If an array of detectors such as a CCD detector 32 is arranged a long distance from the target object 31 then a diffraction pattern is formed at a diffraction plane 33.
  • a Fourier diffraction pattern will form if the detectors 32 are located a distance D from the target object 31 where D is sufficiently long for the diffraction pattern to be formed effectively from a point source. If the diffraction plane is formed closer to the target object 31, by locating the detectors nearer, then a Fresnel diffraction pattern will be formed.
  • An aperture 34 is located post target object to thereby select a region of the target for investigation.
  • the aperture is formed in a mask so that the aperture defines a "support".
  • a support is an area of a function where that function is not zero. In other words outside the support the function is zero. Outside the support the mask blocks the transmittance of radiation. Apertures for use with the present invention need not be finite and sharply defined.
  • aperture describes a localised transmission function of radiation. This may be represented by a complex variable in two dimensions having a modulus value between 0 and 1.
  • An example is a mask having a physical aperture region of varying transmittance.
  • FIG. 2 illustrates schematically the propagation of waves through the arrangement of Figure 1.
  • Incident radiation 30 falls upon the up-stream side of the target object 31 and is scattered by the target object 31 as it is transmitted.
  • a target object wave O(r) is an exit wave function of radiation after interaction with the target object 31.
  • O(r) represents a two-dimensional complex function so that each point in O(r), where r is a two-dimensional coordinate, has associated with it a complex number.
  • O(r) will physically represent an exit wave that would emanate from the target object 31 which is illuminated by a plane wave. For example, in the case of electron scattering, O(r) would represent the phase and amplitude alteration introduced into an incident wave as a result of passing through the target object 31 of interest.
  • the aperture 34 provides a probe function P(r) (or filtering function) which selects a part of the object exit wave function for analysis. It will be understood that rather than selecting an aperture a transmission grating or other such filtering function may be located downstream of the object function.
  • the probe function P(r-R) is an aperture transmission function where an aperture is at a position R.
  • the probe function can be represented as a complex function with its complex value given by a modulus and phase which represent the modulus and phase alterations introduced by the probe into a perfect plane wave incident up it.
  • the exit wave function ⁇ (r,R) 43 is an exit wave function of radiation as it exits the aperture 34.
  • This exit wave ⁇ (r,R) 43 forms a diffraction pattern ⁇ (k,R) 44 at a diffraction plane 33.
  • r is a vector coordinate in real space and k is a vector coordinate in diffraction space.
  • Figure 3 illustrates a prior art algorithm for obtaining a wave function of an object and thus for obtaining image data which may be used subsequently to generate high resolution images of an object.
  • Figure 3 illustrates a method using the arrangement illustrated in Figures 1 and 2 and moving the aperture from a first position after measuring the diffraction pattern to a second position where a second respective diffraction pattern may be measured.
  • O(r) and P(V) represent two-dimensional complex functions, that is, each point in O(r) or P(r), where r is a two-dimensional coordinate, has associated with it a complex number.
  • O(r) will physically represent an exit wave that would emanate from an object function which is illuminated by a plane wave. For example, in the case of electron scattering, O(r) would represent the phase and amplitude alteration into an incident wave as a result of passing through the object of interest.
  • P(V) represents either an illumination function, such as that generated by a caustic or illumination profile formed by a lens or other optical component or a filtering function, such as an aperture or transmission grating mounted downstream of the object function.
  • O(r) or P(r) can be moved relative to one another by various distances R.
  • the nomenclature adopted is written in terms of moving P(r), although equivalently we could instead move O(r) relative to P(r).
  • the complex value of O(r) is altered by forming the product of O(r) with P(r-R) to give a total exit wave function of ⁇ (r), i.e.
  • the algorithm works to find the phase and intensity of the complex function ⁇ (r,R). It requires as input knowledge of the function P(r-R), and one or more (preferably several) measurements of the intensity of the wave function in a plane which is different to that containing the target object 31. It is convenient to use the diffraction plane, which is related to the specimen plane by the Fourier transform. In this case the measured input data is the intensities of the diffraction patterns at one or more probe/aperture positions. However it is also possible to run the algorithm based on a set of defocused images measured at some distance from the exit surface of the specimen/aperture. In this situation the free space propagator is substituted for the Fourier transform. The algorithm is not restricted to use of these two transforms. Other effective transforms could be used to move from one plane of information to the other. In what follows a general transform T is referred to that transforms a wave function from the first plane, called plane 1, to the second plane, called plane 2.
  • step S700 with a guess at the object function Og,n(r), where the subscript g,n represents a guessed wave at the nth iteration of the algorithm.
  • Og,n(r) the object function
  • These functions are in plane 1 (which is the real space plane if the Fourier transform is used).
  • the first guess of Og,n(r) may equal unity at all points r. This corresponds to an absent specimen.
  • Og,n(r) may be set to a random values at each point.
  • a known aperture in terms of position and characteristics is selected at step S701. This provides a probe function P(r-R).
  • the current guess at the object function is multiplied by the aperture or probe at the current position R, P(r-R). This produces the guessed exit wave function (still in plane 1) for position R,
  • step S703 a transformation of ⁇ g,n(r,R) to obtain the corresponding wave function in plane 2 (which would be the diffraction space plane if the Fourier transform is used), for that position R.
  • T is used to represent some general transform that would often be the Fourier transform, but could also be the Fresnel free space propagator, or some other transform suited to a particular application of the algorithm.
  • k is the coordinate in plane 2.
  • k would be the usual reciprocal space coordinate.
  • k would be the xy coordinate in the defocused plane.
  • ⁇ g,n(k,R) is a "guessed" version of the actual wave function in plane 2, since it has been produced by the guessed object function Og,n(r). Successive iterations of the algorithm will produce increasingly accurate versions of ⁇ g,n(k,R). Note that ⁇ g,n(k,R) can be written in the form:
  • the intensity of the diffraction pattern By measuring the intensity of the diffraction pattern by known techniques such as detector array 32 information about the actual transformed exit wave function are known. A measured intensity of the diffraction pattern where the aperture is in a first position thus forms the basis of an estimate of the complex wave function of the diffraction pattern. However the measured intensity does not provide information about the phase of the wave function. Rather the measured intensity is comparable to the squared modulus of ⁇ (r). That is
  • ⁇ c,nM T 1 [ ⁇ c,n( ⁇ )].
  • the update function helps make the effective deconvolution that occurs possible and introduces a weighting factor which causes the object function to be updated most strongly where the probe function has largest amplitude.
  • the selectable constant ⁇ may be set to 1. It may be selected as any value in the range of 0 to 3 and need not be an integer value. It is useful to set ⁇ >1 when there is much noise. ⁇ may be selected ⁇ ⁇ when because of scattering geometry, the detected intensity is of the form of a Gabor hologram or similar. The value ⁇ is used to prevent a divide-by-zero occurring if
  • P(r - R) ⁇ 0.
  • is a small real number as is commonly applied in Weiner Filters and is usually (though not necessarily) smaller than Pmax and can be considerably smaller if the noise present in the recorded data is small.
  • is a parameter which may be set at a fixed value or which may vary. It indicates how noisy the recorded data is and is used to attenuate how the updating is carried out in response to these circumstances. If good conditions exist for data collection that is to say with high beam current (high flux), which would imply low shot-noise, then it is safer to use results gathered to update the guessed estimate. Consequently the value of ⁇ can be a small fraction of Pmax (e.g. less than 1/1 Oth).
  • the algorithm has some similarities to the well known Fienup algorithm. If in this situation, only one position R is used, then the algorithm reduces to being mathematically identical to the basic Fienup algorithm. Where more than one position R is used, the algorithm has considerable advantages over known methods, including the fact that it does not suffer from uniqueness issues, and that a wider field of view may be imaged.
  • the algorithm shown in Figure 7 progresses to selecting a new position R which at least in part overlaps the previous position.
  • the overlap should preferably be more than 20% and is preferably
  • step S708 the algorithm returns to step S702. Instead of the initial estimate of the object function O(r) being loaded in the new guess for O(r) of step S708 is loaded in. On each iteration the new guess for the object function will approximate closer and closer to the actual object function as on each iteration information of the known intensity and thus the known amplitude component of the incident radiation is added to improve the accuracy of the estimate.
  • a known probe function P(r-i?2) at the second position is identified at step S709 and then the step as above mentioned are repeated so that the new guess generated in step S708 is multiplied with the new known probe function identified at step S709. This is illustrated in step S710. Effectively this generates an exit wave function either post specimen or post aperture depending upon the embodiment concerned.
  • the resulting exit wave function is propagated at step S711 to provide an estimate of the scattering pattern which should be detected at that position.
  • the diffraction pattern is measured at step S712 which provides intensity information and thus amplitude information about the transformed wave function. The intensity information is used to correct the amplitude of the transformed wave function whilst phase information is retained at step S713.
  • This corrected wave function is inversely propagated via Fourier transformation (when the image is formed in the far field), Fresnel transformation when the image is formed at a location where Fresnel diffraction dominates or by any other suitable transformation. This is illustrated at step S714.
  • the running estimate of O(r) is then corrected according to the update function shown above at step S715 and the result is a new guess for the object function illustrated in step S716.
  • the illumination or aperture may be made to a third or further position. Again a location where some overlap occurs between previous illuminated locations is preferable. In this way the whole target object may optionally be mapped.
  • the new guess generated at step S716 may be repeated without further positioning knowing known diffraction pattern results.
  • the iterative method is illustrated as being repeated by returning to step S702 in which the new guess generated at step S716 is input to the multiplication stage rather than the initial estimate of the object function supplied at step S700.
  • the iterative method may be repeated until a predetermined event occurs.
  • the iteration may be repeated a predetermined number of times, for example 1000 times or until the sum squared error (SSE) is sufficiently small.
  • SSE is measured in plane 2, as
  • N is the number of pixels in the array representing the wave function.
  • the most up-to-date guess of the object function provides a running estimate for that object function.
  • the running estimate of the object function provides image data at the locations which are either illuminated by the incident radiation or which are selected by location of a post target object aperture. This image data includes amplitude and phase information which can subsequently be used to generate a high resolution image of the selected region of the target object.
  • Embodiments of the algorithm according to the present invention operate on diffraction patterns formed with an aperture at different n positions Ri... R n , where n is two or more. It will also be realised that embodiments of the algorithm may also be applied to diffraction patterns formed by radiation incident upon different regions of a target object 31. As noted above, each aperture position should at least in part overlap a previous aperture position, preferably by at least 20% or more and more preferably by at least 50%.
  • the method begins in step 401.
  • the n aperture positions: Ri... R n are partitioned into, or allocated amongst, m regions: S]...S m . It is not necessary to partition the m regions so that each contains the same number of aperture positions, although doing so will lead to the fastest overall computational time.
  • the m regions may contain each aperture position exactly once, i.e. the union of all m regions is exactly R] ...R n .
  • FIG. 5 shows an illustration of the four regions 501-504, wherein and It will be noted that the shape and arrangement of the four regions is merely an example and their degree of overlap is merely shown for illustration.
  • the four aperture positions R]... R4 are divided into the two regions Si, S2 such that Si comprises R] and R3, whilst S2 comprises aperture positions R 2 and R4.
  • the aperture positions have been arbitrarily assigned to the regions.
  • Si could comprise aperture positions Ri and R4 and no inter-relationship, such as a positional relationship, between regions and aperture positions is required.
  • step 403 a weighting function w ⁇ (r-R) pertaining to the probe function at different aperture positions is calculated for each aperture position.
  • a weighting function for each region Ws(V, S n ) is then calculated. The purpose of weighting is to cause the object function to be updated most strongly where the probe function has the largest amplitude i.e. carries more information.
  • R R, R x
  • w(r-R) is the individual weighting factor for a particular aperture position R and is given by:
  • weighting factor is identical in definition and meaning to that found in the prior art, wherein the weighting factor is not a scalar variable but instead causes the object function to be updated most strongly where the probe function has largest amplitude.
  • the weighting factor is essentially a normalised version of the probe function modified by ⁇ and / as shown above.
  • a normalised weighted probe function at all aperture positions is then calculated by:
  • Equations 10-12 used to calculate aperture position weighting functions may be pre-calculated apriori. That is, the weighting functions may be determined in advance before the algorithm is desired to begin, thus saving computation time. In the first embodiment, Equations 13-15 may also be calculated in advance since the allocation of aperture positions amongst regions does not change during iterations of the algorithm.
  • step 404 calculation of an intermediate object function for each region is assigned to a computation unit selected from amongst two or more computational units.
  • intermediate object functions for each region will be repeatedly calculated. Whilst it may be more convenient to assign the calculation of an intermediate object function to a particular computational unit throughout the repeated iterations, the algorithm is not constrained in this way. That is, a first computational unit may calculate an intermediate object function for region Si during a first iteration and the calculation of an intermediate objection function for region Si during a second iteration may be performed by a second computational unit.
  • computational unit is understood to mean a unit capable of performing independent calculations.
  • an individual processor may represent a computational unit.
  • a core of a multi-core processor may also represent a computational unit.
  • a combination of cores in a multi-core processor or individual processor architectures may also be used.
  • the computational units may be local to one-another i.e. connected to a bus interconnect, or may be remotely located and connected together by a wired or wireless communications medium, such as a network e.g. LAN or WAN network.
  • each computational unit calculates an intermediate object function for a respective region.
  • one or more computational units may calculate intermediate object functions for more than one region, depending upon a speed of execution of those computational units. It will be understood that it would be preferable to balance computing load across the computational units such that each computational unit finishes calculating intermediate objection functions for the one or more regions which it has been allocated at approximately the same time as the other computational units. This allows computational units having different speeds of execution to be used in combination.
  • FIG. 6 an arrangement is shown in Figure 6 comprising a central computational unit and two slave computational units 602, 603.
  • the central computational unit is responsible for performing steps 402-404 of the method illustrated in Figure 4, whilst each of the slave computational units 602, 603 calculates in parallel an intermediate object function for each region Si, S2 of the described example.
  • the designations "central” and “slave” do not imply any functional inter-dependency of the slave computational units 602, 603 on the central computational unit 601 or any performance hierarchy amongst the computational units.
  • the central computational 601 unit may be identical to the slave computational units 602, 603.
  • the central computation unit 601 may calculate an intermediate object function for one region and then calculate the combined object function using an intermediate objection calculated for the other region by one of the slave computational units 602, thus reducing the number of slave computational units required.
  • step 405 computational units calculate in parallel intermediate object function guesses for all of the regions.
  • WO 2005/106531 For each region, the algorithm of WO 2005/106531 is applied which comprises starting with an object function guess and a known probe function and multiplying the two to form a guessed exit wave function in first plane; transforming that exit wave function to form an exit wave function in a second plane which is then corrected against measured intensities of the wave function in the second plane before inversely transforming the function to obtain an improved guess of the exit wave function in the first plane as explained with reference to steps 700-706 in particular.
  • the reader is referred to Figure 3 and WO 2005/106531.
  • WO 2005/106531 gives the equation used to update an object function guess for aj+l th iteration as:
  • an intermediate object function guess after j steps for aperture positions in the m th region O g m j is calculated in the same way as shown in Figure 3 for aperture positions in the m th region.
  • the object function may equal unity at all points r, or may be randomly chosen at each point. Other initial values may also be used. Further, it will be noted that the initial object function for each region may be equal, or may be different in value.
  • the intermediate object function O g m j is updated for the m th region using:
  • n/m is not an integer then a maximum value of j may be chosen appropriately.
  • one region would necessarily comprise 3 aperture positions, whilst the other comprises 2 aperture positions.
  • 7 would need to assume values of 1, 2 and 3, whilst for the other region y would need to assume values of 1 and 2. Therefore, the maximum value of j for each region should be chosen appropriately.
  • step 405 intermediate object function guesses for all of the regions have been calculated.
  • O g n is the intermediate object function guess for region 1 and
  • step 406 the intermediate object function guesses are combined to produce an object function guess O g l for iteration i.
  • step 407 it is determined whether the algorithm has been iterated a sufficient number of times.
  • the algorithm may proceed for a pre-determined number of iterations, for example 1000, or alternatively can be terminated when the sum squared error (SSE) is sufficiently small.
  • SSE sum squared error
  • the SSE can be computed as:
  • N is the numbers of pixels in O. Testing has shown that the above-described prior art solution requiring 85 seconds per iteration has been reduced using certain embodiments of the present invention to 5 second per iteration when using an Intel (RTM) Core2 PC.
  • aperture positions are reassigned amongst regions during iterations of computation of the object function.
  • the assignment of aperture positions does not have to be fixed for all iterations of the algorithm.
  • the assignment of aperture positions may be changed after 50% of the iterations have been performed, or after every 10% or 25% of the iterations. More preferably, the aperture positions may be reassigned amongst the regions for every iteration of the algorithm.
  • the assignment of aperture positions to regions may be performed on a predetermined or random basis.
  • the method begins in step 701 and in step 702 a weighting function for each aperture position is calculated.
  • the weighting function for each aperture may be calculated in advance, as previously discussed.
  • step 703 an initial partitioning of aperture positions amongst a plurality of regions is performed. The initial partitioning may be predetermined, in which case it would allow the calculation of initial weighting functions for each region to also be calculated in advance.
  • step 705 calculation of intermediate objection functions is assigned amongst computational units, as in step 404 of the first embodiment.
  • step 706 a first calculation of an intermediate object function is performed in parallel for each of the regions.
  • the intermediate object functions are then combined in step 707 to form an object function for the first iteration.
  • step 707 the aperture positions are reassigned amongst the regions.
  • region Sl may comprise aperture positions Ri and R3 in the first iteration, which is then changed in step 708 to comprise aperture positions Ri and R4 in the second iteration.
  • weighting functions for the regions are recalculated in step 709. If a predetermined reassignment of aperture positions to regions is utilised, the reassigned weighting functions may be calculated in advance and stored in memory for use in step 709.
  • step 710 intermediate object functions are calculated for each of the regions, which are then combined in step 711 to form an object function guess for that iteration.
  • step 712 it is determined whether enough iterations have been performed in the same way as in step 407 in the first embodiment.
  • Figure 7 shows aperture positions being reassigned amongst regions for every iteration, the second embodiment is not constrained in this way and reassignment may be performed as desired.
  • An advantage of the calculation of intermediate object functions for each region which are then combined to produce an object function for that iteration is that the a level of interdependency between intermediate object functions and a combined object function is introduced, reflecting the actual interdependence between regions in the real object. For example, if intermediate object functions for each region were calculated by 1000 iterations of equation 18, then the intermediate object functions for each region combined to produce a final object function, disparities, seen as, for example, a large difference between object functions for each region could exist. However, frequent combination of the intermediate object functions for to produce a combined object function reduces such differences. That is, the intermediate object functions are combined and the combination used in a subsequent iteration for calculating each intermediate object function, thus avoiding each intermediate function being created in isolation.
  • embodiments of the present invention can be realised in the form of hardware, software or a combination of hardware and software. Any such software may be stored in the form of volatile or non- volatile storage such as, for example, a storage device like a ROM, whether erasable or rewritable or not, or in the form of memory such as, for example, RAM, memory chips, device or integrated circuits or on an optically or magnetically readable medium such as, for example, a CD, DVD, magnetic disk or magnetic tape. It will be appreciated that the storage devices and storage media are embodiments of machine-readable storage that are suitable for storing a program or programs that, when executed, implement embodiments of the present invention.
  • embodiments provide a program comprising code for implementing a system or method as claimed in any preceding claim and a machine readable storage storing such a program. Still further, embodiments of the present invention may be conveyed electronically via any medium such as a communication signal carried over a wired or wireless connection and embodiments suitably encompass the same.

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Medical Informatics (AREA)
  • Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • Theoretical Computer Science (AREA)
  • Chemical & Material Sciences (AREA)
  • Immunology (AREA)
  • Algebra (AREA)
  • Pure & Applied Mathematics (AREA)
  • Analytical Chemistry (AREA)
  • Biochemistry (AREA)
  • Mathematical Physics (AREA)
  • Mathematical Optimization (AREA)
  • Mathematical Analysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Biophysics (AREA)
  • High Energy & Nuclear Physics (AREA)
  • Crystallography & Structural Chemistry (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Optics & Photonics (AREA)
  • Radiology & Medical Imaging (AREA)
  • Biomedical Technology (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Molecular Biology (AREA)
  • Surgery (AREA)
  • Animal Behavior & Ethology (AREA)
  • Public Health (AREA)
  • Veterinary Medicine (AREA)
  • Analysing Materials By The Use Of Radiation (AREA)
  • Investigating Or Analysing Materials By Optical Means (AREA)
PCT/GB2009/051243 2008-09-26 2009-09-23 Improvements in the field of imaging Ceased WO2010035033A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
JP2011528427A JP5575776B2 (ja) 2008-09-26 2009-09-23 イメージングの分野における改良
EP09737121.5A EP2338144B1 (en) 2008-09-26 2009-09-23 Method and apparatus for diffraction imaging
CN200980147396.XA CN102227751B (zh) 2008-09-26 2009-09-23 提供目标对象区域图像数据的方法和装置
US13/121,064 US9202295B2 (en) 2008-09-26 2009-09-23 Field of imaging

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
GBGB0817650.5A GB0817650D0 (en) 2008-09-26 2008-09-26 Improvements in the field of imaging
GB0817650.5 2008-09-26

Publications (1)

Publication Number Publication Date
WO2010035033A1 true WO2010035033A1 (en) 2010-04-01

Family

ID=40019607

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/GB2009/051243 Ceased WO2010035033A1 (en) 2008-09-26 2009-09-23 Improvements in the field of imaging

Country Status (6)

Country Link
US (1) US9202295B2 (enExample)
EP (1) EP2338144B1 (enExample)
JP (1) JP5575776B2 (enExample)
CN (1) CN102227751B (enExample)
GB (1) GB0817650D0 (enExample)
WO (1) WO2010035033A1 (enExample)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2481589A (en) * 2010-06-28 2012-01-04 Phase Focus Ltd Calibration of a probe in Ptychography
WO2012071448A1 (en) * 2010-11-24 2012-05-31 Omnyx LLC Digital microscopy with focus grading in zones distinguished for comparable image structures
WO2013164645A1 (en) * 2012-05-03 2013-11-07 Phase Focus Limited Improvements in providing image data
US9029745B2 (en) 2010-12-03 2015-05-12 Phase Focus Limited Method and apparatus for providing image data
US9274024B2 (en) 2012-01-24 2016-03-01 Phase Focus Limited Method and apparatus for determining object characteristics
US9322791B2 (en) 2012-08-31 2016-04-26 Phase Focus Limited Phase retrieval from ptychography
US9448160B2 (en) 2011-04-27 2016-09-20 Phase Focus Limited Method and apparatus for providing image data for constructing an image of a region of a target object
WO2017085519A1 (en) * 2015-11-19 2017-05-26 Phase Focus Limited Improved method ptychographic detector mapping

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB0822149D0 (en) * 2008-12-04 2009-01-14 Univ Sheffield Provision of image data
FR3001544B1 (fr) * 2013-01-31 2015-02-27 Commissariat Energie Atomique Procede de reglage de la position relative d'un analyte par rapport a un faisceau lumineux
CN107924132B (zh) 2014-08-28 2021-02-12 Asml荷兰有限公司 检查设备、检查方法和制造方法
US11362481B2 (en) 2020-05-01 2022-06-14 Mesa Photonics, LLC Method and apparatus for measuring optical pulses
US12367563B2 (en) * 2021-03-25 2025-07-22 Government Of The United States Of America, As Represented By The Secretary Of Commerce Tomographic reconstruction apparatus and removing diffraction effects in a tomographic image

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2005106531A1 (en) * 2004-04-29 2005-11-10 University Of Sheffield High resolution imaging

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6005916A (en) * 1992-10-14 1999-12-21 Techniscan, Inc. Apparatus and method for imaging with wavefields using inverse scattering techniques
AU716800B2 (en) * 1996-12-24 2000-03-09 Xrt Limited Phase retrieval in phase contrast imaging
US7295691B2 (en) * 2002-05-15 2007-11-13 Ge Medical Systems Global Technology Company, Llc Computer aided diagnosis of an image set
KR101028354B1 (ko) * 2007-03-20 2011-06-14 주식회사 메디슨 초음파 영상을 형성하는 초음파 시스템 및 방법
GB0709796D0 (en) * 2007-05-22 2007-06-27 Phase Focus Ltd Three dimensional imaging

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2005106531A1 (en) * 2004-04-29 2005-11-10 University Of Sheffield High resolution imaging

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
THIBAULT P ET AL: "High-resolution scanning X-ray diffraction microscopy", SCIENCE AMERICAN ASSOCIATION FOR THE ADVANCEMENT OF SCIENCE USA, vol. 321, no. 5887, 18 July 2008 (2008-07-18), pages 379 - 382, XP002564021, ISSN: 0036-8075 *
THIBAULT P ET AL: "Supporting Online Material for High-Resolution Scanning X-ray Diffraction Microscopy", 18 July 2008 (2008-07-18), pages 1 - 10, XP007911178, Retrieved from the Internet <URL:http://www.sciencemag.org/cgi/data/321/5887/379/DC1/1> [retrieved on 20100118] *

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8942449B2 (en) 2010-06-28 2015-01-27 Phase Focus Limited Calibration of a probe in ptychography
GB2481589A (en) * 2010-06-28 2012-01-04 Phase Focus Ltd Calibration of a probe in Ptychography
GB2481589B (en) * 2010-06-28 2014-06-11 Phase Focus Ltd Calibration of a probe in ptychography
WO2012071448A1 (en) * 2010-11-24 2012-05-31 Omnyx LLC Digital microscopy with focus grading in zones distinguished for comparable image structures
US9121764B2 (en) 2010-12-03 2015-09-01 Phase Focus Limited Providing image data
US9029745B2 (en) 2010-12-03 2015-05-12 Phase Focus Limited Method and apparatus for providing image data
US9448160B2 (en) 2011-04-27 2016-09-20 Phase Focus Limited Method and apparatus for providing image data for constructing an image of a region of a target object
US9274024B2 (en) 2012-01-24 2016-03-01 Phase Focus Limited Method and apparatus for determining object characteristics
US9784640B2 (en) 2012-01-24 2017-10-10 Phase Focus Limited Method and apparatus for determining object characteristics
WO2013164645A1 (en) * 2012-05-03 2013-11-07 Phase Focus Limited Improvements in providing image data
US10466184B2 (en) 2012-05-03 2019-11-05 Phase Focus Limited Providing image data
US9322791B2 (en) 2012-08-31 2016-04-26 Phase Focus Limited Phase retrieval from ptychography
WO2017085519A1 (en) * 2015-11-19 2017-05-26 Phase Focus Limited Improved method ptychographic detector mapping

Also Published As

Publication number Publication date
EP2338144B1 (en) 2017-01-11
US20110235862A1 (en) 2011-09-29
GB0817650D0 (en) 2008-11-05
CN102227751A (zh) 2011-10-26
JP2012510608A (ja) 2012-05-10
JP5575776B2 (ja) 2014-08-20
EP2338144A1 (en) 2011-06-29
US9202295B2 (en) 2015-12-01
CN102227751B (zh) 2015-03-04

Similar Documents

Publication Publication Date Title
EP2338144B1 (en) Method and apparatus for diffraction imaging
KR101810637B1 (ko) 티코그래피에서 프로브의 보정
JP4926944B2 (ja) 高分解能撮像
JP6012710B2 (ja) 対象物についてのある領域の画像を構築するための画像データを提供する方法及び装置
JP6556623B2 (ja) 位相回復の改善
US9401042B2 (en) Method and apparatus for imaging a three dimensional target object using incident radiation
JP4594114B2 (ja) 画像処理装置および屈折率分布測定装置
EP2227705B1 (en) Method and apparatus for providing image data
US12266162B2 (en) Enhancing contrast sensitivity and resolution in a grating interferometer by machine learning
JPH08110749A (ja) 位相格子の設計システム
EP2732274B1 (en) Method and apparatus for position determination
KR20250174455A (ko) 다중 차수 모드 신호를 이용한 x-반사 분석 시스템 및 x-반사 분석 방법
US20180328867A1 (en) Improved method ptychographic detector mapping

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 200980147396.X

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 09737121

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 2011528427

Country of ref document: JP

REEP Request for entry into the european phase

Ref document number: 2009737121

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2009737121

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 13121064

Country of ref document: US