US20110262043A1 - Pattern matching method and image processing device - Google Patents
Pattern matching method and image processing device Download PDFInfo
- Publication number
- US20110262043A1 US20110262043A1 US13/122,151 US200913122151A US2011262043A1 US 20110262043 A1 US20110262043 A1 US 20110262043A1 US 200913122151 A US200913122151 A US 200913122151A US 2011262043 A1 US2011262043 A1 US 2011262043A1
- Authority
- US
- United States
- Prior art keywords
- pattern
- image
- template
- pattern matching
- matching
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01J—ELECTRIC DISCHARGE TUBES OR DISCHARGE LAMPS
- H01J37/00—Discharge tubes with provision for introducing objects or material to be exposed to the discharge, e.g. for the purpose of examination or processing thereof
- H01J37/26—Electron or ion microscopes; Electron or ion diffraction tubes
- H01J37/28—Electron or ion microscopes; Electron or ion diffraction tubes with scanning beams
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
- G06T7/74—Determining position or orientation of objects or cameras using feature-based methods involving reference images or patches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/69—Microscopic objects, e.g. biological cells or cellular parts
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01J—ELECTRIC DISCHARGE TUBES OR DISCHARGE LAMPS
- H01J37/00—Discharge tubes with provision for introducing objects or material to be exposed to the discharge, e.g. for the purpose of examination or processing thereof
- H01J37/02—Details
- H01J37/22—Optical or photographic arrangements associated with the tube
- H01J37/222—Image processing arrangements associated with the tube
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10056—Microscopic image
- G06T2207/10061—Microscopic image from scanning electron microscope
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30108—Industrial image inspection
- G06T2207/30148—Semiconductor; IC; Wafer
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/12—Acquisition of 3D measurements of objects
- G06V2201/122—Computational image acquisition in electron microscopy
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01J—ELECTRIC DISCHARGE TUBES OR DISCHARGE LAMPS
- H01J2237/00—Discharge tubes exposing object to beam, e.g. for analysis treatment, etching, imaging
- H01J2237/245—Detection characterised by the variable being measured
- H01J2237/24592—Inspection and quality control of devices
Definitions
- the present invention relates to a pattern matching method and device and, more particularly, to a method of matching for a pattern formed over plural layers and to a device implementing the matching.
- SEMs scanning electron microscopes
- SEMs include equipment known as CD (critical-dimension)-SEM, and in this kind of equipment positioning is performed while relying on pattern matching when patterns on semiconductor wafers or masks placed on stages (sample stages) are measured.
- the pattern matching method is a technique of causing a template representing a unique pattern formed on a sample to search within a given image region, to identify locations having high degrees of match with the template. Since a relative positional relationship between a pattern identified by the pattern matching and a pattern subjected to a measurement is previously known, it is possible to quickly perform positioning relative to the minute pattern subjected to the measurement. Furthermore, in some cases, subjects of measurement themselves are taken as matching patterns.
- Patent Literature 1 a pattern matching apparatus has been proposed that is designed to perform pattern matching in which a hole pattern is subjected to pattern matching and in which design data about a semiconductor pattern is taken as a template.
- a proper degree of match is required between an SEM image subjected to matching and a template.
- a composite shape of its upper and lower portions is a pattern shape represented on an SEM image and so its pattern deformations have more variations than planar patterns.
- the lower portions of hole patterns are deformed to a larger extent.
- the orbit of electrons emitted from the bottoms of the holes is affected by electric charge adhering to the sample surface (electrification). The yield of the detected secondary electrons may be varied.
- a method for the purpose of achieving stable positioning or selective information extraction on a certain layer regardless of the circumstances of the depths of a pattern formed in three dimensions or of the circumstances of charging of a sample and a device for achieving the purpose are hereinafter described.
- a method of performing first pattern matching on image information representing plural layers by the use of a first template selectively representing the shape of an upper portion of a pattern is proposed, as well as a device. Furthermore, in order to perform selective extraction on a certain layer, a method of subtracting information about the shape of the upper portion of the pattern from the image information representing the plural layers or from the shape information is proposed, as well as a device.
- the above-described structure makes it possible to perform pattern matching at a quite high success rate regardless of variations in information obtained from the lower portion of the pattern. Furthermore, selective extraction of a pattern in a lower-level layer is enabled.
- FIG. 1 is a view illustrating one example of processing for extracting shapes of a lower portion of a pattern, based on pattern matching using a template selectively representing an upper portion of the pattern;
- FIG. 2 is a view illustrating another example of processing for extracting shapes of a lower portion of a pattern, based on pattern matching using a template selectively representing an upper portion of the pattern;
- FIG. 4 is a diagram schematically illustrating an SEM system
- FIG. 5 is a chart illustrating one example of template matching processing
- FIG. 6 is a diagram illustrating one example of pattern matching processing
- FIG. 7 is a diagram illustrating one example of two-stage matching for an upper layer pattern and a lower layer pattern.
- pattern matching is a technique for searching for a coincident pattern portion, which may vary in size and shape. Therefore, pattern matching needs to be carried out based on the premise that the subject pattern portion deforms. If the positioning accuracy in pattern matching is insufficient, correct measurement values are not obtained. In the worst case, there is the possibility that totally erroneous positions are measured.
- Pattern deformation variations are combinations of pattern deformations in each layer.
- contact hole bottom surfaces may be more deformed than contact hole front surfaces. Combinations of such variations of deformations of contact hole bottoms and variations of deformations of contact hole front surfaces result in variations of deformations of SEM images of contact holes where pattern front surface and bottom portions are represented in a combined manner.
- a pattern matching method capable of accomplishing higher speeds of matching processing while attaining high matching accuracy even if the pattern is a composite pattern consisting of patterns present in plural layers is hereinafter described, as well as a device for implementing the pattern matching.
- a pattern matching method comprising the steps of performing first pattern matching on an image representing a pattern whose constituent parts are present across plural layers (heights) using a first template in which an upper portion of the pattern is selectively represented; and subtracting information about the upper portion of the pattern from an image of a region including a pattern identified by the first template or performing second pattern matching using a second template in which a lower portion of the pattern is represented.
- the pattern matching method as described above and the device for implementing it make it possible to extract the shape of each hierarchical level if the size and shape of the pattern and the degree of overlap vary. Also, the accuracy of the measurement position can be enhanced. Besides, when the degree of overlap is large, it is possible to make a selection and setting as to how the measurement processing should be performed and how a report of the circumstances should be made.
- FIG. 4 is a diagram schematically illustrating an SEM that is one kind of charged-particle beam apparatus.
- the SEM described in the present embodiment is principally provided with a scanning electron microscope body 401 , a control portion 402 , a host (processor) 407 , a screen display device 409 , and a mouse (pointing device) 410 which are connected as shown.
- the control portion 402 is equipped with an image processing portion B 403 , an imaging control portion 404 , a stage control portion 405 , and a vacuum control portion 406 . A program for providing control by these control portions is registered.
- the imaging control portion 404 provides control for obtaining SEM images from the scanning electron microscope body 401 .
- the stage control portion 405 is intended to control movement of a stage (sample stage) within the scanning electron microscope body 401 .
- the vacuum control portion 406 controls an exhaust vacuum pump within the scanning electron microscope body 401 .
- the image processing portion B 403 enters SEM images from the imaging control portion 404 and processes them.
- the host 407 has an image processing portion A 408 .
- Image processing including many steps checked by a human operator is mainly performed by the image processing portion A ( 408 ), while image processing that needs to be performed continuously at high speed is chiefly done by the image processing portion B 403 because it takes much time cost to transfer image information between the control portion 402 and the host 407 .
- Processing for registering templates offered to the pattern matching described hereinafter is carried out by the image processing portion A 408 of the host 407 .
- SEM images are either transferred from the control portion or stored in a storage device within the host.
- Design data is stored in a storage device within the host.
- Screen images are output to the screen display device 409 . The operator performs arbitrary manipulations on the screen images using the mouse 410 .
- Matching processing and shape extraction processing also described hereinafter is similarly carried out by the image processing portion A 408 of the host 407 .
- the processing is carried out by the image processing portion B 403 according to the need.
- the measurement position accuracy can be enhanced by applying the following algorithm if the device is any instrument (not limited to microscopy apparatus) having the objective of performing automatic measurements based on image information. Furthermore, if the device is not connected with any apparatus for acquiring images in practice, the following algorithm can be utilized provided that the device is in an environment where images can be processed.
- FIG. 5 is a chart illustrating one example of template matching processing.
- the obtained SEM image is edge enhanced and then binarized, followed by execution of matching processing.
- FIG. 6 is a diagram illustrating another example of template matching processing.
- edge extraction processing is performed on the template and SEM image and then smoothing is done. By intentionally blurring the edges in this way, the difference in shape between the edges of the template and SEM image is interpolated. As a result, the rate of success of the matching can be enhanced.
- the template is created based on design data
- the difference in shape between the edges of the template and SEM image is great and, therefore, this technique is effective.
- FIGS. 5 and 6 merely show exemplary examples and that various other pattern matching methods can be applied.
- FIG. 1 is a view illustrating pattern matching using templates and one example of step of extracting shapes of a lower layer pattern.
- An object of the present embodiment is to recognize and measure a deep-bottom pattern that is viewed as a lower-level layer in a semiconductor pattern with round holes (hole pattern).
- the deep-bottom pattern that is viewed as the lower-level layer varies greatly especially in shape, size, and position.
- the step shown in FIG. 1 pertains to a technique for stably performing positioning relying on template matching even if the deformation of shape of the contact hole bottom portions (or intermediate layer) or the like is great.
- pattern information providing fundamental templates is prepared.
- the templates are created based on actual SEM images.
- the templates are created based on design data about semiconductor devices.
- an image 101 representing both a hole bottom (a lower portion of a pattern) and a hole surface (an upper portion of the pattern) and an image 102 selectively representing the hole surface are prepared.
- the image 101 was obtained under the condition where the sample surface was positively charged.
- the image 102 was acquired under the condition where the sample surface was not positively charged.
- the pre-dose technique is available as one method of positively charging sample surfaces.
- the pre-dose is a method of preliminary charging consisting of performing beam scanning so as to charge a sample surface before an image for measurement or inspection is obtained. More specifically, the sample surface is positively charged by scanning a beam having a secondary electron emission efficiency (electrons emitted from the sample/electrons incident on the sample) ⁇ exceeding 1 over the sample (i.e., the condition where exiting electrons are larger in number than electrons incident on the sample).
- the image 101 is an image obtained in such a manner.
- the image 102 is an image formed as a result of execution of beam scanning for the formation of the image under the condition where electrons emitted from the hole bottom cannot ascend to the sample surface.
- an image selectively representing the hole surface can be acquired by scanning the sample with a beam (such as a beam having a secondary electron emission efficiency of less than 1) inducing negative charge opposite to positive charge and then obtaining an image.
- An upper-level template 104 selectively representing an upper portion of a pattern and a lower-level template 109 selectively representing a lower portion of the pattern are formed from the image obtained as described so far.
- the upper-level template 104 is a part or the whole of the image 102 , and is formed by selecting an image where the contour of the upper part of the hole appears.
- information about the hole bottom might be a cause hindering a decision on the degree of match in matching. Therefore, where the hole bottom is displayed on the image, processing is performed in such a way that the hole bottom is masked or information about the portion corresponding to the hole bottom is selectively lost.
- only the image portion representing the upper portion of the hole may be selectively applied to an algorithm for computing the degree of match.
- the lower-level template 109 is formed by selecting the portion of the image 101 which represents the pattern bottom portion (by removing the portion where the upper portion of the pattern appears from the image data).
- the following matching processing is performed after at least two templates are created as described so far.
- Shape information 106 about the upper layer pattern is obtained by performing pattern matching using the upper-level template 104 on the image 105 and selectively extracting the upper layer pattern.
- the pattern matching of only the upper layer can be performed without being affected by variations in the lower-level pattern by masking the inside of the pattern or performing such processing that portions having shape or edge features matching the upper-level template are selectively subjected to matching as described previously.
- Multilayer shape (geometric) information 107 indicates the state in which shape detection on the other layer is done within the range (e.g., a region of the same size as the upper-level template 104 ) identified by the upper-level template 104 . Position error and processing speed deterioration can be prevented by restricting the detection range to the region identified by the upper-level template in this way.
- shape information 108 about the lower layer pattern is extracted. This processing is achieved by subtracting the shape information 106 about the upper layer pattern from the multilayer shape information 107 .
- the lower-level template 109 may be prepared and portions of identical features may be extracted but this is limited to when variations are not so large.
- the lower-level template 109 can be prepared by specifying and extracting those portions from the image 101 or from the design data 103 .
- FIG. 7 is a flowchart illustrating other example of processing of the above-described two-stage matching.
- a pattern search is carried out using an upper layer pattern template 702 to identify an upper layer pattern 703 displayed within an upper layer pattern search region 701 (step (a)). Since the position of the upper layer pattern 703 is identified by the step (a), a search range for extracting a lower layer pattern is determined based on the positional information (step (b)). In the step (b), a region narrower than the upper layer pattern search region 701 is set as the search range 704 for the lower layer pattern utilizing the positional information about the upper layer pattern. Although the size of the search region for the lower layer pattern can be set at will, it is desired that the region be narrower than the upper layer pattern search region 701 and wider than a lower layer pattern template 705 prepared for the lower layer pattern.
- a pre-dose of the upper layer pattern region surrounding the lower layer pattern 706 is then is carried out by electron-beam scanning (step (c)).
- the pre-dose is a technique for extracting electrons released from the bottoms of deep holes (lower layer) by positively charging the sample surface (upper layer) and so it is necessary that the electron beam be scanned relative to the sample surface region surrounding the bottoms of the deep holes. If the pre-dose is done, the amount of electrons released from the bottoms of the deep holes and drawn out to the sample surface increases dramatically. Consequently, on the image of a pattern search region 704 of the lower layer, the lower layer pattern 706 is emphasized (step (d)).
- a pattern search is effected (step (e)) using the lower layer pattern template 705 to identify the position of the lower layer pattern (step (f)).
- the rate of success of matching can be enhanced further by deleting the edge portion 707 of the upper layer pattern displayed on the lower layer pattern template by means of image processing where necessary.
- the process of the pre-dose is not always necessary.
- accurate matching processing can be accomplished because matching processing can be performed based on images where the upper and lower layers are respectively enhanced by using images taken respectively before and after a pre-dose in two-stage matching.
- FIG. 2 is a view illustrating another example of steps of pattern matching using templates.
- the example of steps shown in FIG. 2 is similar to Embodiment 1.
- Embodiment 1 it is assumed that it is difficult to discriminate features in the upper and lower layers on the image 105 obtained by imaging a subject of inspection to be measured because they resemble or that it is difficult to make matching with the upper-level template 104 if information about the upper and lower layers is mixed because variations in the upper layer are large.
- Image information 201 has been obtained by imaging a pattern of only the upper-level layer of a subject of inspection to be measured. Matching with the upper-level template 104 is made using the image information 201 , and position detection and shape extraction are performed. Shape extraction on every layer is effected with the separately obtained image 105 . In the case of the present example, shape extraction about the upper and lower layers is done, and the multilayer shape (geometric) information 107 is obtained. The shape information 108 about the lower layer pattern is extracted by subtracting the shape information 106 about the upper layer pattern from the multilayer shape information 107 .
- FIG. 3 is a diagram illustrating an example of correspondence when measurement processing is varied according to the kind of obtained image.
- a case A 301 is a case where a deep-bottom pattern which is viewed as a lower-level layer is seen without overlapping with an upper-level layer.
- a position to be measured (indicated by an arrow in the figure) can be uniquely determined. In this case, measurements can be made without any particular problems. Therefore, a normal measurement of the lower layer pattern is carried out.
- a case B 302 is a case where a deep-bottom pattern which is viewed as a lower-level layer appears in contact or overlapping with an upper-level layer.
- the measurement is between the upper and lower layers.
- setting 1 is selected.
- setting 2 is selected to generate error information, for terminating the measurement operation.
- a case C 303 is a case where the deep-bottom pattern which should be viewed as a lower-level layer is not seen at all.
- the setting can be selected between setting 1 in which measurement value zero is output and a warning is issued and setting 2 in which error information is generated and the measurement operation is ended.
- the accuracy may be enhanced by making templates automatically displayed at first rest on information registered in the past (such as by learning). Furthermore, in modifying the effective positions of templates, any arbitrary position may be specified rather than preset positions. In addition, in pattern matching of similar structures, it may be possible to cope with deformation of the subject of measurement robustly by imparting a likelihood to the positional relationship between the similarities. Further, the operator may set the likelihood with a parameter.
Abstract
Description
- The present invention relates to a pattern matching method and device and, more particularly, to a method of matching for a pattern formed over plural layers and to a device implementing the matching.
- In semiconductor device fabrication steps, thin-film patterns formed on semiconductor wafers have been made finer with larger scales of integration. It is more important to automate process monitoring for accurate, high-speed processing, in order to inspect whether or not the patterns are formed on the wafers as designed.
- From the past, scanning electron microscopes (SEMs) have been used as equipment for measuring the dimensions of minute patterns.
- SEMs include equipment known as CD (critical-dimension)-SEM, and in this kind of equipment positioning is performed while relying on pattern matching when patterns on semiconductor wafers or masks placed on stages (sample stages) are measured.
- The pattern matching method is a technique of causing a template representing a unique pattern formed on a sample to search within a given image region, to identify locations having high degrees of match with the template. Since a relative positional relationship between a pattern identified by the pattern matching and a pattern subjected to a measurement is previously known, it is possible to quickly perform positioning relative to the minute pattern subjected to the measurement. Furthermore, in some cases, subjects of measurement themselves are taken as matching patterns.
- In
Patent Literature 1, a pattern matching apparatus has been proposed that is designed to perform pattern matching in which a hole pattern is subjected to pattern matching and in which design data about a semiconductor pattern is taken as a template. -
- Patent Literature 1: JP-A-2007-121147 (corresponding to US 2007/0098248)
- In recent years, patterns have been made with larger scales of integration. Concomitantly, semiconductor devices have tended to have multilayer structures or three-dimensional structures. Following further miniaturization, contact holes or the like formed in devices have become deeper. Whether holes have been properly formed is an important item of measurement.
- On the other hand, in pattern matching, a proper degree of match is required between an SEM image subjected to matching and a template. However, with respect to a pattern of holes or the like spanning plural layers, a composite shape of its upper and lower portions is a pattern shape represented on an SEM image and so its pattern deformations have more variations than planar patterns. Especially, the lower portions of hole patterns are deformed to a larger extent. In addition, the orbit of electrons emitted from the bottoms of the holes is affected by electric charge adhering to the sample surface (electrification). The yield of the detected secondary electrons may be varied.
- If the formed image is varied by the efficiency at which secondary electrons are detected or if the degree of deformation of the pattern increases, the rate of success of matching decreases. As a result, there is the anxiety that the measurement accuracy decreases, for example, based on measurement of a wrong part or the throughput deteriorates based on the inability to identify a desired measurement part. In the technique disclosed in
Patent Literature 1, a method that is excellent as a pattern matching method using design data is explained but the technique is not proposed in such a way that deformation of the bottom of a hole pattern or a variation in the efficiency at which secondary electrons are detected is taken as a problem to be solved. - A method for the purpose of achieving stable positioning or selective information extraction on a certain layer regardless of the circumstances of the depths of a pattern formed in three dimensions or of the circumstances of charging of a sample and a device for achieving the purpose are hereinafter described.
- To achieve the above-described purpose, a method of performing first pattern matching on image information representing plural layers by the use of a first template selectively representing the shape of an upper portion of a pattern is proposed, as well as a device. Furthermore, in order to perform selective extraction on a certain layer, a method of subtracting information about the shape of the upper portion of the pattern from the image information representing the plural layers or from the shape information is proposed, as well as a device.
- The above-described structure makes it possible to perform pattern matching at a quite high success rate regardless of variations in information obtained from the lower portion of the pattern. Furthermore, selective extraction of a pattern in a lower-level layer is enabled.
- Other objects, features, and advantages of the present invention will become apparent from the following description of embodiments of the invention with respect to the accompanying drawings.
-
FIG. 1 is a view illustrating one example of processing for extracting shapes of a lower portion of a pattern, based on pattern matching using a template selectively representing an upper portion of the pattern; -
FIG. 2 is a view illustrating another example of processing for extracting shapes of a lower portion of a pattern, based on pattern matching using a template selectively representing an upper portion of the pattern; -
FIG. 3 is a diagram illustrating an example of correspondence when measurement processing is varied according to the kind of obtained image; -
FIG. 4 is a diagram schematically illustrating an SEM system; -
FIG. 5 is a chart illustrating one example of template matching processing; -
FIG. 6 is a diagram illustrating one example of pattern matching processing; and -
FIG. 7 is a diagram illustrating one example of two-stage matching for an upper layer pattern and a lower layer pattern. - Nowadays, it has become more important to measure and inspect multilayer semiconductor devices but the following undesired phenomena occasionally appear when measurement positions are determined using pattern matching.
- Fundamentally, pattern matching is a technique for searching for a coincident pattern portion, which may vary in size and shape. Therefore, pattern matching needs to be carried out based on the premise that the subject pattern portion deforms. If the positioning accuracy in pattern matching is insufficient, correct measurement values are not obtained. In the worst case, there is the possibility that totally erroneous positions are measured.
- Especially, where a multilayer structure or a three-dimensional structure is subjected to pattern matching, variable factors of pattern appear across plural layers. Pattern deformation variations are combinations of pattern deformations in each layer. Especially, in contact holes or the like, contact hole bottom surfaces may be more deformed than contact hole front surfaces. Combinations of such variations of deformations of contact hole bottoms and variations of deformations of contact hole front surfaces result in variations of deformations of SEM images of contact holes where pattern front surface and bottom portions are represented in a combined manner.
- In order to cope with many kinds of pattern deformations, it is conceivable to prepare many templates for pattern matching, the templates being different in pattern size, shape, and others, but the number of the previously prepared templates is exorbitant. Furthermore, there is the anxiety that the matching processing time will be prolonged.
- A pattern matching method capable of accomplishing higher speeds of matching processing while attaining high matching accuracy even if the pattern is a composite pattern consisting of patterns present in plural layers is hereinafter described, as well as a device for implementing the pattern matching.
- A summary of description using the drawings is as follows.
- (1) A pattern matching method comprising the steps of performing first pattern matching on an image representing a pattern whose constituent parts are present across plural layers (heights) using a first template in which an upper portion of the pattern is selectively represented; and subtracting information about the upper portion of the pattern from an image of a region including a pattern identified by the first template or performing second pattern matching using a second template in which a lower portion of the pattern is represented.
- (2) The pattern matching method in which, when the second pattern matching is performed, the second template from which information about a pattern represented in the first pattern has been subtracted is used.
- (3) The pattern matching method in which the first template is derived from an SEM image for which no preliminary charging scanning has been done.
- (4) The pattern matching method in which the second template is derived from an SEM image for which preliminary charging scanning has been done.
- (5) The pattern matching method in which, when the first pattern matching is performed, an SEM image is derived under the condition where no preliminary charging scanning is done.
- (6) The pattern matching method in which, when the second pattern matching is performed, an SEM image is derived under the condition where preliminary charging scanning has been done.
- (7) The pattern matching method in which a decision is made as to whether or not a measurement is performed according to the degree of overlap between upper and lower portions of a pattern whose constitutive parts are present across plural layers (heights).
- (8) The pattern matching method in which a subsequent method of measurement or the like is set according to the degree of overlap between the upper and lower portions of a pattern whose constitutive parts are present across plural layers (heights).
- The pattern matching method as described above and the device for implementing it make it possible to extract the shape of each hierarchical level if the size and shape of the pattern and the degree of overlap vary. Also, the accuracy of the measurement position can be enhanced. Besides, when the degree of overlap is large, it is possible to make a selection and setting as to how the measurement processing should be performed and how a report of the circumstances should be made.
-
FIG. 4 is a diagram schematically illustrating an SEM that is one kind of charged-particle beam apparatus. The SEM described in the present embodiment is principally provided with a scanningelectron microscope body 401, acontrol portion 402, a host (processor) 407, ascreen display device 409, and a mouse (pointing device) 410 which are connected as shown. - The
control portion 402 is equipped with an imageprocessing portion B 403, animaging control portion 404, astage control portion 405, and avacuum control portion 406. A program for providing control by these control portions is registered. Theimaging control portion 404 provides control for obtaining SEM images from the scanningelectron microscope body 401. - The
stage control portion 405 is intended to control movement of a stage (sample stage) within the scanningelectron microscope body 401. - The
vacuum control portion 406 controls an exhaust vacuum pump within the scanningelectron microscope body 401. The imageprocessing portion B 403 enters SEM images from theimaging control portion 404 and processes them. Thehost 407 has an imageprocessing portion A 408. - Image processing including many steps checked by a human operator is mainly performed by the image processing portion A (408), while image processing that needs to be performed continuously at high speed is chiefly done by the image
processing portion B 403 because it takes much time cost to transfer image information between thecontrol portion 402 and thehost 407. - Processing for registering templates offered to the pattern matching described hereinafter is carried out by the image
processing portion A 408 of thehost 407. SEM images are either transferred from the control portion or stored in a storage device within the host. Design data is stored in a storage device within the host. Screen images are output to thescreen display device 409. The operator performs arbitrary manipulations on the screen images using the mouse 410. - Matching processing and shape extraction processing also described hereinafter is similarly carried out by the image
processing portion A 408 of thehost 407. However, where the processing capability of the host processor or the image transfer rate from the control portion is considered to be low, the processing is carried out by the imageprocessing portion B 403 according to the need. - The measurement position accuracy can be enhanced by applying the following algorithm if the device is any instrument (not limited to microscopy apparatus) having the objective of performing automatic measurements based on image information. Furthermore, if the device is not connected with any apparatus for acquiring images in practice, the following algorithm can be utilized provided that the device is in an environment where images can be processed.
-
FIG. 5 is a chart illustrating one example of template matching processing. In the present matching processing, the obtained SEM image is edge enhanced and then binarized, followed by execution of matching processing.FIG. 6 is a diagram illustrating another example of template matching processing. In this processing, edge extraction processing is performed on the template and SEM image and then smoothing is done. By intentionally blurring the edges in this way, the difference in shape between the edges of the template and SEM image is interpolated. As a result, the rate of success of the matching can be enhanced. Especially, where the template is created based on design data, the difference in shape between the edges of the template and SEM image is great and, therefore, this technique is effective. It is to be noted that the examples of processing shown inFIGS. 5 and 6 merely show exemplary examples and that various other pattern matching methods can be applied. -
FIG. 1 is a view illustrating pattern matching using templates and one example of step of extracting shapes of a lower layer pattern. An object of the present embodiment is to recognize and measure a deep-bottom pattern that is viewed as a lower-level layer in a semiconductor pattern with round holes (hole pattern). The deep-bottom pattern that is viewed as the lower-level layer varies greatly especially in shape, size, and position. The step shown inFIG. 1 pertains to a technique for stably performing positioning relying on template matching even if the deformation of shape of the contact hole bottom portions (or intermediate layer) or the like is great. - First, pattern information providing fundamental templates is prepared. In some cases, the templates are created based on actual SEM images. In other cases, the templates are created based on design data about semiconductor devices.
- In the former case, an
image 101 representing both a hole bottom (a lower portion of a pattern) and a hole surface (an upper portion of the pattern) and animage 102 selectively representing the hole surface are prepared. - The
image 101 was obtained under the condition where the sample surface was positively charged. Theimage 102 was acquired under the condition where the sample surface was not positively charged. The pre-dose technique is available as one method of positively charging sample surfaces. The pre-dose is a method of preliminary charging consisting of performing beam scanning so as to charge a sample surface before an image for measurement or inspection is obtained. More specifically, the sample surface is positively charged by scanning a beam having a secondary electron emission efficiency (electrons emitted from the sample/electrons incident on the sample) δ exceeding 1 over the sample (i.e., the condition where exiting electrons are larger in number than electrons incident on the sample). - When the beam for forming an image is scanned after the preliminary charging as described so far has been performed, electrons produced at the hole bottom are pulled toward positive charges on the surface of the sample. Therefore, the electrons produced at the hole bottom can be captured at high efficiency. The
image 101 is an image obtained in such a manner. - On the other hand, the
image 102 is an image formed as a result of execution of beam scanning for the formation of the image under the condition where electrons emitted from the hole bottom cannot ascend to the sample surface. In particular, it is conceivable to obtain an image prior to execution of preliminary charging. Furthermore, in a case where charging is induced in any situation or in a case where the hole is shallow and electrons released from hole bottom are guided to the sample surface without inducing charging, an image selectively representing the hole surface can be acquired by scanning the sample with a beam (such as a beam having a secondary electron emission efficiency of less than 1) inducing negative charge opposite to positive charge and then obtaining an image. - An upper-
level template 104 selectively representing an upper portion of a pattern and a lower-level template 109 selectively representing a lower portion of the pattern are formed from the image obtained as described so far. The upper-level template 104 is a part or the whole of theimage 102, and is formed by selecting an image where the contour of the upper part of the hole appears. In the case of the present example, information about the hole bottom might be a cause hindering a decision on the degree of match in matching. Therefore, where the hole bottom is displayed on the image, processing is performed in such a way that the hole bottom is masked or information about the portion corresponding to the hole bottom is selectively lost. Besides, only the image portion representing the upper portion of the hole may be selectively applied to an algorithm for computing the degree of match. - The lower-
level template 109 is formed by selecting the portion of theimage 101 which represents the pattern bottom portion (by removing the portion where the upper portion of the pattern appears from the image data). - On the other hand, in the case of the latter example where templates are formed based on design data, information about the upper and lower portions of the pattern is previously stored in the
design data 103. Therefore, necessary information is extracted from a GDS file or the like where the design data is stored, and what are imaged two-dimensionally from the information can be applied in place of theimages level template 104 and lower-level template 109 without performing preliminary charging. - The following matching processing is performed after at least two templates are created as described so far.
- First, pattern matching of only the upper layer is performed using the upper-
level template 104, and its position and shape are extracted.Shape information 106 about the upper layer pattern is obtained by performing pattern matching using the upper-level template 104 on theimage 105 and selectively extracting the upper layer pattern. - The pattern matching of only the upper layer can be performed without being affected by variations in the lower-level pattern by masking the inside of the pattern or performing such processing that portions having shape or edge features matching the upper-level template are selectively subjected to matching as described previously.
- Then, a shape extraction is performed on all the layers within the range where the pattern positions have been detected. Multilayer shape (geometric)
information 107 indicates the state in which shape detection on the other layer is done within the range (e.g., a region of the same size as the upper-level template 104) identified by the upper-level template 104. Position error and processing speed deterioration can be prevented by restricting the detection range to the region identified by the upper-level template in this way. - Then, shape
information 108 about the lower layer pattern is extracted. This processing is achieved by subtracting theshape information 106 about the upper layer pattern from themultilayer shape information 107. Alternatively, the lower-level template 109 may be prepared and portions of identical features may be extracted but this is limited to when variations are not so large. The lower-level template 109 can be prepared by specifying and extracting those portions from theimage 101 or from thedesign data 103. - If a deep, bottom pattern shape which seems as a lower-level layer is extracted and identified as described so far, it is unlikely that the measurement position is in error even in the case where the shape, size, or position varies greatly.
-
FIG. 7 is a flowchart illustrating other example of processing of the above-described two-stage matching. In the present embodiment, a pattern search is carried out using an upperlayer pattern template 702 to identify anupper layer pattern 703 displayed within an upper layer pattern search region 701 (step (a)). Since the position of theupper layer pattern 703 is identified by the step (a), a search range for extracting a lower layer pattern is determined based on the positional information (step (b)). In the step (b), a region narrower than the upper layerpattern search region 701 is set as thesearch range 704 for the lower layer pattern utilizing the positional information about the upper layer pattern. Although the size of the search region for the lower layer pattern can be set at will, it is desired that the region be narrower than the upper layerpattern search region 701 and wider than a lowerlayer pattern template 705 prepared for the lower layer pattern. - A pre-dose of the upper layer pattern region surrounding the
lower layer pattern 706 is then is carried out by electron-beam scanning (step (c)). The pre-dose is a technique for extracting electrons released from the bottoms of deep holes (lower layer) by positively charging the sample surface (upper layer) and so it is necessary that the electron beam be scanned relative to the sample surface region surrounding the bottoms of the deep holes. If the pre-dose is done, the amount of electrons released from the bottoms of the deep holes and drawn out to the sample surface increases dramatically. Consequently, on the image of apattern search region 704 of the lower layer, thelower layer pattern 706 is emphasized (step (d)). - Within the
pattern search region 704 of the lower layer in which the lower layer pattern has been emphasized as described so far, a pattern search is effected (step (e)) using the lowerlayer pattern template 705 to identify the position of the lower layer pattern (step (f)). The rate of success of matching can be enhanced further by deleting theedge portion 707 of the upper layer pattern displayed on the lower layer pattern template by means of image processing where necessary. - Furthermore, the process of the pre-dose is not always necessary. However, accurate matching processing can be accomplished because matching processing can be performed based on images where the upper and lower layers are respectively enhanced by using images taken respectively before and after a pre-dose in two-stage matching.
-
FIG. 2 is a view illustrating another example of steps of pattern matching using templates. The example of steps shown inFIG. 2 is similar toEmbodiment 1. However, inEmbodiment 1, it is assumed that it is difficult to discriminate features in the upper and lower layers on theimage 105 obtained by imaging a subject of inspection to be measured because they resemble or that it is difficult to make matching with the upper-level template 104 if information about the upper and lower layers is mixed because variations in the upper layer are large. -
Image information 201 has been obtained by imaging a pattern of only the upper-level layer of a subject of inspection to be measured. Matching with the upper-level template 104 is made using theimage information 201, and position detection and shape extraction are performed. Shape extraction on every layer is effected with the separately obtainedimage 105. In the case of the present example, shape extraction about the upper and lower layers is done, and the multilayer shape (geometric)information 107 is obtained. Theshape information 108 about the lower layer pattern is extracted by subtracting theshape information 106 about the upper layer pattern from themultilayer shape information 107. - An example of error handling in a case where the lower-level layer of a contact hole pattern is automatically measured is hereinafter described. Where an automatic measurement of a pattern is made in CD-SEM, if the measurement is made on the assumption that there is given information though the given information is not obtained on the image, the measurement is carried out on an object despite it is different from the subject of measurement. Consequently, the reliability of the measurement results is in danger of deteriorating. In the present example, a method which, when given information is not obtained, is capable of selectively extracting only reliable information by issuing a warning to that effect or terminating the measurement and a device are described.
-
FIG. 3 is a diagram illustrating an example of correspondence when measurement processing is varied according to the kind of obtained image. - A
case A 301 is a case where a deep-bottom pattern which is viewed as a lower-level layer is seen without overlapping with an upper-level layer. A position to be measured (indicated by an arrow in the figure) can be uniquely determined. In this case, measurements can be made without any particular problems. Therefore, a normal measurement of the lower layer pattern is carried out. - A
case B 302 is a case where a deep-bottom pattern which is viewed as a lower-level layer appears in contact or overlapping with an upper-level layer. - In the instance of such cases, if the pattern width of the lower layer is measured, the measurement is between the upper and lower layers. Where such a subject of measurement is measured, setting 1 is selected. Where a measurement of such a subject of measurement is not desired, setting 2 is selected to generate error information, for terminating the measurement operation.
- A
case C 303 is a case where the deep-bottom pattern which should be viewed as a lower-level layer is not seen at all. - Also, in this case, the setting can be selected between setting 1 in which measurement value zero is output and a warning is issued and setting 2 in which error information is generated and the measurement operation is ended. Thus, it is possible to obtain measurement results fulfilling user's intention.
- In the above embodiment, in the method of registering a measuring position and positioning templates, the accuracy may be enhanced by making templates automatically displayed at first rest on information registered in the past (such as by learning). Furthermore, in modifying the effective positions of templates, any arbitrary position may be specified rather than preset positions. In addition, in pattern matching of similar structures, it may be possible to cope with deformation of the subject of measurement robustly by imparting a likelihood to the positional relationship between the similarities. Further, the operator may set the likelihood with a parameter.
- The above description has been provided concerning embodiments. The present invention is not limited to them. It is obvious for those skilled in the art that various changes and modifications may be made within the spirit of the present invention and the scopes of the accompanying claims.
-
-
- 401: scanning electron microscope body
- 402: control portion
- 403: image processing portion B
- 404: imaging control portion
- 405: stage control portion
- 406: vacuum control portion
- 407: host
- 408: image processing portion A
- 409: screen display device
- 410: mouse
Claims (10)
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2008257940A JP5063551B2 (en) | 2008-10-03 | 2008-10-03 | Pattern matching method and image processing apparatus |
JP2008-257940 | 2008-10-03 | ||
PCT/JP2009/067226 WO2010038859A1 (en) | 2008-10-03 | 2009-10-02 | Pattern matching method and image processing device |
Publications (2)
Publication Number | Publication Date |
---|---|
US20110262043A1 true US20110262043A1 (en) | 2011-10-27 |
US8953894B2 US8953894B2 (en) | 2015-02-10 |
Family
ID=42073617
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/122,151 Active 2029-12-25 US8953894B2 (en) | 2008-10-03 | 2009-10-02 | Pattern matching method and image processing device |
Country Status (3)
Country | Link |
---|---|
US (1) | US8953894B2 (en) |
JP (1) | JP5063551B2 (en) |
WO (1) | WO2010038859A1 (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130216141A1 (en) * | 2010-07-01 | 2013-08-22 | Hitachi High-Technologies Corporation | Pattern matching method, image processing device, and computer program |
US20170249766A1 (en) * | 2016-02-25 | 2017-08-31 | Fanuc Corporation | Image processing device for displaying object detected from input picture image |
US10545017B2 (en) | 2012-07-06 | 2020-01-28 | Hitachi High-Technologies Corporation | Overlay error measuring device and computer program for causing computer to measure pattern |
WO2020191121A1 (en) * | 2019-03-20 | 2020-09-24 | Carl Zeiss Smt Inc. | Method for imaging a region of interest of a sample using a tomographic x-ray microscope, microscope, system and computer program |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2012155637A (en) * | 2011-01-28 | 2012-08-16 | Hitachi High-Technologies Corp | Pattern matching device and computer program |
JP5639925B2 (en) * | 2011-02-25 | 2014-12-10 | 株式会社日立ハイテクノロジーズ | Pattern matching device and computer program |
US9715724B2 (en) * | 2014-07-29 | 2017-07-25 | Applied Materials Israel Ltd. | Registration of CAD data with SEM images |
JP5857106B2 (en) * | 2014-10-14 | 2016-02-10 | 株式会社日立ハイテクノロジーズ | Pattern matching device and computer program |
JP7395566B2 (en) * | 2019-04-02 | 2023-12-11 | 株式会社半導体エネルギー研究所 | Inspection method |
Citations (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20010048478A1 (en) * | 1999-01-25 | 2001-12-06 | Hironobu Taoka | Pattern distortion detecting method and apparatus and recording medium for pattern distortion detection |
US20030173516A1 (en) * | 2001-04-27 | 2003-09-18 | Atsushi Takane | Semiconductor inspection system |
US20040131246A1 (en) * | 2002-10-21 | 2004-07-08 | Takahiro Ikeda | Micropattern measuring method, micropattern measuring apparatus, and computer-readable recording medium on which a micropattern measuring program is recorded |
US20040217288A1 (en) * | 1999-10-07 | 2004-11-04 | Fumihiro Sasajima | Microstructured pattern inspection method |
US20050045821A1 (en) * | 2003-04-22 | 2005-03-03 | Nobuharu Noji | Testing apparatus using charged particles and device manufacturing method using the testing apparatus |
US20050086618A1 (en) * | 2003-09-25 | 2005-04-21 | Takeshi Ito | Apparatus and method for verifying an integrated circuit pattern |
US20050194535A1 (en) * | 2004-03-08 | 2005-09-08 | Ebara Corporation | Sample surface inspection method and inspection system |
US20060193508A1 (en) * | 2005-02-25 | 2006-08-31 | Takumichi Sutani | Pattern measuring method and pattern measuring device |
US20060288325A1 (en) * | 2005-06-15 | 2006-12-21 | Atsushi Miyamoto | Method and apparatus for measuring dimension of a pattern formed on a semiconductor wafer |
US20070023653A1 (en) * | 2005-06-17 | 2007-02-01 | Yasutaka Toyoda | Method and apparatus of pattern inspection and semiconductor inspection system using the same |
US20070092129A1 (en) * | 2005-09-14 | 2007-04-26 | Akiyuki Sugiyama | System and method of image processing, and scanning electron microscope |
US20070098249A1 (en) * | 2005-10-31 | 2007-05-03 | Yumiko Miyano | Pattern shape evaluation apparatus, pattern shape evaluation method, method of manufacturing semiconductor device, and program |
US20080032429A1 (en) * | 2005-11-09 | 2008-02-07 | Da Chen | Methods, defect review tools, and systems for locating a defect in a defect review process |
US20080138916A1 (en) * | 2006-04-26 | 2008-06-12 | Tadashi Mitsui | Pattern shape evaluation method, program, and semiconductor device manufacturing method |
US20090039263A1 (en) * | 2007-08-09 | 2009-02-12 | Ryoichi Matsuoka | Pattern measurement apparatus |
US20090084954A1 (en) * | 2007-09-27 | 2009-04-02 | Hitachi High-Technologies Corporation | Method for inspecting and measuring sample and scanning electron microscope |
US20090238443A1 (en) * | 2008-03-18 | 2009-09-24 | Hidetoshi Sato | Pattern measurement methods and pattern measurement equipment |
US20120207397A1 (en) * | 2009-10-22 | 2012-08-16 | Hitachi High-Technologies Corporation | Pattern Matching Method and Pattern Matching Apparatus |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4261743B2 (en) * | 1999-07-09 | 2009-04-30 | 株式会社日立製作所 | Charged particle beam equipment |
JP2005061837A (en) * | 2003-08-11 | 2005-03-10 | Jeol Ltd | Defect inspecting method using scanning charged particle beam system |
JP4136883B2 (en) * | 2003-10-03 | 2008-08-20 | 株式会社日立ハイテクノロジーズ | Defect observation method |
JP4824987B2 (en) | 2005-10-28 | 2011-11-30 | 株式会社日立ハイテクノロジーズ | Pattern matching apparatus and semiconductor inspection system using the same |
-
2008
- 2008-10-03 JP JP2008257940A patent/JP5063551B2/en active Active
-
2009
- 2009-10-02 US US13/122,151 patent/US8953894B2/en active Active
- 2009-10-02 WO PCT/JP2009/067226 patent/WO2010038859A1/en active Application Filing
Patent Citations (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20010048478A1 (en) * | 1999-01-25 | 2001-12-06 | Hironobu Taoka | Pattern distortion detecting method and apparatus and recording medium for pattern distortion detection |
US20040217288A1 (en) * | 1999-10-07 | 2004-11-04 | Fumihiro Sasajima | Microstructured pattern inspection method |
US20030173516A1 (en) * | 2001-04-27 | 2003-09-18 | Atsushi Takane | Semiconductor inspection system |
US20040131246A1 (en) * | 2002-10-21 | 2004-07-08 | Takahiro Ikeda | Micropattern measuring method, micropattern measuring apparatus, and computer-readable recording medium on which a micropattern measuring program is recorded |
US20050045821A1 (en) * | 2003-04-22 | 2005-03-03 | Nobuharu Noji | Testing apparatus using charged particles and device manufacturing method using the testing apparatus |
US20050086618A1 (en) * | 2003-09-25 | 2005-04-21 | Takeshi Ito | Apparatus and method for verifying an integrated circuit pattern |
US20050194535A1 (en) * | 2004-03-08 | 2005-09-08 | Ebara Corporation | Sample surface inspection method and inspection system |
US20060193508A1 (en) * | 2005-02-25 | 2006-08-31 | Takumichi Sutani | Pattern measuring method and pattern measuring device |
US20060288325A1 (en) * | 2005-06-15 | 2006-12-21 | Atsushi Miyamoto | Method and apparatus for measuring dimension of a pattern formed on a semiconductor wafer |
US20070023653A1 (en) * | 2005-06-17 | 2007-02-01 | Yasutaka Toyoda | Method and apparatus of pattern inspection and semiconductor inspection system using the same |
US20090152463A1 (en) * | 2005-06-17 | 2009-06-18 | Hitachi High-Technologies Corporation | Method and apparatus of pattern inspection and semiconductor inspection system using the same |
US20070092129A1 (en) * | 2005-09-14 | 2007-04-26 | Akiyuki Sugiyama | System and method of image processing, and scanning electron microscope |
US20070098249A1 (en) * | 2005-10-31 | 2007-05-03 | Yumiko Miyano | Pattern shape evaluation apparatus, pattern shape evaluation method, method of manufacturing semiconductor device, and program |
US20080032429A1 (en) * | 2005-11-09 | 2008-02-07 | Da Chen | Methods, defect review tools, and systems for locating a defect in a defect review process |
US20080138916A1 (en) * | 2006-04-26 | 2008-06-12 | Tadashi Mitsui | Pattern shape evaluation method, program, and semiconductor device manufacturing method |
US20090039263A1 (en) * | 2007-08-09 | 2009-02-12 | Ryoichi Matsuoka | Pattern measurement apparatus |
US20090084954A1 (en) * | 2007-09-27 | 2009-04-02 | Hitachi High-Technologies Corporation | Method for inspecting and measuring sample and scanning electron microscope |
US20090238443A1 (en) * | 2008-03-18 | 2009-09-24 | Hidetoshi Sato | Pattern measurement methods and pattern measurement equipment |
US20120207397A1 (en) * | 2009-10-22 | 2012-08-16 | Hitachi High-Technologies Corporation | Pattern Matching Method and Pattern Matching Apparatus |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130216141A1 (en) * | 2010-07-01 | 2013-08-22 | Hitachi High-Technologies Corporation | Pattern matching method, image processing device, and computer program |
US9141879B2 (en) * | 2010-07-01 | 2015-09-22 | Hitachi High-Technologies Corporation | Pattern matching method, image processing device, and computer program |
US10545017B2 (en) | 2012-07-06 | 2020-01-28 | Hitachi High-Technologies Corporation | Overlay error measuring device and computer program for causing computer to measure pattern |
US20170249766A1 (en) * | 2016-02-25 | 2017-08-31 | Fanuc Corporation | Image processing device for displaying object detected from input picture image |
US10930037B2 (en) * | 2016-02-25 | 2021-02-23 | Fanuc Corporation | Image processing device for displaying object detected from input picture image |
WO2020191121A1 (en) * | 2019-03-20 | 2020-09-24 | Carl Zeiss Smt Inc. | Method for imaging a region of interest of a sample using a tomographic x-ray microscope, microscope, system and computer program |
US20210407127A1 (en) * | 2019-03-20 | 2021-12-30 | Carl Zeiss Smt Inc. | Method for imaging a region of interest of a sample using a tomographic x-ray microscope, microscope, system and computer program |
Also Published As
Publication number | Publication date |
---|---|
US8953894B2 (en) | 2015-02-10 |
JP5063551B2 (en) | 2012-10-31 |
JP2010086925A (en) | 2010-04-15 |
WO2010038859A1 (en) | 2010-04-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8953894B2 (en) | Pattern matching method and image processing device | |
TWI492165B (en) | Pattern matching method and device | |
TWI512684B (en) | Defect observation method and defect observation device | |
US8045789B2 (en) | Method and apparatus for inspecting defect of pattern formed on semiconductor device | |
US7889909B2 (en) | Pattern matching method and pattern matching program | |
US20210358101A1 (en) | Processing image data sets | |
US8867818B2 (en) | Method of creating template for matching, as well as device for creating template | |
US20140016854A1 (en) | Pattern matching device and computer program | |
US8669523B2 (en) | Contour-based defect detection using an inspection apparatus | |
KR102521386B1 (en) | Dimension measuring device, dimension measuring method, and semiconductor manufacturing system | |
US10074167B2 (en) | Reducing registration and design vicinity induced noise for intra-die inspection | |
KR101709433B1 (en) | Sample observation device | |
JP2011090470A (en) | Pattern matching method and pattern matching device | |
US20120305767A1 (en) | Pattern Inspection Method, Pattern Inspection Program, and Electronic Device Inspection System | |
JP5988615B2 (en) | Semiconductor evaluation apparatus and computer program | |
US20130248709A1 (en) | Defect inspecting apparatus | |
TWI798521B (en) | System and computer-implemented method for selecting one or more deign files for use in test image to design alignment, and non-transitory computer-readable medium | |
WO2017130365A1 (en) | Overlay error measurement device and computer program | |
JP7062563B2 (en) | Contour extraction method, contour extraction device, and program | |
TWI791191B (en) | Method for extracting pattern contour information from an inspection image, contour extraction apparatus, and related non-transistory computer readable medium | |
CN114556416A (en) | Inspection system and non-transitory computer readable medium | |
US10161882B1 (en) | Method of examining locations in a wafer with adjustable navigation accuracy and system thereof | |
JP2003218181A (en) | Defect image pickup device | |
US20230194253A1 (en) | Pattern Inspection/Measurement Device, and Pattern Inspection/Measurement Program | |
KR20190007449A (en) | Pattern evaluation device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: HITACHI HIGH-TECHNOLOGIES CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SATO, YOSHIMICHI;IKEDA, MITSUJI;SASAJIMA, FUMIHIRO;SIGNING DATES FROM 20110418 TO 20110419;REEL/FRAME:026501/0953 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551) Year of fee payment: 4 |
|
AS | Assignment |
Owner name: HITACHI HIGH-TECH CORPORATION, JAPAN Free format text: CHANGE OF NAME AND ADDRESS;ASSIGNOR:HITACHI HIGH-TECHNOLOGIES CORPORATION;REEL/FRAME:052259/0227 Effective date: 20200212 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |