TW202302263A - Observation device, observation method, and object to be observed - Google Patents

Observation device, observation method, and object to be observed Download PDF

Info

Publication number
TW202302263A
TW202302263A TW111102335A TW111102335A TW202302263A TW 202302263 A TW202302263 A TW 202302263A TW 111102335 A TW111102335 A TW 111102335A TW 111102335 A TW111102335 A TW 111102335A TW 202302263 A TW202302263 A TW 202302263A
Authority
TW
Taiwan
Prior art keywords
image
crack
aforementioned
light
unit
Prior art date
Application number
TW111102335A
Other languages
Chinese (zh)
Inventor
坂本剛志
佐野育
Original Assignee
日商濱松赫德尼古斯股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 日商濱松赫德尼古斯股份有限公司 filed Critical 日商濱松赫德尼古斯股份有限公司
Publication of TW202302263A publication Critical patent/TW202302263A/en

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/84Systems specially adapted for particular applications
    • G01N21/88Investigating the presence of flaws or contamination
    • G01N21/8851Scan or image signal processing specially adapted therefor, e.g. for scan signal adjustment, for detecting different kinds of defects, for compensating for structures, markings, edges
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B23MACHINE TOOLS; METAL-WORKING NOT OTHERWISE PROVIDED FOR
    • B23KSOLDERING OR UNSOLDERING; WELDING; CLADDING OR PLATING BY SOLDERING OR WELDING; CUTTING BY APPLYING HEAT LOCALLY, e.g. FLAME CUTTING; WORKING BY LASER BEAM
    • B23K26/00Working by laser beam, e.g. welding, cutting or boring
    • B23K26/70Auxiliary operations or equipment
    • B23K26/702Auxiliary equipment
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B23MACHINE TOOLS; METAL-WORKING NOT OTHERWISE PROVIDED FOR
    • B23KSOLDERING OR UNSOLDERING; WELDING; CLADDING OR PLATING BY SOLDERING OR WELDING; CUTTING BY APPLYING HEAT LOCALLY, e.g. FLAME CUTTING; WORKING BY LASER BEAM
    • B23K26/00Working by laser beam, e.g. welding, cutting or boring
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B23MACHINE TOOLS; METAL-WORKING NOT OTHERWISE PROVIDED FOR
    • B23KSOLDERING OR UNSOLDERING; WELDING; CLADDING OR PLATING BY SOLDERING OR WELDING; CUTTING BY APPLYING HEAT LOCALLY, e.g. FLAME CUTTING; WORKING BY LASER BEAM
    • B23K26/00Working by laser beam, e.g. welding, cutting or boring
    • B23K26/02Positioning or observing the workpiece, e.g. with respect to the point of impact; Aligning, aiming or focusing the laser beam
    • B23K26/03Observing, e.g. monitoring, the workpiece
    • B23K26/032Observing, e.g. monitoring, the workpiece using optical means
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/17Systems in which incident light is modified in accordance with the properties of the material investigated
    • G01N21/59Transmissivity
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/84Systems specially adapted for particular applications
    • G01N21/88Investigating the presence of flaws or contamination
    • G01N21/8806Specially adapted optical and illumination features
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B7/00Mountings, adjusting means, or light-tight connections, for optical elements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0004Industrial image inspection
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B23MACHINE TOOLS; METAL-WORKING NOT OTHERWISE PROVIDED FOR
    • B23KSOLDERING OR UNSOLDERING; WELDING; CLADDING OR PLATING BY SOLDERING OR WELDING; CUTTING BY APPLYING HEAT LOCALLY, e.g. FLAME CUTTING; WORKING BY LASER BEAM
    • B23K2101/00Articles made by soldering, welding or cutting
    • B23K2101/36Electric or electronic devices
    • B23K2101/40Semiconductor devices
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/84Systems specially adapted for particular applications
    • G01N21/88Investigating the presence of flaws or contamination
    • G01N21/8851Scan or image signal processing specially adapted therefor, e.g. for scan signal adjustment, for detecting different kinds of defects, for compensating for structures, markings, edges
    • G01N2021/8854Grading and classifying of flaws
    • G01N2021/8858Flaw counting
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N2201/00Features of devices classified in G01N21/00
    • G01N2201/10Scanning
    • G01N2201/104Mechano-optical scan, i.e. object and beam moving
    • G01N2201/1047Mechano-optical scan, i.e. object and beam moving with rotating optics and moving stage
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N2201/00Features of devices classified in G01N21/00
    • G01N2201/12Circuits of general importance; Signal processing
    • G01N2201/127Calibration; base line adjustment; drift compensation

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Optics & Photonics (AREA)
  • Immunology (AREA)
  • Pathology (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Chemical & Material Sciences (AREA)
  • Analytical Chemistry (AREA)
  • Biochemistry (AREA)
  • General Health & Medical Sciences (AREA)
  • Mechanical Engineering (AREA)
  • Plasma & Fusion (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Signal Processing (AREA)
  • Quality & Reliability (AREA)
  • Theoretical Computer Science (AREA)
  • Microscoopes, Condenser (AREA)
  • Investigating Or Analysing Materials By Optical Means (AREA)
  • Laser Beam Processing (AREA)
  • Length Measuring Devices By Optical Means (AREA)

Abstract

The invention provides an observation device, an observation method, and an object to be observed, which can more accurately acquire information about the position of a modified region. This observation device is provided with: an imaging unit which has a condensing lens for condensing, on an object, transmitted light that is trans missive with respect to the object, and which images the object by means of the transmitted light; a moving unit for moving the condenser lens relative to the object; and a control unit for controlling at least the imaging unit and the moving unit, the object has a first surface and a second surface on the opposite side of the first surface, and the object is provided with a mark in which the measured value of the position in the Z direction intersecting the first surface and the second surface is known.

Description

觀察裝置、觀察方法和觀察對象物Observation device, observation method and observation object

本發明涉及觀察裝置、觀察方法和觀察對象物。The present invention relates to an observation device, an observation method, and an observation object.

已知有一種鐳射加工裝置,其為了將具備半導體基板和形成於半導體基板的表面上的功能元件層的晶圓分別沿著多條線切斷,而從半導體基板的背面側對晶圓照射鐳射光,來分別沿著多條線在半導體基板的內部形成多排改質區域。專利文獻1(日本特開2017-64746號公報)所記載的鐳射加工裝置包括紅外線攝影機,能夠從半導體基板的背面側觀察形成於半導體基板的內部的改質區域和形成於功能元件層的加工損傷等。There is known a laser processing apparatus that irradiates laser light on the wafer from the back side of the semiconductor substrate in order to cut a wafer including a semiconductor substrate and a functional element layer formed on the surface of the semiconductor substrate along a plurality of lines, respectively. light to form multiple rows of modified regions inside the semiconductor substrate along multiple lines respectively. The laser processing apparatus described in Patent Document 1 (Japanese Patent Laid-Open No. 2017-64746) includes an infrared camera, and can observe the modified region formed inside the semiconductor substrate and the processing damage formed on the functional element layer from the back side of the semiconductor substrate. wait.

但是,在使用從對象物透射的透射光觀察對象物時,例如,存在著藉由使包括透射光的光源、檢測器的攝影單元在Z方向(例如透射光的光軸方向)上移動,一邊使透射光的聚光點在Z方向上移動一邊在多個位置對對象物進行攝影的情況。在此情況下,能夠考慮藉由對檢測出改質區域時的攝影單元的移動量乘以與物鏡的NA、對象物的折射率相應的校正係數,來算出改質區域在對象物內部的位置的測算值。 但是,根據本發明的發明者的發現,隨著裝置狀態、觀察深度(從透射光的射入面至聚光點在Z方向上的距離)的變化,檢測出改質區域時的攝影單元的移動量有時會產生偏差。作為其原因,首先考慮因物鏡的聚光模糊引起的觀察位置偏移。即,在使攝影單元的物鏡的球面像差校正量為一定的情況下,該一定的球面像差校正對於理想狀態有時為弱校正。在此情況下,對象物內的透射光的聚光位置相對變淺,其結果是,檢測出某改質區域時的攝影單元的移動量相對地變大(觀察位置變得更深)。 同樣地,在使攝影單元的物鏡的球面像差校正量為一定的情況下,該一定的球面像差校正對於理想狀態為過校正的情況下,對象物內的透射光的聚光位置相對地變深,其結果是,檢測出某改質區域時的攝影單元的移動量相對地變小(觀察位置變得更淺)。 此外,作為移動量產生偏差的原因,考慮校正環透鏡的操作前後的偏移。即,在攝影單元的物鏡為校正環透鏡的情況下,即使為了調節校正環的像差校正量而操作校正環,相對於像差校正量的變化量,校正環的操作量有時也不一定,其結果是,在校正環的操作前後觀察位置有時會偏移。此外,攝影單元的物鏡的機差、物鏡的拆裝等也是移動量的偏差的一個原因。 這樣,在因各種原因,檢測出改質區域時的攝影單元的移動量產生了偏差的狀態下,當對該移動量乘以一定的校正係數運算改質區域的位置的測算值時,該測算值也會產生偏差。其結果是,難以獲取改質區域的準確的位置。 於是,本發明的目的是提供能夠更準確獲取關於改質區域的位置的資訊的觀察裝置、觀察方法和觀察對象物。 本發明的觀察裝置具備:攝影部,其具有用於使相對於對象物具有透射性的透射光向對象物聚光的聚光透鏡,用於利用透射光對對象物進行攝影;用於使聚光透鏡相對於對象物相對地移動的移動部;和用於至少控制攝影部和移動部的控制部,對象物具有第1面和第1面的相反側的第2面,在對象物設置有與第1面和第2面交叉的Z方向上的位置的實測值已知的標記,控制部執行攝影處理和匯出處理,其中,在攝影處理中,控制攝影部和移動部,使透射光從第1面射入至對象物的內部,使聚光透鏡沿Z方向移動並利用透射光對對象物進行攝影,來獲取作為對象物的內部圖像的包含標記的像的標記圖像,匯出處理在攝影處理之後匯出校正係數,校正係數使得該校正係數與拍攝得到標記圖像時的聚光透鏡的移動量的乘積的值即測算值成為實測值。 本發明的觀察方法具備:準備對象物的準備步驟,對象物具有第1面和第1面的相反側的第2面,在對象物形成有與第1面和第2面交叉的Z方向上的位置的實測值已知的標記;攝影步驟,使相對於對象物具有透射性的透射光從第1面射入至對象物的內部,使用來將透射光聚光的聚光透鏡沿Z方向移動並利用透射光對對象物進行攝影,來獲取作為對象物的內部圖像的包含標記的像;和匯出步驟,在攝影步驟之後匯出校正係數,校正係數使得該校正係數與拍攝得到標記圖像時的聚光透鏡的移動量的乘積的值即測算值成為實測值。 在這些觀察裝置和觀察方法的對象物中,設置有與該第1面和第2面交叉的Z方向上的位置的實測值已知的標記。在這些觀察裝置和觀察方法中,藉由一邊使聚光透鏡移動一邊對這樣的對象物進行攝影,能夠獲取作為對象物的內部圖像的包含標記的像的標記圖像。而且,匯出校正係數,校正係數使得該校正係數與拍攝得到該標記圖像時的聚光透鏡的移動量的乘積的值(測算值)為已知的標記的位置的實測值。即,根據該觀察裝置和觀察方法,能夠匯出與攝影得到標記圖像時的裝置狀態以及拍攝得到標記圖像時的聚光透鏡的移動量(即觀察深度)相應的校正係數。藉此,在利用透射光觀察改質區域來對改質區域的位置的測算值進行運算時,使用該校正係數能夠更準確地獲取關於改質區域的位置的資訊。 本發明的觀察裝置也可以構成為,在對象物中,形成有Z方向上的位置彼此不同且該位置的實測值已知的多個標記,在攝影處理中,控制部使聚光透鏡沿Z方向相對移動,使得透射光的聚光點位於Z方向上的對象物的內部的多個位置來對對象物進行攝影,藉此獲取包括多個標記各自的像的多個標記圖像,在匯出處理中,控制部匯出多個校正係數亦可,多個校正係數使得校正係數與拍攝得到多個標記圖像的各個標記圖像時的聚光透鏡的各個移動量的乘積的值即測算值各自分別成為多個標記的實測值的每一個。在此情況下,能夠匯出與多個移動量相應的校正係數。因此,在利用透射光觀察改質區域來對改質區域的位置的測算值進行運算時,能夠在Z方向上的更寬廣的範圍內,更準確地獲取關於改質區域的位置的資訊。 本發明的觀察裝置也可構成為,在對象物中,作為標記形成有在沿第1面和第2面的X方向上排列的改質區域和從改質區域延伸的裂紋,在攝影處理中,使聚光透鏡沿Z方向移動,來使透射光的聚光點移動而利用透射光對對象物進行攝影,由此獲取包括裂紋中的沿與X方向和Z方向交叉的方向延伸的裂紋的像的內部圖像作為標記圖像。 根據本發明的發明者的發現,當在對象物的內部例如藉由鐳射加工形成改質區域時,有時形成從該改質區域向各種各樣的方向延伸的裂紋。而且,該裂紋中沿與對對象物的鐳射射入面交叉的Z方向與作為鐳射加工的前進方向的X方向交叉的方向延伸的裂紋,與改質區域相較之下,能夠利用從對象物透射的透射光精準地檢測出。因此,如上所述,如果令包含該裂紋的像的內部圖像為標記圖像,則能夠減輕攝影得到該標記圖像時的聚光透鏡的移動量的偏差。其結果是,能夠匯出更準確的校正係數。 本發明的觀察裝置也可以構成為,攝影部具有校正環透鏡,該校正環透鏡包括聚光透鏡和設置於聚光透鏡的用於對在對象物產生的像差進行校正的校正環。這樣,在聚光透鏡設置有校正環的情況下,存在在校正環的操作前後裝置狀態發生變化的問題。因此,如上所述那樣匯出與裝置狀態相應的校正係數更有效。 本發明的觀察裝置也可以構成為具備:用於設置對象物的設置部;和設置於設置部的對象物。這樣,藉由始終設置設有標記的對象物,能夠在任意時間進行校正係數的匯出。 本發明的觀察裝置包括:攝影部,其具有用於使相對於對象物具有透射性的透射光向對象物聚光的聚光透鏡,用於利用透射光對對象物進行攝影;用於使聚光透鏡相對於對象物相對地移動的移動部;和用於至少控制攝影部和移動部的控制部,對象物具有第1面和第1面的相反側的第2面,在對象物中設有改質區域和從改質區域延伸的裂紋,控制部執行攝影處理和運算處理,其中,在攝影處理中,控制攝影部和移動部,使透射光從第1面射入至對象物,使聚光透鏡沿與第1面和第2面交叉的Z方向移動並利用透射光對對象物進行攝影,來獲取作為內部圖像的包含改質區域和/或裂紋的像的檢測圖像,運算處理在攝影處理之後,將校正係數乘以拍攝得到檢測圖像時的聚光透鏡的移動量,來運算改質區域和/或裂紋在Z方向上的位置的測算值,控制部保存有與移動量相應的多個校正係數。 該觀察裝置,如上所述,保存有與聚光透鏡的移動量相應的校正係數。因此,藉由使用該校正係數來運算改質區域的位置的測算值,能夠獲取關於更準確的改質區域的位置的資訊。 本發明的觀察對象物具有第1面和第1面的相反側的第2面,設置有標記,觀察對象物用於匯出校正係數,該校正係數用於從標記的位置的實測值運算與第1面和第2面交叉的Z方向上的標記的位置的測算值。使用該觀察對象物能夠如上所述的那樣匯出校正係數。 根據本發明,能夠提供能夠更準確地獲取關於改質區域的位置的資訊的觀察裝置、觀察方法和觀察對象物。 However, when observing an object using transmitted light transmitted from the object, for example, by moving an imaging unit including a light source of the transmitted light and a detector in the Z direction (for example, the direction of the optical axis of the transmitted light), while The case of photographing an object at a plurality of positions while moving the focused point of the transmitted light in the Z direction. In this case, it is conceivable to calculate the position of the modified region inside the object by multiplying the movement amount of the imaging unit when the modified region is detected by a correction coefficient corresponding to the NA of the objective lens and the refractive index of the object. The estimated value of . However, according to the findings of the inventors of the present invention, as the state of the device and the depth of observation (the distance in the Z direction from the incident surface of the transmitted light to the condensing point) change, the imaging unit when the modified region is detected The amount of movement sometimes deviates. As the cause, first, the shift of the observation position due to the focus blur of the objective lens is considered. That is, when the spherical aberration correction amount of the objective lens of the imaging unit is constant, the constant spherical aberration correction may be weakly corrected in an ideal state. In this case, the converging position of transmitted light in the object becomes relatively shallow, and as a result, the movement amount of the imaging unit when a certain modified region is detected becomes relatively large (observation position becomes deeper). Similarly, when the spherical aberration correction amount of the objective lens of the imaging unit is constant, and the constant spherical aberration correction is overcorrected for the ideal state, the condensing position of the transmitted light in the object is relatively As a result, the movement amount of the imaging unit when a certain modified region is detected becomes relatively small (observation position becomes shallower). In addition, as a cause of variation in the amount of movement, a deviation before and after the operation of the correction ring lens is considered. That is, when the objective lens of the imaging unit is a correction ring lens, even if the correction ring is operated to adjust the aberration correction amount of the correction ring, the operation amount of the correction ring may not be constant with respect to the change amount of the aberration correction amount. , as a result, the viewing position sometimes shifts before and after the operation of the correction collar. In addition, mechanical differences of the objective lens of the imaging unit, detachment of the objective lens, and the like are also one cause of the variation in the amount of movement. In this way, when the movement amount of the imaging unit when the modified region is detected is deviated due to various reasons, when the estimated value of the position of the modified region is calculated by multiplying the movement amount by a fixed correction coefficient, the estimated Values can also be biased. As a result, it is difficult to obtain an accurate position of the modified region. Therefore, an object of the present invention is to provide an observation device, an observation method, and an observation object capable of obtaining more accurate information on the position of a modified region. The observation device of the present invention includes: an imaging unit having a condenser lens for condensing transmitted light having transmittance with respect to the object to the object, for photographing the object with the transmitted light; A moving part that the optical lens relatively moves with respect to the object; and a control part for at least controlling the photographing part and the moving part, the object has a first surface and a second surface on the opposite side of the first surface, and the object is provided with The control unit executes the imaging processing and export processing of the marker whose actual measurement value of the position in the Z direction intersecting the first surface and the second surface is known. In the imaging processing, the imaging unit and the moving unit are controlled so that the transmitted light It enters the inside of the object from the first surface, moves the condenser lens in the Z direction, and shoots the object with the transmitted light to obtain a marker image including a marker image as an internal image of the object, and collects The export process exports correction coefficients such that the calculated value, which is the product of the correction coefficient and the movement amount of the condenser lens when the marker image is captured, becomes the actual measured value after the photographing process. The observation method of the present invention includes: a preparation step of preparing an object, the object has a first surface and a second surface on the opposite side of the first surface, and the object is formed in the Z direction intersecting the first surface and the second surface. In the photographing step, the transmitted light that is transparent to the object is injected into the inside of the object from the first surface, and the condensing lens for condensing the transmitted light is used in the Z direction moving and photographing the object with transmitted light to obtain an image including a mark as an internal image of the object; and an exporting step of exporting a correction coefficient after the photographing step such that the correction coefficient is consistent with the photographed mark The calculated value, which is the product of the moving amount of the condenser lens at the time of imaging, becomes the actual measured value. The objects of these observation devices and observation methods are provided with marks whose actual measured values are known at positions in the Z direction intersecting the first surface and the second surface. In these observation devices and observation methods, by imaging such an object while moving the condenser lens, it is possible to acquire a marker image including an image of the marker as an internal image of the object. Then, a correction coefficient is exported so that the value (measured value) of the product of the correction coefficient and the movement amount of the condenser lens when the mark image is captured is an actual value of the known position of the mark. That is, according to the observation device and observation method, it is possible to derive correction coefficients corresponding to the state of the device when the marker image is captured and the amount of movement of the condenser lens when the marker image is captured (that is, the observation depth). Thereby, when calculating the estimated value of the position of the modified region by observing the modified region with transmitted light, the information on the position of the modified region can be obtained more accurately by using the correction coefficient. The observation device of the present invention may also be configured such that the object is formed with a plurality of marks whose positions in the Z direction are different from each other and whose actual measured values are known, and the control unit moves the condensing lens along the Z direction during the photographing process. Directions are relatively moved so that the focal point of the transmitted light is located at a plurality of positions inside the object in the Z direction to photograph the object, thereby acquiring a plurality of marker images including images of the plurality of markers. In the export process, the control unit may export a plurality of correction coefficients, and the value of the product of the correction coefficient and each movement amount of the condenser lens when each of the plurality of marker images is captured is calculated. Each of the values is each of the measured values of the plurality of markers. In this case, correction coefficients corresponding to a plurality of movement amounts can be exported. Therefore, when calculating the estimated position of the modified region by observing the modified region with transmitted light, more accurate information on the position of the modified region can be obtained in a wider range in the Z direction. The observation device of the present invention may be configured such that modified regions aligned in the X-direction along the first surface and the second surface and cracks extending from the modified regions are formed as marks on the object, and in the imaging process , moving the condenser lens along the Z direction to move the condensing point of the transmitted light and photographing the object with the transmitted light, thereby acquiring cracks including cracks extending in a direction intersecting the X direction and the Z direction. The internal image of the image is used as the marker image. According to the findings of the inventors of the present invention, when a modified region is formed inside an object by, for example, laser processing, cracks extending in various directions from the modified region may be formed. Moreover, among the cracks, the crack extending in the direction intersecting the Z direction intersecting the laser incident surface of the object and the X direction which is the advancing direction of the laser processing can be compared with the modified region by utilizing the The transmitted transmitted light is precisely detected. Therefore, as described above, if the internal image including the image of the crack is used as the marker image, it is possible to reduce the variation in the amount of movement of the condenser lens when the marker image is captured. As a result, more accurate correction coefficients can be exported. The observation device of the present invention may be configured such that the imaging unit has a correction ring lens including a condenser lens and a correction ring provided on the condenser lens for correcting aberrations occurring in the object. Thus, in the case where the condenser lens is provided with a correction ring, there is a problem that the state of the device changes before and after the operation of the correction ring. Therefore, it is more effective to export the correction coefficient according to the state of the device as described above. The observation device of the present invention may be configured to include: a setting unit for setting an object; and an object set on the setting unit. In this way, by always setting the marked object, it is possible to export the correction coefficient at any time. The observation device of the present invention includes: a photographing unit having a condensing lens for converging the transmitted light having transmittance to the target object on the target object for photographing the target object with the transmitted light; A moving part that the optical lens relatively moves with respect to the object; and a control part for controlling at least the photographing part and the moving part, the object has a first surface and a second surface on the opposite side of the first surface, and the object is provided with There is a modified region and a crack extending from the modified region, and the control unit executes imaging processing and calculation processing, wherein in the imaging processing, the imaging unit and the moving unit are controlled so that transmitted light enters the object from the first surface, so that The condenser lens is moved in the Z direction intersecting the first surface and the second surface, and the object is photographed with transmitted light to obtain a detection image including a modified region and/or a crack image as an internal image, and the calculation Processing After the photographing process, the correction coefficient is multiplied by the movement amount of the condenser lens when the detection image is captured to calculate the estimated value of the position of the modified region and/or crack in the Z direction, and the control unit saves and moves The amount of corresponding multiple correction coefficients. This observation device stores a correction coefficient corresponding to the movement amount of the condenser lens as described above. Therefore, by calculating the estimated value of the position of the modified region using the correction coefficient, more accurate information on the position of the modified region can be obtained. The observation object of the present invention has a first surface and a second surface on the opposite side of the first surface, on which a mark is provided, and the observation object is used to derive a correction coefficient for calculating and calculating from an actual measured value at the position of the mark. The measured value of the position of the mark in the Z direction where the first surface and the second surface intersect. Using this observation object, it is possible to derive the correction coefficient as described above. According to the present invention, it is possible to provide an observation device, an observation method, and an observation object capable of more accurately acquiring information on the position of a modified region.

下面,參照附圖對一個實施型態進行詳細說明。其中,在各附圖的說明中,有時對於相同或相當的部分賦予相同符號,省略重複的說明。此外,在各圖中,有時表示由X軸、Y軸和Z軸規定的直角座標系。作為一例,X方向和Y方向是彼此交叉(正交)的第1水平方向和第2水平方向,Z方向是與X方向和Y方向交叉(正交)的鉛垂方向。 如圖1所示,鐳射加工裝置1具備載置台2、鐳射照射單元3(照射部)、多個攝影單元4、5、6、驅動單元7、控制部8和顯示器150(顯示部)。鐳射加工裝置1是藉由向對象物11照射鐳射光L來在對象物11形成改質區域12的裝置。 載置台2例如藉由吸附黏貼於對象物11的膜來支撐對象物11。載置台2能夠分別沿著X方向和Y方向移動,且能夠以與Z方向平行的軸線為中心線旋轉。 鐳射照射單元3將對於對象物11具有透射性的鐳射光L聚光來照射於對象物11。當鐳射光L聚光至由載置台2支撐的對象物11的內部時,在與鐳射光L的聚光點C對應的部分,鐳射光L特別會被吸收,能夠在對象物11的內部形成改質區域12。 改質區域12是密度、折射率、機械強度或其他物理特性與周圍的非改質區域不同的區域。作為改質區域12,例如有熔融處理區域、裂縫區域、絕緣破壞區域、折射率變化區域等。改質區域12具有裂紋容易從改質區域12延伸至鐳射光L的射入側及其相反側的特性。這樣的改質區域12的特性被利用於對象物11的切斷。 作為一例,當使載置台2沿著X方向移動,並使聚光點C相對於對象物11沿著X方向相對地移動時,以沿著X方向排成1排的方式形成多個改質點12s。1個改質點12s是藉由1個脈衝的鐳射光L的照射而形成的。1排改質區域12是排成1排的多個改質點12s的集合。相鄰的改質點12s根據聚光點C相對於對象物11的相對移動速度和鐳射光L的反復頻率,存在彼此相連的情況,以及彼此分開的情況。 攝影單元4拍攝形成於對象物11中的改質區域12,以及從改質區域12延伸出的裂紋的前端。 攝影單元5和攝影單元6在控制部8的控制下,利用從對象物11透射的光來拍攝被載置台2所支撐的對象物11。攝影單元5、6進行拍攝而獲得的圖像,作為一例,用於進行鐳射光L的照射位置的對準。 驅動單元7支撐鐳射照射單元3和多個攝影單元4、5、6。驅動單元7使鐳射照射單元3和多個攝影單元4、5、6沿著Z方向移動。 控制部8控制載置台2、鐳射照射單元3、多個攝影單元4、5、6和驅動單元7的動作。控制部8作為包含處理器、記憶體、暫存器和通訊裝置等的電腦裝置而構成。在控制部8中,處理器執行記憶體等中所讀取的軟體(程式),控制記憶體和暫存器中的資料的讀出或寫入,以及通訊裝置所進行的通訊。 顯示器150具有作為接受使用者進行的資訊的輸入的輸入部的功能、和作為對使用者顯示資訊的顯示部的功能。 [對象物的結構] 本實施型態的對象物11如圖2及圖3所示為晶圓20。晶圓20具備半導體基板21和功能元件層22。在本實施型態中,說明晶圓20包括功能元件層22,但是晶圓20既可以具有功能元件層22,也可以不具有功能元件層22,也可以是裸晶圓。半導體基板21具有表面21a(第2面)和背面21b(第1面)。半導體基板21例如為矽基板。功能元件層22形成於半導體基板21的表面21a。功能元件層22包含沿著表面21a二維排列的多個功能元件22a。功能元件22a是例如光電二極體等的受光元件、鐳射二極體等的發光元件、記憶體等的電路元件等。功能元件22a也存在堆疊多個層而三維地構成的情況。另外,在半導體基板21,雖然設有顯示結晶方位的缺口21c,但是也可以取代缺口21c而設置定向平面。 晶圓20分別沿著多條線15按每個功能元件22a被切斷。多條線15從晶圓20的厚度方向觀察時,通過多個功能元件22a各自之間。更具體而言,線15從晶圓20的厚度方向觀察時,通過格線區域23的中心(寬度方向的中心)。格線區域23在功能元件層22以通過相鄰的功能元件22a之間的方式延伸。在本實施型態中,多個功能元件22a沿著表面21a以矩陣狀排列,多條線15設定為格子狀。另外,線15雖是虛擬線,但也可是實際上劃出的線。 [鐳射照射單元的結構] 如圖4所示,鐳射照射單元3具有光源31、空間光調變器32和聚光透鏡33。光源31例如藉由脈衝振盪方式來輸出鐳射光L。空間光調變器32調變從光源31輸出的鐳射光L。空間光調變器32例如是反射型液晶(LCOS:Liquid Crystal on Silicon)的空間光調變器(SLM:Spatial Light Modulator)。聚光透鏡33將藉由空間光調變器32調變的鐳射光L聚光。其中,聚光透鏡33也可以是校正環透鏡。 在本實施型態中,鐳射照射單元3分別沿著多條線15從半導體基板21的背面21b側對晶圓20照射鐳射光L,由此分別沿著多條線15在半導體基板21的內部形成2排改質區域12a、12b。改質區域12a是2排改質區域12a、12b中的最靠近表面21a的改質區域。改質區域12b是2排改質區域12a、12b中的最靠近改質區域12a的改質區域,且是最靠近背面21b的改質區域。 2排改質區域12a、12b在晶圓20的厚度方向(Z方向)上相鄰。2排改質區域12a、12b是藉由使2個聚光點C1、C2相對於半導體基板21沿著線15相對地移動而形成的。藉由空間光調變器32調變鐳射光L,使得例如聚光點C2相對於聚光點C1位於行進方向的後側且位於鐳射光L的射入側。其中,關於改質區域的形成,可以是單焦點,也可以是多焦點,可以是1個路徑,也可以是多個路徑。 鐳射照射單元3沿多個線15中的各個線15從半導體基板21的背面21b側對晶圓20照射鐳射光L。作為一例,相對於作為厚度400μm的單晶矽<100>基板的半導體基板21,使2個聚光點C1、C2分別對焦到與表面21a相距54μm的位置和至128μm的位置,沿多個線15的各個線15從半導體基板21的背面21b側對晶圓20照射鐳射光L。此時,例如在選取遍及2排改質區域12a、12b的裂紋14到達半導體基板21的表面21a的條件的情況下,令鐳射光L的波長為1099nm,脈衝寬度為700n秒,反復頻率為120kHz。此外,令聚光點C1的鐳射光L的輸出為2.7W,聚光點C2的鐳射光L的輸出為2.7W,令2個聚光點C1、C2相對於半導體基板21的相對移動速度為800mm/秒。其中,例如在加工路徑數為5的情況下,對於上述的晶圓20,例如也可以令ZH80(與表面21a相距328μm的位置)、ZH69(與表面21a相距283μm的位置)、ZH57(與表面21a相距234μm的位置)、ZH26(與表面21a相距107μm的位置)、ZH12(與表面21a相距49.2μm的位置)為加工位置。在此情況下,例如,也可以是,鐳射光L的波長為1080nm,脈衝寬度為400nsec,反復頻率為100kHz,移動速度為490mm/秒。 這樣的2排改質區域12a、12b和裂紋14的形成是在如下那樣的情況下實施的。即,在之後的步驟中,例如藉由研磨半導體基板21的背面21b而使半導體基板21薄化,並且使裂紋14露出到背面21b,分別沿著多條線15將晶圓20切斷為多個半導體元件的情況。 [檢查用攝影單元的結構] 如圖5所示,攝影單元4(攝影部)具有光源41、反射鏡42、物鏡43(聚光透鏡)和光檢測部44。攝影單元4對晶圓20進行攝影。光源41輸出相對於半導體基板21具有透射性的光I1。光源41例如由鹵素燈和濾光片構成,輸出近紅外線區域的光I1。從光源41輸出的光I1會被反射鏡42反射而通過物鏡43,從半導體基板21的背面21b側照射至晶圓20。此時,載置台2如上述那樣支撐形成有2排改質區域12a、12b的晶圓20。 物鏡43用於使相對於半導體基板21具有透射性的光(透射光)I1朝向半導體基板21聚光。物鏡43使被半導體基板21的表面21a反射後的光I1通過。即,物鏡43使在半導體基板21中傳播後的光I1通過。物鏡43的數值孔徑(NA)例如是0.45以上。物鏡43具有校正環43a。校正環43a例如藉由調節構成物鏡43的多個透鏡的彼此之間的距離,來校正半導體基板21內的光I1所產生的像差。其中,作為校正像差的手段,不限於校正環43a,也可以是空間光調變器等其他校正手段。光檢測部44對從物鏡43和反射鏡42透射後的光I1進行檢測。光檢測部44例如由InGaAs攝影機構成,對近紅外線區域的光I1進行檢測。其中,檢測(攝影)近紅外線區域的光I1的手段不限於InGaAs攝影機,也可以是透射型共聚焦顯微鏡等進行透射型的攝影的其它攝影手段。 攝影單元4能夠拍攝出2排改質區域12a、12b各者以及多個裂紋14a、14b、14c、14d各自的前端(詳情後述)。裂紋14a是從改質區域12a向表面21a側延伸的裂紋。裂紋14b是從改質區域12a向背面21b側延伸的裂紋。裂紋14c是從改質區域12b向表面21a側延伸的裂紋。裂紋14d是從改質區域12b向背面21b側延伸的裂紋。 [對準校正用攝影單元的結構] 如圖6所示,攝影單元5具有光源51、反射鏡52、透鏡53和光檢測部54。光源51輸出相對於半導體基板21具有透射性的光I2。光源51例如由鹵素燈及濾光片構成,輸出近紅外線區域的光I2。光源51也可與攝影單元4的光源41共用。從光源51輸出的光I2會被反射鏡52反射而通過透鏡53,從半導體基板21的背面21b側照射於晶圓20。 透鏡53使被半導體基板21的表面21a反射後的光I2通過。即,透鏡53使在半導體基板21中傳播後的光I2通過。透鏡53的數值孔徑是0.3以下。即,攝影單元4的物鏡43的數值孔徑比透鏡53的數值孔徑大。光檢測部54檢測藉由透鏡53和反射鏡52的光I2。光檢測部54例如由InGaAs攝影機構成,檢測近紅外線區域的光I2。 攝影單元5在控制部8的控制下,從背面21b側將光I2照射於晶圓20,並且檢測從表面21a(功能元件層22)返回的光I2,藉此拍攝功能元件層22。並且,攝影單元5同樣在控制部8的控制下,從背面21b側將光I2照射於晶圓20,並且檢測從半導體基板21的改質區域12a、12b的形成位置返回的光I2,藉此獲取包含改質區域12a、12b的區域的圖像。這些圖像用於進行鐳射光L的照射位置的校準。攝影單元6除了相較於透鏡53為低倍率(例如,在攝影單元5中是6倍,在攝影單元6中是1.5倍)這點以外,具有與攝影單元5同樣的結構,並與攝影單元5同樣地用於校準。 [檢查用攝影單元的拍攝原理] 使用圖5所示的攝影單元4,如圖7所示,對於跨2排改質區域12a、12b的裂紋14到達了表面21a的半導體基板21,使焦點F(物鏡43的焦點)從背面21b側向表面21a側移動。在該情況下,如果使焦點F從背面21b側對焦到從改質區域12b向背面21b側延伸的裂紋14的前端14e,則能夠確認到該前端14e(圖7的右側的圖像)。然而,即便使焦點F從背面21b側對焦到裂紋14本身和到達了表面21a的裂紋14的前端14e,也無法進行確認(圖7的左側的圖像)。另外,如果使焦點F從背面21b側對焦到半導體基板21的表面21a,則能夠確認到功能元件層22。 並且,使用圖5所示的攝影單元4,如圖8所示,對於跨2排改質區域12a、12b的裂紋14未到達表面21a的半導體基板21,使焦點F從背面21b側向表面21a側移動。在該情況下,即便使焦點F從背面21b側對焦到從改質區域12a向表面21a側延伸的裂紋14的前端14e,也無法確認到該前端14e(圖8的左側的圖像)。然而,如果使焦點F從背面21b側對焦到相對於表面21a位於與背面21b相反側的區域(即,相對於表面21a位於功能元件層22側的區域),關於表面21a使與焦點F對稱的虛擬焦點Fv位於該前端14e,則能夠確認到該前端14e(圖8的右側的圖像)。另外,虛擬焦點Fv是考慮了半導體基板21的折射率的關於表面21a與焦點F對稱的點。 以上那樣無法確認到裂紋14,推測是因為裂紋14的寬度比作為照明光的光I1的波長小。圖9和圖10是形成於作為矽基板的半導體基板21的內部的改質區域12及裂紋14的SEM(Scanning Electron Microscope)圖像。圖9的(b)是圖9的(a)所示的區域A1的放大圖像,圖10的(a)是圖9的(b)所示的區域A2的放大圖像,圖10的(b)是圖10的(a)所示的區域A3的放大圖像。像這樣,裂紋14的寬度是120nm左右,比近紅外線區域的光I1的波長(例如,1.1~1.2μm)還小。 根據以上事項所設想的攝影原理如下所述。如圖11的(a)所示,如果使焦點F位於空氣中,則光I1不會返回,因此會獲得漆黑的圖像(圖11的(a)的右側的圖像)。如圖11的(b)所示,如果使焦點F位於半導體基板21的內部,則被表面21a反射的光I1會返回,故會獲得白淨的圖像(圖11的(b)的右側的圖像)。如圖11的(c)所示,如果使焦點F從背面21b側對焦到改質區域12,則會因改質區域12使被表面21a反射而返回的光I1的一部分產生吸收、散射等,所以會獲得在白淨的背景中顯示出漆黑的改質區域12的圖像(圖11的(c)的右側的圖像)。 如圖12的(a)及(b)所示,如果使焦點F從背面21b側對焦到裂紋14的前端14e,則例如會因產生於前端14e附近的光學特異性(應力集中、歪曲、原子密度的不連續性等)使光被局限在前端14e附近等,由此使被表面21a反射而返回的光I1的一部分產生散射、反射、干涉、吸收等,所以會獲得在白淨的背景中顯示出漆黑的前端14e的圖像(圖12的(a)及(b)的右側的圖像)。如圖12的(c)所示,如果使焦點F從背面21b側對焦到裂紋14的前端14e附近以外的部分,則被表面21a反射的光I1有至少一部分會返回,故會獲得淨白的圖像(圖12的(c)的右側的圖像)。 [內部觀察的實施型態] 圖13是表示形成了改質區域的對象物的圖。圖13的(a)是以使改質區域露出的方式被切斷的對象物的截面照片。圖13的(b)是利用從對象物透射的光進行攝影而得到的對象物的圖像的一例。圖13的(c)是利用從對象物透射的光進行攝影而得到的對象物的圖像的另一例。如圖13的(a)所示,藉由鐳射光L的聚光,形成在對象物(這裡是半導體基板21)的改質區域12包括:位於半導體基板21的鐳射光L的射入面的相反側的面即表面21a側的缺陷(Void)區域12m;和比缺陷區域12m靠鐳射光L的射入面即背面21b側的缺陷上方區域12n。 當利用相對於半導體基板21具有透射性的光I1對形成了這樣的改質區域12的半導體基板21進行攝影時,如圖13的(b)、(c)所示,有時能夠確認到沿與Z方向和X方向交叉的方向延伸(相對於X方向具有角度)的裂紋14k的像。從Z方向看時,裂紋14k在圖13的(b)的例子中與Y方向大致平行,在圖13的(c)的例子中相對於Y方向稍微傾斜。這些裂紋14k的像,在一邊使光I1的聚光點沿Z方向移動一邊在多個位置對半導體基板21進行攝影時,與改質區域12相較之下,能夠在Z方向上的限定範圍內被清晰地檢測出。 圖14是關於Z方向上的改質區域和裂紋的位置的圖表。在圖14中,缺陷下端、缺陷上端、缺陷上方區域下端、和缺陷上方區域上端的繪製,是藉由截面觀察實際測定到的實測值。下端是指表面21a側的端部,上端是指背面21b側的端部。因此,例如缺陷上方區域下端是指缺陷上方區域12n的表面21a側的端部。 此外,圖14的圖表中的直接觀察和背面反射觀察的繪製,是基於利用光I1攝影得到的圖像中包含裂紋14k清晰的像的內部圖像被攝影得到時的Z方向的物鏡43的移動量(下面有時簡稱為“移動量”)運算得到的測算值,作為一例,是藉由基於AI的圖像判斷得到的值。直接觀察,是使光I1從背面21b射入,且不經過表面21a的反射直接使光I1的聚光點對準裂紋14k的情況(在上述的例子中,從背面21b側使焦點F與裂紋14k一致的情況),背面反射觀察,是使光I1從背面21b射入,被表面21a反射了的光I1的聚光點對準裂紋14k的情況(在上述的例子中,相對於表面21a,從背面21b側使焦點F對準背面21b的相反側的區域,使關於表面21a與焦點F對稱的虛擬焦點Fv對準裂紋14k的情況)。 如圖14所示,在直接觀察中,在使改質區域12的形成位置在Z方向上位於不同的4個的情況C1~C4下,都在缺陷上方區域下端與缺陷上方區域上端之間檢測出裂紋14k,在背面反射觀察中,在情況C1下在大致與缺陷上方區域下端相同的位置檢測出裂紋14k,在情況C2~C4下在缺陷上方區域下端與缺陷上端之間檢測出裂紋14k。Z方向上的改質區域12的寬度,是缺陷下端與缺陷上方區域上端之間的距離。這樣,裂紋14k與改質區域12本身相較之下,能夠在Z方向上被更精準地檢測出。 因此,藉由獲取在Z方向上出現裂紋14k時的內部圖像的移動量,能夠更準確地獲取關於改質區域12的位置的資訊。其中,圖14的縱軸表示與背面的距離,這裡的背面是相對於光I1的射入面的背面,對半導體基板21而言是表面21a。此外,圖15是在截面照片對情況C1下的檢測結果進行繪製而得到的。 在本實施型態中,基於以上那樣的發現,藉由內部觀察檢測裂紋14k,獲取關於改質區域12的位置的資訊。接著對本實施型態的觀察方法進行說明。在該觀察方法中,裂紋14k是檢測對象的對象裂紋。 在本實施型態中,在獲取關於改質區域12的位置的資訊時,首先,對Z方向上的裂紋14k的位置進行運算。此時,對檢測出裂紋14k的內部圖像被攝影得到時的物鏡43在Z方向上的移動量,乘以規定的校正係數。如圖14、15所示,裂紋14k在Z方向上的改質區域12的範圍內被檢測出。因此,運算得到的裂紋14k的位置,也就是改質區域12在Z方向上的位置的測算值。 首先,對關於校正係數的發現進行說明。如圖16所示,為了調節半導體基板21內的光I1的聚光點的位置,使用驅動單元7使攝影單元4沿Z方向移動了移動量Fz。此時,如果沒有半導體基板21,則光I1的聚光點的移動量也為移動量Fz。但是,在光I1的聚光點形成在半導體基板21的內部的情況下,光I1的聚光點的移動量為與移動量Fz不同的移動量Hz(在圖示的例子中被放大)。移動量Hz規定半導體基板21內的實際的攝影位置,即,檢測對象(例如改質區域12、裂紋14k)的位置。 另一方面,控制部8能夠直接獲取的資訊,是作為驅動單元7的控制時的輸入值的攝影單元4的移動量Fz(即,沒有半導體基板21的情況下的聚光點的移動量Fz)。因此,控制部8為了獲取半導體基板21內的實際的檢測對象的位置,需要對移動量Fz乘以某係數。此時應用的係數是校正係數。該校正係數,考慮到物鏡43的NA、半導體基板21的折射率,能夠設定為一定的值(例如在半導體基板21為矽的情況下,為4左右)。但是,在令校正係數為一定的值的情況下,可能產生如下的問題。 圖17是表示檢測對象在Z方向的位置與檢測出檢測對象時的移動量的關係的圖表。圖17的圖表的橫軸的“深度位置”,是設置了檢測對象的Z方向的位置(使光I1的聚光點對準的位置),圖17的圖表的縱軸的“Z軸移動量”,是在Z方向的各個位置檢測出檢測對象時的移動量Fz。在半導體基板21內以40μm左右的間隔設置檢測對象,以使得在令移動量Fz為10μm時能夠檢測出檢測對象。 在圖17中,一併記載了裝置狀態不同的多個情況下的曲線圖,在關注1個裝置狀態(1個曲線圖)的情況下,當檢測對象在Z方向上的位置不同時,本來應當固定為10μm的移動量Fz產生偏差。在多個裝置狀態(多個曲線圖)之間進行比較的情況下也相同。這樣在移動量Fz產生偏差的原因,首先,考慮因物鏡43的聚光模糊而引起的觀察位置偏移。即,在使攝影單元4的物鏡43的球面像差校正量為一定的情況下,該一定的球面像差校正對於理想狀態為弱校正。在此情況下,半導體基板21內的光I1的聚光位置相對地變淺,其結果是,檢測出某改質區域12時的攝影單元4的移動量Fz相對地變大(觀察位置變得更深)。 同樣,在使攝影單元4的物鏡43的球面像差校正量為一定的情況下,該一定的球面像差校正對於理想狀態為過校正時,半導體基板21內的光I1的聚光位置相對變深,其結果是,檢測出某改質區域12時的攝影單元4的移動量Fz相對地變小(觀察位置變得更淺)。 此外,作為移動量Fz產生偏差的原因,能夠考慮校正環透鏡的操作前後的偏移。即,在攝影單元4的物鏡43為校正環透鏡的情況下,即使為了利用校正環43a調節像差校正量而操作校正環43a,相對於像差校正量的變化量,校正環43a的操作量有時也不一定,其結果是,在校正環43a的操作前後存在觀察位置偏移的情況。進一步,攝影單元4的物鏡43的機差、物鏡43的拆裝等也為移動量Fz的偏差的原因。 當這樣對產生了偏差的移動量Fz乘以一定的校正係數而運算測算值時,運算結果也會產生偏差。據此,為了獲取關於檢測對象的準確位置的資訊,需要根據裝置狀態、Z方向的位置使用適當的校正係數。於是,在本實施型態的觀察方法中,如下所述地進行校正係數的匯出。進行校正係數的匯出的時機是任意的,作為一例,是在如物鏡43的拆裝時之裝置狀態發生了變化的時機。 圖18是表示用於匯出本實施型態的觀察方法中的校正係數的步驟的流程圖。如圖18所示,將用於匯出校正係數的對象物(觀察對象物)60移動至攝影單元4的物鏡43的下部(步驟S1)。如圖1所示,鐳射加工裝置1還具備與載置台2不同的載置台(設置部)2A,對象物60載置在該載置台2A。載置台2A例如藉由驅動單元7能夠沿X方向和Y方向移動。 圖19是表示用於匯出校正係數的對象物的側視圖。如圖19所示,對象物60包括背面60b和背面(第1面)60b的相反側的表面(第2面)60a。在對象物60中,藉由鐳射加工,形成了在沿背面60b和表面60a的X方向上排列的改質區域12和從改質區域12延伸的裂紋(裂紋14、14k)。尤其是,在對象物60形成了從改質區域12沿與Z方向和X方向交叉的方向延伸的裂紋14k。此外,在對象物60中,以在Z方向上排列的方式形成了多排改質區域12。改質區域12在Z方向上的間隔,換算成移動量Fz為10μm以下。 在對象物60中,以使改質區域12露出的方式形成了切斷面,各個改質區域12在Z方向上的位置,例如作為裂紋14k的位置,藉由該切斷面的觀察被實測而為已知。該已知的實測值,可以由控制部8保存,也可以保存在控制部8能夠聯絡的任意的儲存裝置中。這樣,在這裡,實施準備對象物60的準備步驟,該對象物60包含背面60b和背面60b的相反側的表面60a,作為與背面60b和表面60a交叉的Z方向上的位置的實測值已知的標記,改質區域12和裂紋14k形成在對象物60中。 在接下來的步驟中,如圖20所示,利用相對於對象物60具有透射性的光(透射光)I1,進行對象物60的攝影(步驟S2:攝影步驟)。在該步驟S2中,藉由對攝影單元4(攝影部)的控制,執行下述的攝影處理:利用光I1對從改質區域12延伸的裂紋中的沿與Z方向和X方向交叉的方向延伸的裂紋14k進行攝影。Y方向是與X方向及Z方向交叉的方向的一例,X方向是用於在對象物60形成改質區域12的鐳射加工的加工前進方向(即改質區域12的排列方向),Z方向是與背面60b和表面60a交叉的方向。 在該步驟S2中,控制部8藉由控制攝影單元4和驅動單元7,使光I1從對象物60的背面60b射入至對象物60,並使攝影單元4沿Z方向移動而使光I1的聚光點(焦點F,虛擬焦點Fv)沿Z方向移動,使光I1的聚光點位於對象物60的內部的多個位置而多次對對象物60進行攝影。藉此,獲取多個內部圖像GD。在本實施型態中,物鏡43與攝影單元4一體地移動。因此,使攝影單元4移動也就是使物鏡43移動,攝影單元4的移動量與物鏡43的移動量是同等的。 使光I1的聚光點移動的範圍,可以是對象物60的厚度的整個範圍,但是在這裡,能夠選取一部分的範圍RA,該一部分的範圍RA包含為了形成改質區域12(在這裡,作為一例是改質區域12a、12b)而使鐳射光的聚光點對準的Z方向的位置。進行多次攝影時攝影單元4在Z方向上的移動間隔,即,對象物60的攝影間隔是任意的,從更準確地檢測裂紋14k的觀點來看,更精細地設定較佳。攝影間隔,作為一例為1μm以內,這裡是0.2μm。 進一步,這裡,控制部8控制攝影單元4和驅動單元7,以執行對象物60的直接觀察和背面反射觀察。更具體而言,控制部8,首先,執行下述的第1攝影處理:使光I1從背面60b射入至對象物60,並使攝影單元4沿Z方向移動,一邊使沒有經過表面60a的反射的光I1的聚光點(焦點F)從背面60b側向表面60a側移動,一邊在Z方向的多個位置對對象物60進行攝影,作為內部圖像GD獲取多個第1內部圖像GD1。該第1攝影處理是直接觀察。 除此之外,控制部8執行下述的第2攝影處理:使光I1從背面60b射入至對象物,使攝影單元4沿Z方向移動,藉此一邊使被表面60a反射的光I1的聚光點(虛擬焦點Fv)從表面60a側向背面60b側移動,一邊在多個位置對對象物60進行攝影,由此,作為內部圖像GD獲取多個第2內部圖像GD2。該第2攝影處理由於是從光I1的射入面的背面(這裡,根據與半導體基板21的表面背面的關係,稱為表面60a)側進行的觀察,因此是背面反射觀察。 在接下來的步驟中,保存關於藉由步驟S2的攝影而獲取的內部圖像GD的攝影資料(步驟S3)。如上所述,在步驟S2中,控制部8一邊藉由對驅動單元7的控制使攝影單元4(即光I1的聚光點)沿Z方向移動一邊進行攝影。因此,控制部8能夠獲取攝影得到各個內部圖像時的移動量Fz。這裡,關於各個內部圖像GD的移動量Fz的資訊能夠與各個內部圖像GD相關聯,並作為攝影資料被保存。攝影資料,不論控制部8和鐳射加工裝置1的內外,能夠被保存在控制部8能夠聯絡的任意儲存裝置中。 攝影單元4(物鏡43)的移動量Fz,作為一例,是從使光I1的聚光點對準半導體基板21的背面21b的狀態的位置起,以使光I1的聚光點對準半導體基板21的內部的所希望的位置的方式使攝影單元4沿Z方向移動的情況下的攝影單元4的移動量。 接著,控制部8從規定的儲存裝置輸入攝影資料(步驟S4)。然後,控制部8判斷裂紋14k的形成狀態(步驟S5)。這裡,作為一例,控制部8藉由圖像識別自動地判斷多個內部圖像GD中裂紋14k的像相對地清晰的內部圖像GD(標記圖像)(進行AI判斷)。這裡,對藉由AI判斷來檢測裂紋、改質區域的演算法的一例進行說明。 圖29和圖30是對裂紋檢測進行說明的圖。圖29圖示了內部觀察結果(半導體基板21的內部圖像)。控制部8對於圖29的(a)所示的半導體基板21的內部圖像,首先,檢測直線組140。在直線組140的檢測中,例如能夠使用Hough變換或LSD(Line Segment Detector:直線段檢測演算法)等演算法。Hough變換,是對通過圖像上的所有點的全部直線進行檢測,對更多地通過特徵點的直線賦予權重而檢測直線的方法。LSD是,藉由計算圖像內的亮度值的斜率和角度而推算成為線段的區域,藉由將該區域近似為矩形而檢測直線的方法。 接著,控制部8藉由如圖30所示的那樣對直線組140運算與裂紋線的類似度,從直線組140檢測裂紋14。裂紋線,如圖30的上圖所示,具有相對於線上的亮度值在Y方向上前後非常明亮的特徵。因此,控制部8例如將檢測出的直線組140的所有畫素的亮度值與Y方向的前後比較,將其差在前後都是閾值以上的畫素數量作為類似度的分數。然後,將檢測出的多個直線組140中與裂紋線的類似度的分數最高者作為該圖像的代表值。代表值越高,成為存在裂紋14的可能性越高的指標。控制部8藉由將多個圖像的代表值相比較,將分數相對較高的圖像作為裂紋圖像候選。 圖31~圖33是對傷痕檢測進行說明的圖。圖31圖示了內部觀察結果(半導體基板21的內部圖像)。控制部8對於圖31的(a)所示的半導體基板21的內部的圖像,將圖像內的角落(邊緣的聚集)作為關鍵點檢測出,檢測其位置、大小、方向而檢測出特徵點250。這樣檢測特徵點的方法,已知Eigen,Harris,Fast,SIFT,SURF,STAR,MSER,ORB,AKAZE等。 這裡,如圖32所示,傷痕280因為圓形、矩形等形狀以一定間隔排列,因此作為角落的特徵強。因此,藉由統計圖像內的特徵點250的特徵量,能夠高精度地檢測出傷痕280。如圖33所示,比較在深度方向上轉移而攝影得到的每個圖像的特徵量合計,能夠確認表示每個改質層的裂紋排量的值的變化。控制部8將該變化的峰值推算為傷痕280的位置。藉由這樣統計特徵量,不僅能夠推算傷痕位置,而且能夠推算脈衝間距。 對於以上的AI判斷的說明,是關於沿X方向延伸的裂紋14和傷痕280者,但是沿與Z方向和X方向交叉的方向延伸的裂紋14k,也能夠利用相同的演算法,藉由比較多個內部圖像ID的代表值,將分數相對較高的判斷為該裂紋14k的像相對清晰的內部圖像ID。 作為一例,圖21是在Z方向上彼此不同的位置攝影得到的多個內部圖像GD。在圖21中,以(d)所示的內部圖像GDd的攝影位置為中心,(c)是向背面60b側去1μm的攝影位置的內部圖像GDc,(b)是向背面60b側去3μm的攝影位置的內部圖像GDb,(a)是向背面60b側去5μm的攝影位置的內部圖像GDa,(e)是向表面60a側去1μm的攝影位置的內部圖像GDe,(f)是向表面60a側去3μm的攝影位置的內部圖像GDf,(g)是向表面60a側去5μm的攝影位置的內部圖像GDg。這裡的攝影位置是對象物60的內部的值。 在圖21所示的例子中,內部圖像GDd中裂紋14k的像最清晰,於是控制部8判斷內部圖像GDd是相對地分數較高且該裂紋14k的像相對地清晰的內部圖像。即,這裡,判斷在內部圖像GDd中檢測出了裂紋14k(令內部圖像GDd為標記圖像)。控制部8能夠獲取攝影得到內部圖像GDd時的移動量Fz。控制部8藉由對Z方向的位置不同的多排改質區域12和從各個改質區域12延伸的裂紋14k,進行同樣的步驟、處理,能夠獲取檢測出從多排改質區域12的各者延伸的裂紋14k時的移動量Fz。 即,在本實施型態中,在對象物60中形成了在Z方向上的位置彼此不同並且作為標記該位置的實測值已知的多個改質區域12和裂紋14k,在攝影處理中,控制部8藉由使攝影單元4沿Z方向移動,使光I1的聚光點位於Z方向上的對象物60的內部的多個位置來對對象物60進行攝影,藉此獲取包含從多個改質區域12的各者延伸的各個裂紋14k的清晰的像的多個內部圖像GD(與上述的內部圖像GDd相當的圖像,亦即標記圖像)。然後,控制部8如圖22的第2欄Q2所示的那樣,獲取攝影得到各個標記圖像時的移動量Fz。 如圖22的第1欄Q1所示的那樣,控制部8將多排改質區域12(裂紋14k)在Z方向上的位置的實測值,作為與對象物60的背面(相對於光I1的射入面的背面,這裡是表面60a)相距的距離而獲取。 接著,控制部8匯出校正係數(步驟S6:匯出步驟)。如圖22所示,控制部8,作為用於對在Z方向上排列的多個改質區域12各自的位置的測算值進行運算的資訊,是獲取檢測出從各個改質區域12延伸的裂紋14k時的移動量Fz(第2欄Q2)。另一方面,控制部8能夠獲取在Z方向上排列的多個改質區域12各自的位置的實測值(第1欄Q1)。因此,控制部8對於在Z方向上排列的多個改質區域12的各者,能夠以使得對移動量Fz乘以校正係數而得到的值即改質區域12的位置的測算值成為對應的實測值的方式,匯出校正係數。換言之,控制部8,作為校正係數=實測值/移動量Fz,匯出校正係數。 即,控制部8執行下述的匯出處理:以使得對攝影得到包含裂紋14k清晰的像的內部圖像GD即標記圖像時的移動量Fz乘以校正係數而得到的值即測算值,成為改質區域12在Z方向上的位置的實測值的方式,匯出校正係數。圖22的第3欄Q3表示這樣匯出的校正係數。之後,控制部8保存表示匯出的校正係數的資料(步驟S7),結束處理。 藉由上述方式匯出的校正係數,是基於利用來自攝影單元4的光I1攝影得到的內部圖像GD而匯出的。因此,校正係數反映了進行內部圖像GD的攝影時的攝影單元4的裝置狀態。此外,校正係數基於在Z方向上在對象物60的多個位置攝影得到的內部圖像GD而被匯出。因此,校正係數是將使光I1的聚光點在對象物60內對準的Z方向上的位置、和與該位置相應的像差校正量考慮在內而得到的。 接著,在本實施型態的觀察方法中,藉由對包含沒有實測到Z方向的位置的改質區域12的對象物進行觀察,實施用於獲取關於該改質區域12在Z方向上的位置的資訊的一系列的步驟。圖23是表示本實施型態的觀察方法中,用於獲取關於改質區域在Z方向上的位置的資訊的步驟的流程圖。 如圖23所示,這裡,準備形成了改質區域的對象物。在此,進行鐳射加工(步驟S11:準備步驟)。但是,作為觀察方法的一個步驟,鐳射加工的步驟並不是必不可少的,例如也可以準備使用其他鐳射加工裝置(或者,利用鐳射加工裝置1在其他時間)形成了改質區域12的對象物。 在該步驟S11中,如圖24所示,準備包括半導體基板21的對象物。半導體基板21包括背面(第1面)21b和背面21b的相反側的表面(第2面)21a。在半導體基板21中,設定了在沿背面21b和表面21a的X方向上延伸的線15。半導體基板21為了使背面21b為鐳射光L的射入面,以背面21b面對鐳射照射單元3的方式被載置台2支撐。在此狀態下,控制部8一邊控制鐳射照射單元3,一邊控制驅動單元7和/或載置台2的移動機構,使半導體基板21沿X方向相對移動,使鐳射光L的聚光點C沿線15相對於半導體基板21移動。 此時,控制部8顯示用於使空間光調變器32將鐳射光L分為多個(在此2個)的鐳射光L1、L2的圖案。由此,在半導體基板21的內部,以在Z方向上間隔距離Dz、並且在X方向上間隔距離Dx的方式,形成鐳射光L1、L2各自的聚光點C1、C2。其結果是,在半導體基板21中,沿線15形成多個(在此2排)的改質區域12a、12b。因此,在此,X方向為聚光點C1、C2前進的加工前進方向。 這樣,在此,控制部8執行下述的鐳射加工處理:藉由對鐳射照射單元3(照射部)的控制,沿作為線15的延伸方向的X方向對半導體基板21照射鐳射光L,在半導體基板21形成沿X方向排列的多個改質區域12和從改質區域12延伸的裂紋(裂紋14、14k)。在圖24及之後的附圖中,省略了形成在半導體基板21的表面21a的功能元件層22。 接著,進行內部觀察。即,在接下來的步驟中,使半導體基板21移動至觀察位置(步驟S12)。更具體而言,控制部8藉由控制驅動單元7和/或載置台2的移動機構,使半導體基板21相對移動至攝影單元4的物鏡43的正下方。此外,在另外準備了形成了改質區域12的半導體基板21的情況下,例如也可以由用戶將該半導體基板21載置在觀察位置。 接著,如圖25所示,利用相對於半導體基板21具有透射性的光(透射光)I1進行半導體基板21的攝影(步驟S13:攝影步驟)。在該步驟S13中,執行下述的攝影處理:藉由對攝影單元4(攝影部)的控制,一邊使光I1從半導體基板21的背面21b射入至半導體基板21的內部,一邊利用光I1對從改質區域12延伸的裂紋中的沿與Z方向和X方向交叉的方向延伸的裂紋14k即對象裂紋進行攝影。Y方向是與作為加工前進方向的X方向、及同背面21b和表面21a交叉的Z方向交叉的方向的一例。 更具體而言,在步驟S13中,控制部8藉由控制驅動單元7(移動部)和攝影單元4,使攝影單元4沿Z方向移動,使光I1的聚光點位於半導體基板21的內部的多個位置而對半導體基板21進行攝影,由此獲取多個內部圖像ID。如上前述,在本實施型態中,物鏡43與攝影單元4一體地移動。因此,使攝影單元4移動也就是使物鏡43移動,攝影單元4的移動量與物鏡43的移動量是同等的。 此時,控制部8藉由對驅動單元7的控制,使攝影單元4在Z方向上移動,一邊使光I1的聚光點(焦點F,虛擬焦點Fv)在Z方向上移動,一邊進行多次半導體基板21的攝影。使光I1的聚光點移動的範圍,可以是半導體基板21的厚度的整個範圍,但是在此,在步驟S11的鐳射加工時,能夠選取一部分的範圍RA,該一部分的範圍RA包含為了形成改質區域12a、12b而使鐳射光L1、L2的聚光點C1、C2對準的Z方向的位置。進行多次攝影時攝影單元4在Z方向上的移動間隔,即,半導體基板21的攝影間隔是任意的,但是從更準確地檢測裂紋14k的觀點來看,更精細地設定較佳。攝影間隔作為一例在1μm以內,在此為0.2μm。 進一步,在此,控制部8控制攝影單元4和驅動單元7,以執行半導體基板21的直接觀察和背面反射觀察。更具體而言,控制部8首先執行下述的第1攝影處理,使光I1從背面21b射入至半導體基板21,並使攝影單元4沿Z方向移動,由此,一邊使沒有經過表面21a的反射的光I1的聚光點(焦點F)從背面21b側向表面21a側移動,一邊在Z方向的多個位置對半導體基板21進行攝影,由此,作為內部圖像ID獲取多個第1內部圖像ID1。該第1攝影處理是直接觀察。 並且,控制部8執行下述的第2攝影處理:使光I1從背面21b射入至對象物,使攝影單元4沿Z方向移動,藉此,一邊使在表面21a反射了的光I1的聚光點(虛擬焦點Fv)從表面21a側向背面21b側移動、一邊在多個位置對半導體基板21進行攝影,由此,作為內部圖像ID獲取多個第2內部圖像ID2。該第2攝影處理由於是相對於光I1的射入面從背面(在此,在半導體基板21的結構上,稱為表面21a)側進行的觀察,因此是背面反射觀察。 在接著的步驟中,保存關於藉由步驟S13的攝影獲取的內部圖像ID的攝影資料(步驟S14)。如上所述,在步驟S13中,控制部8一邊藉由對驅動單元7的控制使攝影單元4(即光I1的聚光點)沿Z方向移動一邊進行攝影。因此,控制部8能夠獲取攝影得到各個內部圖像時的攝影單元4的移動量Fz。在此,關於各個內部圖像ID的移動量Fz的資訊,能夠與各個內部圖像ID相對應,並作為攝影資料被保存。攝影資料,不論控制部8和鐳射加工裝置1的內外,能夠被保存在控制部8能夠聯絡的任意儲存裝置中。 攝影單元4(物鏡43)的移動量,作為一例,能夠選取從使光I1的聚光點對準了半導體基板21的背面21b的狀態的位置起,以使光I1的聚光點對準半導體基板21的內部的所希望的位置的方式使攝影單元4沿Z方向移動了的情況下的攝影單元4的移動量。 接著,控制部8從規定的儲存裝置輸入攝影資料(步驟S15)。然後,控制部8判斷裂紋14k的形成狀態(步驟S16)。在此,作為一例,控制部8藉由圖像識別,自動地判斷多個內部圖像ID中裂紋14k的像相對地清晰的內部圖像ID(進行AI判斷)。AI判斷的一例如上所述。圖26是在Z方向上彼此不同的位置攝影得到的多個內部圖像ID。 在圖26中,以(d)所示的內部圖像IDd的攝影位置為中心,(c)是向背面21b側去1μm的攝影位置的內部圖像IDc,(b)是向背面21b側去3μm的攝影位置的內部圖像IDb,(a)是向背面21b側去5μm的攝影位置的內部圖像IDa,(e)是向表面21a側去1μm的攝影位置的內部圖像IDe,(f)是向表面21a側去3μm的攝影位置的內部圖像IDf,(g)是向表面21a側去5μm的攝影位置的內部圖像IDg。其中,這裡的攝影位置,是半導體基板21的內部的值。 在圖26所示的例子中,根據在內部圖像IDd中裂紋14k的像最清晰的情況,藉由控制部8判斷內部圖像IDd是相對地分數較高且該裂紋14k的像相對地清晰的內部圖像(即,判斷為在內部圖像IDd中檢測出了裂紋14k)。控制部8能夠獲取攝影得到內部圖像IDd時的移動量。因此,控制部8能夠基於攝影得到內部圖像IDd時的移動量,運算裂紋14k的裂紋位置。 這樣,控制部8執行下述的攝影處理:藉由對攝影單元4和驅動單元7的控制,使光I1從背面21b射入至半導體基板21,一邊使攝影單元4(物鏡43)沿Z方向移動一邊利用光I1來對半導體基板21進行攝影,由此獲取包含裂紋14k清晰的像的內部圖像ID即檢測圖像。 此外,控制部8執行下述的運算處理:基於多個內部圖像ID和攝影得到各個內部圖像ID時的攝影單元4的移動量Fz,運算沿與Z方向和X方向交叉的方向延伸的裂紋14k即對象裂紋在Z方向上的位置即裂紋位置。更具體而言,控制部8在運算處理中,判斷多個內部圖像ID中裂紋14k的像清晰的內部圖像ID,基於攝影得到所判斷的該內部圖像ID時的移動量Fz來運算裂紋位置。裂紋位置例如能夠藉由對移動量Fz乘以規定的校正係數而運算。校正係數藉由上述的步驟S1~7已經被匯出。 即,控制部8保存了與移動量Fz相應的多個校正係數,在運算處理中,使用與攝影得到檢測圖像時的移動量Fz對應的校正係數,運算與裂紋14k的裂紋位置對應的改質區域12的位置的測算值。 控制部8能夠對藉由直接觀察而獲取的第1內部圖像ID1、和藉由背面反射觀察而獲取的第2內部圖像ID2這兩者進行上述的裂紋14k的裂紋位置的運算。由此,控制部8能夠運算與第1內部圖像ID1對應且相對地位於背面21b側的裂紋14k的裂紋位置、和與第2內部圖像ID2對應且相對地位於表面21a側的裂紋14k的裂紋位置。 即,在此情況下,控制部8執行第1運算處理和第2運算處理,其中,在第1運算處理中,判斷多個第1內部圖像ID1中裂紋14k清晰的第1內部圖像,基於攝影得到所判斷的該第1內部圖像時的攝影單元4的移動量,來運算作為裂紋位置的第1裂紋位置Z1,在第2運算處理中,判斷多個第2內部圖像ID2中裂紋14k清晰的第2內部圖像,基於攝影得到所判斷的該第2內部圖像時的攝影單元4的移動量,來運算作為裂紋位置的第2裂紋位置Z2(關於第1裂紋位置Z1和第2裂紋位置Z2的一例,參照圖15)。相對地位於背面21b側的第1裂紋位置Z1與相對地位於表面21a側的第2裂紋位置Z2之間的距離,界定改質區域12中的形成了裂紋14k的部分(裂紋起始部)的寬度。 接著,在步驟S16中,控制部8基於所獲取的裂紋位置等來推算改質區域12的位置等。即,這裡,控制部8執行下述的推算處理:基於改質區域12的形成條件(這裡是鐳射加工的加工條件)和裂紋位置,來推算改質區域12的背面21b側的端部(缺陷上方區域上端)在Z方向上的位置、改質區域12的表面21a側的端部(缺陷下端)在Z方向上的位置、和改質區域12在Z方向上的寬度(缺陷上方區域上端與缺陷下端的間隔)中的至少一個。 這裡,控制部8基於直接觀察來運算背面21b側的裂紋14k(上方裂紋)的第1裂紋位置Z1,基於背面反射觀察來運算表面21a側的裂紋14k(下方裂紋)的第2裂紋位置Z2。因此,控制部8,作為上方裂紋的第1裂紋位置Z1和下方裂紋的第2裂紋位置Z2的間隔,能夠運算半導體基板21內部的裂紋起始部的寬度。 然後,控制部8,例如藉由對運算得到的裂紋起始部的寬度乘以關於鐳射加工的加工條件的係數,能夠運算半導體基板21的內部的改質區域12在Z方向上的寬度。這裡的係數,例如基於鐳射加工時的鐳射光L的波長、像差校正量、脈衝寬度和脈衝能量等影響改質區域12的形成的各種條件而決定。這裡的係數在一例中為3.0左右。 這樣,控制部8在推算處理中,能夠基於改質區域12的形成條件(鐳射加工的加工條件)和第1裂紋位置Z1與第2裂紋位置Z2的間隔,推算改質區域12在Z方向上的寬度。 另一方面,控制部8藉由從上方裂紋的第1裂紋位置Z1減去所假設的改質區域12的整體的寬度即假設改質區域寬度,能夠運算改質區域12的表面21a側的下端的位置。假設改質區域寬度,例如能夠基於鐳射加工時的鐳射光L的波長、像差校正量、脈衝寬度和脈衝能量等影響改質區域12的形成的各種條件而決定。假設改質區域寬度作為一例是20μm左右。 此外,控制部8藉由從下方裂紋的第2裂紋位置Z2減去所假設的缺陷區域12m的寬度即假設缺陷區域寬度,能夠運算改質區域12的表面21a的下端的位置。假設缺陷區域寬度例如能夠基於鐳射加工時的鐳射L的波長、像差校正量、脈衝寬度和脈衝能量等影響改質區域12的形成的各種條件而決定。假設缺陷區域寬度作為一例是10μm左右。 進一步,控制部8藉由對下方裂紋的第2裂紋位置Z2加上所假設的缺陷上方區域12n的寬度即假設缺陷上方區域寬度,能夠運算改質區域12的背面21b側的上端的位置。假設缺陷上方區域寬度例如能夠基於鐳射加工時的鐳射光L的波長、像差校正量、脈衝寬度和脈衝能量等影響改質區域12的形成的各種條件而決定。假設缺陷上方區域寬度,作為一例是10μm左右。 如上前述,控制部8在步驟S16中,推算並獲取關於改質區域12的位置的各種資訊。在接著的步驟中,控制部8將步驟S16的判斷結果的資訊輸出至任意的儲存裝置(步驟S17),並保存在該儲存裝置中(步驟S18)。之後,根據需要,以能夠接受使用者的輸入的狀態使顯示器150顯示各種資訊(步驟S19),完成處理。顯示器150顯示的資訊,例如是第1裂紋位置Z1、第2裂紋位置Z2、起始部寬度、改質區域12的端部的位置、和改質區域12在Z方向上的寬度等。這樣,控制部8在步驟S19中,藉由對顯示器150的控制,執行使顯示器150顯示裂紋位置的資訊的顯示處理。 如上所述,利用鐳射加工裝置1的觀察方法結束。在本實施型態中,藉由鐳射加工裝置1中的攝影單元4、驅動單元7和控制部8進行觀察方法。換言之,在鐳射加工裝置1中,由攝影單元4、驅動單元7和控制部8構成觀察裝置1A(參照圖1),其中,攝影單元4用於利用相對於對象物60和半導體基板21具有透射性的光I1來對對象物60和半導體基板21進行攝影,驅動單元7用於使攝影單元4相對於對象物60和半導體基板21移動,控制部8用於至少控制攝影單元4和驅動單元7。 圖27是表示改質區域的位置的測算值與實測值的誤差的圖表。如圖27所示,在本實施型態的觀察方法中,使用在步驟S1~S7中考慮裝置狀態、Z方向的位置(圖27的深度位置)、像差校正量而匯出的校正係數,在步驟S11~S19運算了改質區域12的測算值。由此,本實施型態中的測算值與實測值的誤差,大致處於6μm以內。而使用了一定(固定)的校正係數的比較例的誤差,與本實施型態相較之下會大到19μm左右。 如上所說明的那樣,在本實施型態的觀察裝置1A和觀察方法的對象物60,設置了與其背面60b和表面60a交叉的Z方向上的位置的實測值已知的標記(在此為改質區域12和裂紋14k)。在本實施型態的觀察裝置1A和觀察方法中,藉由一邊使攝影單元4移動一邊對這樣的對象物60進行攝影,能夠獲取作為對象物60的內部圖像GD的包含裂紋14k的像的標記圖像。而且,以使得對攝影得到該標記圖像時的移動量Fz乘以校正係數而得到的值(測算值)為已知的改質區域12的位置的實測值的方式,匯出校正係數。即,根據該觀察裝置1A和觀察方法,能夠匯出與攝影得到標記圖像時的裝置狀態、和攝影得到標記圖像時的攝影單元4的移動量(即觀察深度)相應的校正係數。因此,在利用光I1觀察改質區域12來運算改質區域12的位置的測算值時,如果使用該校正係數,就能夠更準確地獲取關於改質區域12的位置的資訊。 此外,在本實施型態的觀察裝置1A中,在對象物60形成了在Z方向上的位置彼此不同,且該位置的實測值已知的多個改質區域12和裂紋14k,在攝影處理中,控制部8藉由在Z方向上使聚光點位於對象物60的內部的多個位置而對對象物60進行攝影,獲取包含從多個改質區域12的各個改質區域12延伸的各個裂紋14k的像的多個標記圖像。然後,在匯出處理中,控制部8以使得對攝影得到多個裂紋14k各者時的攝影單元4的各個移動量Fz乘以校正係數而得到的各個值即測算值,成為多個改質區域12各自的位置的實測值的方式,匯出多個校正係數。因此,在利用光I1觀察改質區域12來運算改質區域12的位置的測算值時,能夠在Z方向上的更寬廣的範圍更準確地獲取關於改質區域12的位置的資訊。 此外,在觀察裝置1A中,在對象物60中形成了在Z方向上排列的改質區域12和從改質區域12延伸的裂紋14、14k,在攝影處理中,藉由使攝影單元4沿Z方向移動,一邊使光I1的聚光點移動一邊利用光I1來對對象物60進行攝影,作為標記圖像獲取包含裂紋14、14k中的沿與X方向和Z方向交叉的方向延伸的裂紋14k的像的內部圖像GD。 根據本發明的發明者的發現,在對象物60的內部,例如藉由鐳射加工形成改質區域12時,有時形成從該改質區域12向各種方向延伸的裂紋。而且,該裂紋中,沿與Z方向和X方向交叉的方向延伸的裂紋14k,與改質區域12相較之下,利用從對象物60透射的光I1被精準地檢測出,其中,Z方向是與作為對象物60的鐳射光射入面的背面60b交叉的方向,X方向是鐳射加工的前進方向。因此,如果如上所述的那樣使包含該裂紋14k的像的內部圖像GD為標記圖像,就能夠減輕攝影得到該標記圖像時的攝影單元4的移動量Fz的偏差。其結果是,能夠匯出更準確的校正係數。 在本實施型態的觀察裝置1A中,攝影單元4具有校正環透鏡,該校正環透鏡包括物鏡43和設置在物鏡43且用於對在對象物60產生的像差進行校正的校正環43a。這樣,在用於使光I1朝向對象物60聚光的物鏡43設置了校正環43a的情況下,在校正環43a的操作前後可能發生裝置狀態的變化。因此,如上所述的那樣匯出與裝置狀態相應的校正係數更有效。 這裡,本實施型態的觀察裝置1A包括:攝影單元4,其具有用於使相對於半導體基板21具有透射性的光I1朝向半導體基板21聚光的物鏡43,用於利用光I1來對半導體基板21進行攝影;用於使物鏡43相對於半導體基板21移動的驅動單元7;和用於至少控制攝影單元4和驅動單元7的控制部8。半導體基板21具有背面21b和背面21b的相反側的表面21a,在半導體基板21中設置了改質區域12和從改質區域12延伸的裂紋14、14k。控制部8執行攝影處理和運算處理,其中,前述攝影處理,藉由對攝影單元4和驅動單元7的控制,使光I1從背面21b射入至半導體基板21,一邊使攝影單元4沿Z方向移動一邊利用光I1來對半導體基板21進行攝影,由此獲取包含裂紋14k的像的內部圖像ID即檢測圖像,前述運算處理,在攝影處理之後,藉由對攝影得到檢測圖像時的攝影單元4的移動量Fz乘以校正係數,來運算裂紋14k在Z方向上的位置即裂紋位置。控制部8保存了與移動量Fz相應的多個校正係數。 該觀察裝置1A,如上所述,保存了與攝影單元4的移動量Fz相應的校正係數。因此,藉由使用該校正係數來運算裂紋14k的位置的測算值,能夠獲取關於更準確的改質區域12的位置的資訊。 本實施型態的對象物60,具有背面60b和背面60b的相反側的表面60a,而作為標記設置了改質區域12和從改質區域12延伸的裂紋14k,對象物60用於匯出校正係數,該校正係數用於根據改質區域12的位置的實測值來運算與背面60b和表面60a交叉的Z方向上的改質區域12和裂紋14k的位置的測算值。使用該對象物60,能夠如上所述地匯出校正係數。 以上的實施型態用於說明本發明的一個型態。因此,本發明不限定於上述實施型態,能夠被任意地變形。 例如,在上述實施型態中,作為使物鏡43相對於半導體基板21沿Z方向移動的手段,例示了使物鏡43與攝影單元4一起移動的驅動單元7。但是,例如也可以利用致動器僅使物鏡43沿Z方向移動。 此外,在上述實施型態中,說明了在步驟S5、S16中,控制部8自動地進行圖像的判斷的例子,但是控制部8也可以基於用戶的判斷結果獲取裂紋14k的裂紋位置。在此情況下,控制部8例如使顯示器150顯示多個內部圖像GD、ID,並且使顯示器150顯示催促從多個內部圖像GD、ID判斷(選擇)裂紋14k的像清晰的一個內部圖像的資訊。並且,控制部8能夠藉由顯示器150接受該判斷結果的輸入,基於與判斷結果對應的內部圖像GD、ID的移動量Fz來運算裂紋14k的裂紋位置。在此情況下,顯示器150是用於顯示資訊的顯示部,並且也是接受輸入的輸入接受部。在此情況下,控制部8的用於進行圖像識別等的處理負荷被減輕。 此外,在上述實施型態中,在步驟S2、S13中對1個改質區域12的觀察進行直接觀察和背面反射觀察這兩者,獲取了作為內部圖像GD、ID的第1內部圖像GD1、ID1和第2內部圖像GD2、ID2。但是,在步驟S2、S13中,也可以僅進行直接觀察和背面反射觀察中的一者。在此情況下,因為能夠獲得第1內部圖像GD1、ID1和第2內部圖像GD2、ID2中的一者,所以也可以基於該一者,進行校正係數的匯出、改質區域12的端部的位置、寬度的推算。 這裡,在觀察裝置1A中,也可以始終設置用於匯出校正係數的對象物60。即,觀察裝置1A也可以包括:用於設置對象物60的設置部(例如圖1的載置台2A)和設置在設置部的對象物60。這樣,藉由始終設置作為標記設置了改質區域12和裂紋14k的對象物60,能夠在任意的時間進行校正係數的匯出。 進一步,在上述實施型態中,在匯出校正係數時,將包含裂紋14k的內部圖像作為標記圖像,利用了攝影得到該標記圖像時的移動量Fz。即,藉由檢測裂紋14k,基於檢測出該裂紋14k的位置進行了校正係數的匯出。但是,在觀察裝置1A和觀察方法中,在匯出校正係數時,也可以構成為檢測改質區域12本身。在此情況下,控制部8的處理如下所述。 即,控制部8執行攝影處理和匯出處理,在攝影處理中,藉由對攝影單元4和驅動單元7的控制,使光I1從背面60b射入至對象物60的內部,一邊使攝影單元4沿Z方向移動而使光I1的聚光點沿Z方向移動,一邊利用光I1來對對象物60進行攝影,藉此獲取作為對象物60的內部圖像GD的包含改質區域12的像的標記圖像,在匯出處理中,在攝影處理之後,以使得對攝影得到標記圖像時的攝影單元4的移動量Fz乘以校正係數而得到的值即測算值為改質區域12的位置的實測值的方式,匯出校正係數。 這樣在使改質區域12本身為檢測對象的情況下,為了在Z方向上的更多的位置匯出校正係數,在Z方向上要以更小的間隔(例如移動量Fz為5μm,對象物60內部為20μm左右)對內部圖像進行攝影時,存在1個改質區域12的像遍及多個內部圖像地被攝影的問題。在此情況下,Z方向上的某位置的改質區域12的像與Z方向的另一位置的改質區域12的像在1個內部圖像中重複。 為了解決這樣的問題,在令改質區域12為檢測對象的情況下,能夠考慮如圖28所示的那樣,使沿Z方向排列的多排改質區域12各自沿Y方向彼此偏移。這樣,能夠避免Z方向上的某位置的改質區域12的像與Z方向上的另一位置的改質區域12的像在1個內部圖像中重複,因此能夠以更小的間隔對內部圖像進行攝影,能夠匯出更多位置的校正係數。 此外,在用於匯出校正係數的檢測對象為改質區域12的情況下,在之後的用於獲取關於半導體基板21的改質區域12的位置的資訊的步驟中,也可以代替裂紋14k的檢測而進行改質區域12本身的檢測。 這裡,在上述實施型態中,在匯出校正係數時,列舉了使用對象物60的例子,在該對象物60設置了實測值已知的改質區域12和裂紋14k作為標記。但是,用於匯出校正係數的對象物不限定於此。例如,可以將在厚度已知的晶圓的1個面,作為標記黏貼了測試圖標的東西作為對象物,或者也可以將在厚度已知的晶圓的1個面作為標記形成了規定圖案而得到東西作為對象物。在這些情況下,藉由準備厚度不同的多個對象物,能夠匯出Z方向的多個位置的校正係數。 Hereinafter, an embodiment will be described in detail with reference to the drawings. However, in the description of each drawing, the same reference numerals may be assigned to the same or corresponding parts, and overlapping descriptions may be omitted. In addition, in each drawing, a rectangular coordinate system defined by an X axis, a Y axis, and a Z axis may be shown. As an example, the X direction and the Y direction are a first horizontal direction and a second horizontal direction intersecting (orthogonal) with each other, and the Z direction is a vertical direction intersecting (orthogonal) the X direction and the Y direction. As shown in FIG. 1 , the laser processing apparatus 1 includes a stage 2, a laser irradiation unit 3 (irradiation unit), a plurality of imaging units 4, 5, 6, a drive unit 7, a control unit 8, and a display 150 (display unit). The laser processing apparatus 1 is an apparatus for forming a modified region 12 on an object 11 by irradiating the object 11 with laser light L. The stage 2 supports the object 11 by, for example, a film adhered to the object 11 by suction. The mounting table 2 is movable along the X direction and the Y direction, and is rotatable about an axis parallel to the Z direction as a center line. The laser irradiation unit 3 condenses the laser light L having transmittance to the object 11 to irradiate the object 11 . When the laser light L is condensed to the inside of the object 11 supported by the stage 2, the laser light L is particularly absorbed at the part corresponding to the condensing point C of the laser light L, and a laser light L can be formed inside the object 11. Modified area 12. The modified region 12 is a region that differs in density, refractive index, mechanical strength, or other physical properties from the surrounding non-modified region. As the modified region 12, there are, for example, a melt-processed region, a crack region, a dielectric breakdown region, a refractive index change region, and the like. The modified region 12 has a characteristic that cracks easily extend from the modified region 12 to the incident side of the laser light L and the opposite side. Such properties of the modified region 12 are utilized for cutting the object 11 . As an example, when the stage 2 is moved in the X direction and the focused point C is relatively moved in the X direction with respect to the object 11, a plurality of modified spots are formed in a row along the X direction. 12s. One modified spot 12s is formed by irradiating one pulse of laser light L. One row of modified regions 12 is a collection of a plurality of modified spots 12s arranged in one row. Adjacent modified spots 12s may be connected to each other or may be separated from each other depending on the relative movement speed of the focused spot C with respect to the object 11 and the repetition frequency of the laser light L. The imaging unit 4 photographs the modified region 12 formed in the object 11 and the tip of a crack extending from the modified region 12 . The imaging unit 5 and the imaging unit 6 image the object 11 supported by the mounting table 2 using the light transmitted from the object 11 under the control of the control unit 8 . Images captured by the imaging units 5 and 6 are used, for example, to align the irradiation position of the laser light L. FIG. The driving unit 7 supports the laser irradiation unit 3 and the plurality of imaging units 4 , 5 , 6 . The driving unit 7 moves the laser irradiation unit 3 and the plurality of imaging units 4, 5, 6 in the Z direction. The control unit 8 controls the operations of the mounting table 2 , the laser irradiation unit 3 , the plurality of imaging units 4 , 5 , and 6 , and the driving unit 7 . The control unit 8 is configured as a computer device including a processor, a memory, a register, a communication device, and the like. In the control unit 8 , the processor executes software (program) read from the memory, etc., and controls reading or writing of data in the memory and temporary register, and communication by the communication device. The display 150 has a function as an input unit that accepts input of information from a user, and a function as a display unit that displays information to the user. [Structure of Object] The object 11 of this embodiment is a wafer 20 as shown in FIGS. 2 and 3 . The wafer 20 includes a semiconductor substrate 21 and a functional element layer 22 . In this embodiment, it is described that the wafer 20 includes the functional element layer 22 , but the wafer 20 may have the functional element layer 22 or may not have the functional element layer 22 , or may be a bare wafer. The semiconductor substrate 21 has a front surface 21a (second surface) and a back surface 21b (first surface). The semiconductor substrate 21 is, for example, a silicon substrate. The functional element layer 22 is formed on the surface 21 a of the semiconductor substrate 21 . The functional element layer 22 includes a plurality of functional elements 22a arranged two-dimensionally along the surface 21a. The functional element 22a is, for example, a light receiving element such as a photodiode, a light emitting element such as a laser diode, or a circuit element such as a memory. The functional element 22a may also be formed three-dimensionally by stacking a plurality of layers. In addition, although the notch 21c showing the crystal orientation is provided in the semiconductor substrate 21, an orientation flat may be provided instead of the notch 21c. The wafer 20 is cut along the plurality of lines 15 for each functional element 22a. The plurality of lines 15 pass between each of the plurality of functional elements 22 a when viewed from the thickness direction of the wafer 20 . More specifically, the line 15 passes through the center of the ruled line region 23 (the center in the width direction) when viewed from the thickness direction of the wafer 20 . The ruled line region 23 extends in the functional element layer 22 so as to pass between adjacent functional elements 22a. In this embodiment, a plurality of functional elements 22a are arranged in a matrix along the surface 21a, and a plurality of lines 15 are set in a grid. In addition, although the line 15 is an imaginary line, it may be an actually drawn line. [Structure of Laser Irradiating Unit] As shown in FIG. 4 , the laser irradiating unit 3 has a light source 31 , a spatial light modulator 32 and a condenser lens 33 . The light source 31 outputs laser light L by, for example, pulse oscillation. The spatial light modulator 32 modulates the laser light L output from the light source 31 . The spatial light modulator 32 is, for example, a reflective liquid crystal (LCOS: Liquid Crystal on Silicon) spatial light modulator (SLM: Spatial Light Modulator). The condensing lens 33 condenses the laser light L modulated by the spatial light modulator 32 . Wherein, the condenser lens 33 may also be a correction ring lens. In the present embodiment, the laser irradiation unit 3 irradiates the laser light L to the wafer 20 from the back surface 21b side of the semiconductor substrate 21 along the plurality of lines 15, thereby irradiating the laser light L inside the semiconductor substrate 21 along the plurality of lines 15 respectively. Two rows of modified regions 12a, 12b are formed. The modified region 12a is the modified region closest to the surface 21a among the two rows of modified regions 12a and 12b. The modified region 12b is the modified region closest to the modified region 12a among the two rows of modified regions 12a and 12b, and is the modified region closest to the back surface 21b. The two rows of modified regions 12 a and 12 b are adjacent to each other in the thickness direction (Z direction) of the wafer 20 . The two rows of modified regions 12 a and 12 b are formed by relatively moving the two light-converging points C1 and C2 along the line 15 with respect to the semiconductor substrate 21 . The laser light L is modulated by the spatial light modulator 32 , so that, for example, the focal point C2 is located on the rear side of the traveling direction and on the incident side of the laser light L relative to the focal point C1 . Here, regarding the formation of the modified region, it may be single focus or multi focus, and may be one path or multiple paths. The laser irradiation unit 3 irradiates the wafer 20 with laser light L along each of the plurality of lines 15 from the back surface 21 b side of the semiconductor substrate 21 . As an example, with respect to the semiconductor substrate 21 which is a single-crystal silicon <100> substrate with a thickness of 400 μm, two light-converging points C1 and C2 are respectively focused on a position 54 μm away from the surface 21 a and a position 128 μm away from the surface 21 a. Each line 15 of 15 irradiates laser light L to the wafer 20 from the back surface 21 b side of the semiconductor substrate 21 . At this time, for example, under the condition that the cracks 14 throughout the two rows of modified regions 12a, 12b reach the surface 21a of the semiconductor substrate 21, the wavelength of the laser light L is 1099nm, the pulse width is 700n seconds, and the repetition frequency is 120kHz. . In addition, let the output of the laser light L at the converging point C1 be 2.7 W, and the output of the laser light L at the converging point C2 be 2.7 W, and let the relative moving speeds of the two converging points C1 and C2 with respect to the semiconductor substrate 21 be 800mm/sec. Wherein, for example, when the number of processing passes is 5, for the above-mentioned wafer 20, ZH80 (position 328 μm away from surface 21a), ZH69 (position 283 μm away from surface 21a), ZH57 (position 283 μm away from surface 21a) and ZH57 (position 328 μm away from surface 21a) can also be used for the above-mentioned wafer 20, for example. 21a at a distance of 234 μm), ZH26 (a position at a distance of 107 μm from the surface 21a), and ZH12 (a position at a distance of 49.2 μm from the surface 21a) are processing positions. In this case, for example, the laser light L may have a wavelength of 1080 nm, a pulse width of 400 nsec, a repetition frequency of 100 kHz, and a moving speed of 490 mm/sec. The formation of such two rows of modified regions 12a, 12b and cracks 14 is carried out as follows. That is, in subsequent steps, for example, by grinding the back surface 21b of the semiconductor substrate 21, the semiconductor substrate 21 is thinned, and the cracks 14 are exposed to the back surface 21b, and the wafer 20 is cut into multiple pieces along the plurality of lines 15, respectively. case of a semiconductor element. [Structure of Imaging Unit for Inspection] As shown in FIG. The photographing unit 4 photographs the wafer 20 . The light source 41 outputs light I1 that is transparent to the semiconductor substrate 21 . The light source 41 is composed of, for example, a halogen lamp and a filter, and outputs light I1 in the near-infrared region. The light I1 output from the light source 41 is reflected by the mirror 42 , passes through the objective lens 43 , and is irradiated to the wafer 20 from the rear surface 21 b side of the semiconductor substrate 21 . At this time, the stage 2 supports the wafer 20 on which the two rows of modified regions 12 a and 12 b are formed as described above. The objective lens 43 is used to condense light (transmitted light) I1 that is transparent to the semiconductor substrate 21 toward the semiconductor substrate 21 . The objective lens 43 passes the light I1 reflected by the surface 21 a of the semiconductor substrate 21 . That is, the objective lens 43 passes the light I1 having propagated through the semiconductor substrate 21 . The numerical aperture (NA) of the objective lens 43 is 0.45 or more, for example. The objective lens 43 has a correction ring 43a. The correction ring 43 a corrects the aberration generated by the light I1 in the semiconductor substrate 21 by, for example, adjusting the distance between a plurality of lenses constituting the objective lens 43 . Here, the means for correcting aberrations is not limited to the correction ring 43a, and other correction means such as a spatial light modulator may also be used. The light detection unit 44 detects the light I1 transmitted through the objective lens 43 and the reflection mirror 42 . The photodetector 44 is composed of, for example, an InGaAs camera, and detects the light I1 in the near-infrared region. However, the means for detecting (imaging) the light I1 in the near-infrared region is not limited to an InGaAs camera, and other imaging means that perform transmission-type imaging such as a transmission confocal microscope may be used. The imaging unit 4 is capable of imaging each of the two rows of modified regions 12a, 12b and the front ends of each of the plurality of cracks 14a, 14b, 14c, and 14d (details will be described later). The crack 14a is a crack extending from the modified region 12a toward the surface 21a. The crack 14b is a crack extending from the modified region 12a to the rear surface 21b side. The crack 14c is a crack extending from the modified region 12b to the surface 21a side. The crack 14d is a crack extending from the modified region 12b to the rear surface 21b side. [Structure of Imaging Unit for Alignment Correction] As shown in FIG. 6 , the imaging unit 5 has a light source 51 , a mirror 52 , a lens 53 , and a photodetector 54 . The light source 51 outputs light I2 that is transparent to the semiconductor substrate 21 . The light source 51 is composed of, for example, a halogen lamp and a filter, and outputs light I2 in the near-infrared region. The light source 51 can also be shared with the light source 41 of the imaging unit 4 . The light I2 output from the light source 51 is reflected by the mirror 52 , passes through the lens 53 , and is irradiated onto the wafer 20 from the rear surface 21 b side of the semiconductor substrate 21 . The lens 53 passes the light I2 reflected by the surface 21 a of the semiconductor substrate 21 . That is, the lens 53 passes the light I2 that has propagated through the semiconductor substrate 21 . The numerical aperture of the lens 53 is 0.3 or less. That is, the numerical aperture of the objective lens 43 of the imaging unit 4 is larger than the numerical aperture of the lens 53 . The light detection unit 54 detects the light I2 passing through the lens 53 and the reflection mirror 52 . The photodetector 54 is composed of, for example, an InGaAs camera, and detects the light I2 in the near-infrared region. Under the control of the control unit 8 , the imaging unit 5 irradiates the wafer 20 with light I2 from the back surface 21b side, and detects the light I2 returned from the surface 21a (functional element layer 22 ), thereby imaging the functional element layer 22 . Also, under the control of the control unit 8, the imaging unit 5 irradiates the wafer 20 with light I2 from the back surface 21b side, and detects the light I2 returning from the positions where the modified regions 12a, 12b are formed on the semiconductor substrate 21, whereby An image of a region including the modified regions 12a, 12b is acquired. These images are used to calibrate the irradiation position of the laser beam L. FIG. The photographing unit 6 has the same structure as the photographing unit 5 except that it has a lower magnification than the lens 53 (for example, 6 times in the photographing unit 5 and 1.5 times in the photographing unit 6), and has the same structure as the photographing unit 5. 5 is also used for calibration. [Principle of imaging by imaging unit for inspection] Using imaging unit 4 shown in FIG. 5, as shown in FIG. (The focal point of the objective lens 43) moves from the back surface 21b side to the surface 21a side. In this case, when the focal point F is focused on the front end 14e of the crack 14 extending from the modified region 12b to the back side 21b side from the rear surface 21b side, the front end 14e can be confirmed (right image in FIG. 7 ). However, even if the focal point F is focused on the crack 14 itself and the tip 14e of the crack 14 that has reached the surface 21a from the back surface 21b side, it cannot be confirmed (the image on the left side of FIG. 7 ). In addition, when the focal point F is focused on the front surface 21 a of the semiconductor substrate 21 from the rear surface 21 b side, the functional element layer 22 can be confirmed. And, using the imaging unit 4 shown in FIG. 5, as shown in FIG. 8, for the semiconductor substrate 21 in which the crack 14 spanning the two rows of modified regions 12a, 12b has not reached the surface 21a, the focal point F is directed from the back surface 21b side to the surface 21a. Move sideways. In this case, even if the focal point F is focused on the front end 14e of the crack 14 extending from the modified region 12a to the front surface 21a side from the rear surface 21b side, the front end 14e cannot be confirmed (the left image in FIG. 8 ). However, if the focal point F is focused from the rear surface 21b side to a region located on the opposite side to the rear surface 21b with respect to the surface 21a (that is, a region located on the functional element layer 22 side with respect to the surface 21a), the focal point F is symmetrical with respect to the surface 21a. When the virtual focal point Fv is positioned at the front end 14e, the front end 14e can be confirmed (the image on the right side of FIG. 8 ). In addition, the virtual focal point Fv is a point symmetrical to the focal point F with respect to the surface 21 a in consideration of the refractive index of the semiconductor substrate 21 . It is presumed that the crack 14 cannot be confirmed as described above because the width of the crack 14 is smaller than the wavelength of the light I1 which is the illumination light. 9 and 10 are SEM (Scanning Electron Microscope) images of modified regions 12 and cracks 14 formed inside semiconductor substrate 21 which is a silicon substrate. (b) of FIG. 9 is an enlarged image of the region A1 shown in (a) of FIG. 9 , (a) of FIG. 10 is an enlarged image of the region A2 shown in (b) of FIG. 9 , and ( b) is an enlarged image of the area A3 shown in (a) of FIG. 10 . Thus, the width of the crack 14 is about 120 nm, which is smaller than the wavelength (for example, 1.1 to 1.2 μm) of the light I1 in the near-infrared region. The principle of photography conceived from the above matters is as follows. As shown in (a) of FIG. 11 , since the light I1 does not return when the focal point F is placed in the air, a pitch-black image (image on the right side of (a) of FIG. 11 ) is obtained. As shown in (b) of FIG. 11 , if the focal point F is located inside the semiconductor substrate 21, the light I1 reflected by the surface 21a will return, so a clean image (the figure on the right side of (b) of FIG. 11 ) will be obtained. picture). As shown in (c) of FIG. 11 , if the focal point F is focused on the modified region 12 from the rear surface 21b side, part of the light I1 returned after being reflected by the surface 21a is absorbed, scattered, etc. by the modified region 12, Therefore, an image (image on the right side of (c) of FIG. 11 ) showing the jet-black modified region 12 against a clean background is obtained. As shown in (a) and (b) of FIG. 12, if the focal point F is focused on the front end 14e of the crack 14 from the back surface 21b side, for example, due to optical specificity (stress concentration, distortion, atomic crack) generated near the front end 14e, Density discontinuity, etc.) makes the light localized near the front end 14e, etc., thereby causing a part of the light I1 reflected by the surface 21a to be scattered, reflected, interfered, absorbed, etc., so it will be displayed in a white background An image of the jet-black front end 14e is obtained (images on the right side of (a) and (b) of FIG. 12 ). As shown in (c) of FIG. 12 , if the focal point F is focused from the back surface 21b side to a portion other than the vicinity of the front end 14e of the crack 14, at least a part of the light I1 reflected by the surface 21a will return, so that a clean white image will be obtained. image (the image on the right side of (c) in FIG. 12 ). [Example of Internal Observation] FIG. 13 is a view showing an object on which a modified region is formed. (a) of FIG. 13 is a cross-sectional photograph of an object cut so as to expose a modified region. (b) of FIG. 13 is an example of an image of an object captured by light transmitted through the object. (c) of FIG. 13 is another example of an image of an object captured by light transmitted through the object. As shown in (a) of FIG. 13 , the modified region 12 formed on the object (here, the semiconductor substrate 21 ) by condensing the laser light L includes: Void region 12m on the side opposite to surface 21a; and Void region 12n on the back surface 21b side, which is a surface where laser light L is incident from defect region 12m. When the semiconductor substrate 21 on which such a modified region 12 is formed is photographed by light I1 which is transparent to the semiconductor substrate 21, as shown in (b) and (c) of FIG. An image of a crack 14k extending in a direction intersecting the Z direction and the X direction (having an angle with respect to the X direction). When viewed from the Z direction, the crack 14k is approximately parallel to the Y direction in the example of FIG. 13( b ), and is slightly inclined relative to the Y direction in the example of FIG. 13( c ). When the image of these cracks 14k is photographed at a plurality of positions on the semiconductor substrate 21 while moving the converging point of the light I1 in the Z direction, the range in the Z direction can be limited compared with that of the modified region 12 was clearly detected. FIG. 14 is a graph regarding the location of modified regions and cracks in the Z direction. In FIG. 14 , the plots of the lower end of the defect, the upper end of the defect, the lower end of the area above the defect, and the upper end of the area above the defect are actual measured values actually measured by cross-sectional observation. The lower end means the end on the front surface 21a side, and the upper end means the end on the back surface 21b side. Therefore, for example, the lower end of the region above the defect refers to the end of the region 12n above the defect on the surface 21a side. In addition, the drawing of the direct observation and the back reflection observation in the graph of FIG. 14 is based on the movement of the objective lens 43 in the Z direction when the internal image including the clear image of the crack 14k in the image obtained by using the light I1 is photographed. The measured value obtained by calculation of the amount (hereinafter sometimes simply referred to as "movement amount") is, for example, a value obtained by image judgment based on AI. Direct observation is the case where light I1 is incident from the back surface 21b, and the converging point of light I1 is directly aligned with the crack 14k without being reflected by the surface 21a (in the above example, the focal point F is aligned with the crack 14k from the back surface 21b side). 14k), the back reflection observation is the case where light I1 is incident from the back surface 21b, and the converging point of the light I1 reflected by the surface 21a is aligned with the crack 14k (in the above example, with respect to the surface 21a, When the focal point F is aligned with the area on the opposite side of the rear surface 21b from the rear surface 21b side, and the virtual focal point Fv symmetrical to the focal point F with respect to the surface 21a is aligned with the crack 14k). As shown in FIG. 14 , in direct observation, in the cases C1 to C4 where the formation positions of the modified regions 12 are located at four different positions in the Z direction, all detection is between the lower end of the region above the defect and the upper end of the region above the defect. A crack 14k is detected. In the back reflection observation, the crack 14k is detected at approximately the same position as the lower end of the area above the defect in case C1, and the crack 14k is detected between the lower end of the area above the defect and the upper end of the defect in cases C2~C4. The width of the modified region 12 in the Z direction is the distance between the lower end of the defect and the upper end of the region above the defect. In this way, the crack 14 k can be detected more accurately in the Z direction than the modified region 12 itself. Therefore, by acquiring the movement amount of the internal image when the crack 14k appears in the Z direction, information on the position of the modified region 12 can be acquired more accurately. Here, the vertical axis in FIG. 14 represents the distance from the back surface, where the back surface is the back surface with respect to the incident surface of the light I1 , and is the surface 21 a for the semiconductor substrate 21 . In addition, Fig. 15 is obtained by plotting the detection results in the case C1 on the cross-sectional photograph. In this embodiment, based on the above findings, the crack 14k is detected by internal observation, and the information on the position of the modified region 12 is acquired. Next, the observation method of this embodiment will be described. In this observation method, the crack 14k is a crack to be detected. In this embodiment, when obtaining information on the position of the modified region 12, first, the position of the crack 14k in the Z direction is calculated. At this time, the movement amount of the objective lens 43 in the Z direction when the internal image of the detected crack 14k is captured is multiplied by a predetermined correction coefficient. As shown in FIGS. 14 and 15 , cracks 14 k are detected within the range of the modified region 12 in the Z direction. Therefore, the calculated position of the crack 14 k is the calculated value of the position of the modified region 12 in the Z direction. First, the discovery about the correction coefficient will be described. As shown in FIG. 16 , in order to adjust the position of the converging point of the light I1 in the semiconductor substrate 21 , the imaging unit 4 is moved by a movement amount Fz in the Z direction using the drive unit 7 . At this time, if there is no semiconductor substrate 21, the moving amount of the converging point of the light I1 is also the moving amount Fz. However, when the converging point of the light I1 is formed inside the semiconductor substrate 21 , the moving amount of the converging point of the light I1 is a moving amount Hz different from the moving amount Fz (magnified in the illustrated example). The amount of movement Hz defines the actual imaging position in the semiconductor substrate 21 , that is, the position of the detection target (for example, the modified region 12 and the crack 14 k ). On the other hand, the information that can be directly acquired by the control unit 8 is the movement amount Fz of the imaging unit 4 (that is, the movement amount Fz of the focal point when there is no semiconductor substrate 21 ) as an input value for the control of the drive unit 7 . ). Therefore, in order to obtain the actual position of the detection target in the semiconductor substrate 21 , the control unit 8 needs to multiply the movement amount Fz by a certain coefficient. The coefficient applied at this time is the correction coefficient. This correction coefficient can be set to a constant value (for example, about 4 when the semiconductor substrate 21 is silicon) in consideration of the NA of the objective lens 43 and the refractive index of the semiconductor substrate 21 . However, when the correction coefficient is set to a constant value, the following problems may arise. 17 is a graph showing the relationship between the position of the detection target in the Z direction and the movement amount when the detection target is detected. The "depth position" on the horizontal axis of the graph in FIG. 17 is the position in the Z direction where the detection object is set (the position where the light I1 is aligned), and the "Z-axis movement amount" on the vertical axis of the graph in FIG. 17 is " is the movement amount Fz when the detection object is detected at each position in the Z direction. The detection objects are arranged at intervals of about 40 μm in the semiconductor substrate 21 so that the detection objects can be detected when the movement amount Fz is 10 μm. In FIG. 17 , graphs in a plurality of cases with different device states are collectively described. In the case of focusing on one device state (one graph), when the position of the detection object in the Z direction is different, the original The amount of movement Fz that should be fixed at 10 μm deviates. The same applies when comparing a plurality of device states (a plurality of graphs). The cause of the shift in the amount of movement Fz in this way is, first, the shift in the observation position due to the focus blur of the objective lens 43 . That is, when the amount of spherical aberration correction of the objective lens 43 of the imaging unit 4 is constant, the constant spherical aberration correction is weak correction relative to the ideal state. In this case, the condensing position of the light I1 in the semiconductor substrate 21 becomes relatively shallow, and as a result, the movement amount Fz of the imaging unit 4 when a certain modified region 12 is detected becomes relatively large (the observation position becomes Deeper). Similarly, when the spherical aberration correction amount of the objective lens 43 of the imaging unit 4 is constant, and the constant spherical aberration correction is overcorrected compared to the ideal state, the condensing position of the light I1 in the semiconductor substrate 21 is relatively changed. As a result, the movement amount Fz of the imaging unit 4 when a certain modified region 12 is detected becomes relatively small (the observation position becomes shallower). In addition, as a cause of variation in the amount of movement Fz, it is conceivable to consider a deviation before and after the operation of the correction ring lens. That is, in the case where the objective lens 43 of the photographing unit 4 is a correction ring lens, even if the correction ring 43a is operated in order to adjust the aberration correction amount by the correction ring 43a, the operation amount of the correction ring 43a is equal to the change amount of the aberration correction amount. It may not always be the case, and as a result, the observation position may be shifted before and after the operation of the correction ring 43a. Furthermore, mechanical error of the objective lens 43 of the imaging unit 4, attachment and detachment of the objective lens 43, and the like are also causes of variation in the movement amount Fz. When calculating the measured value by multiplying the shifted amount of movement Fz by a certain correction coefficient in this way, the calculation result also deviates. Accordingly, in order to obtain information about the exact position of the detection object, it is necessary to use an appropriate correction coefficient according to the state of the device and the position in the Z direction. Therefore, in the observation method of the present embodiment, exporting of the correction coefficients is performed as follows. The timing of exporting the correction coefficient is arbitrary, and as an example, it is a timing when the state of the device changes such as when the objective lens 43 is attached or detached. Fig. 18 is a flowchart showing the procedure for exporting the correction coefficients in the observation method of this embodiment. As shown in FIG. 18 , the object (observation object) 60 for exporting the correction coefficient is moved to the lower part of the objective lens 43 of the imaging unit 4 (step S1 ). As shown in FIG. 1 , the laser processing apparatus 1 further includes a mounting table (installation portion) 2A different from the mounting table 2 , and an object 60 is mounted on the mounting table 2A. The mounting table 2A is movable in the X direction and the Y direction by the drive unit 7, for example. Fig. 19 is a side view showing an object for exporting correction coefficients. As shown in FIG. 19 , the object 60 includes a rear surface 60 b and a surface (second surface) 60 a opposite to the rear surface (first surface) 60 b. In object 60 , modified regions 12 aligned in the X direction along back surface 60 b and surface 60 a and cracks (cracks 14 , 14 k ) extending from modified regions 12 are formed by laser processing. In particular, cracks 14 k extending from the modified region 12 in a direction intersecting the Z direction and the X direction are formed in the object 60 . In addition, in the object 60, a plurality of rows of modified regions 12 are formed so as to be aligned in the Z direction. The distance between the modified regions 12 in the Z direction is equal to or less than 10 μm in terms of the amount of movement Fz. In the object 60, a cut surface is formed so that the modified region 12 is exposed, and the position of each modified region 12 in the Z direction, for example, the position of the crack 14k, is actually measured by observing the cut surface. rather known. The known measured values may be stored by the control unit 8 or may be stored in any storage device that the control unit 8 can communicate with. In this way, here, the preparation step of preparing the object 60 including the back surface 60b and the surface 60a on the opposite side of the back surface 60b is carried out, and the actual measurement value of the position in the Z direction intersecting the back surface 60b and the surface 60a is known. , the modified region 12 and the crack 14k are formed in the object 60 . In the next step, as shown in FIG. 20 , the object 60 is photographed using light (transmitted light) I1 having a transmittance to the object 60 (step S2: photographing step). In this step S2, by controlling the imaging unit 4 (imaging unit), the following imaging processing is performed: using the light I1 to scan cracks extending from the modified region 12 in the direction intersecting the Z direction and the X direction. Extended crack 14k for photography. The Y direction is an example of a direction intersecting the X direction and the Z direction. The X direction is the processing advance direction of laser processing for forming the modified regions 12 on the object 60 (that is, the direction in which the modified regions 12 are arranged), and the Z direction is the direction in which the modified regions 12 are arranged. A direction intersecting the back surface 60b and the surface 60a. In this step S2, the control unit 8 controls the imaging unit 4 and the driving unit 7 to cause the light I1 to enter the object 60 from the back surface 60b of the object 60, and moves the imaging unit 4 in the Z direction to make the light I1 The focal point (focus F, virtual focal point Fv) of the light I1 is moved in the Z direction so that the focal point of the light I1 is positioned at a plurality of positions inside the object 60 and the object 60 is photographed multiple times. Thereby, a plurality of internal images GD are acquired. In this embodiment, the objective lens 43 moves integrally with the imaging unit 4 . Therefore, to move the imaging unit 4 is to move the objective lens 43 , and the moving amount of the imaging unit 4 is equal to the moving amount of the objective lens 43 . The range in which the converging point of the light I1 is moved may be the entire range of the thickness of the object 60, but here, a part of the range RA can be selected. One example is the position in the Z direction where the laser beam focus point is aligned by modifying the regions 12a, 12b). The moving interval of the imaging unit 4 in the Z direction when performing multiple imaging, that is, the imaging interval of the object 60 is arbitrary, and it is preferable to set it more finely from the viewpoint of detecting the crack 14k more accurately. The imaging interval is within 1 μm as an example, and here is 0.2 μm. Further, here, the control unit 8 controls the imaging unit 4 and the driving unit 7 so as to perform direct observation and rear reflection observation of the object 60 . More specifically, the control unit 8 firstly executes the first photographing process of causing the light I1 to enter the object 60 from the back surface 60b, and moving the photographing unit 4 in the Z direction, while moving the light I1 that does not pass through the surface 60a. The converging point (focal point F) of the reflected light I1 moves from the back surface 60b side to the front surface 60a side, and the object 60 is photographed at a plurality of positions in the Z direction, and a plurality of first internal images are acquired as the internal image GD. GD1. This first photographing process is direct observation. In addition, the control unit 8 executes the second photographing process of causing the light I1 to enter the object from the back surface 60b and moving the photographing unit 4 in the Z direction, thereby reducing the intensity of the light I1 reflected by the surface 60a. The focus point (virtual focal point Fv) is moved from the front 60a side to the back 60b side, and the object 60 is photographed at a plurality of positions, thereby acquiring a plurality of second internal images GD2 as the internal image GD. This second photographing process is observation from the back side of the incident surface of the light I1 (here, referred to as the front surface 60 a in relation to the front and back of the semiconductor substrate 21 ) side, and thus is back reflection observation. In the next step, photographing data on the internal image GD acquired by photographing in step S2 is saved (step S3). As described above, in step S2 , the control unit 8 performs imaging while moving the imaging unit 4 (that is, the focal point of the light I1 ) in the Z direction by controlling the driving unit 7 . Therefore, the control unit 8 can acquire the movement amount Fz when each internal image is captured. Here, the information on the amount of movement Fz of each internal image GD can be associated with each internal image GD and stored as photographic data. The photographic data can be stored in any storage device that can be communicated with the control unit 8 regardless of inside or outside the control unit 8 and the laser processing apparatus 1 . The amount of movement Fz of the imaging unit 4 (objective lens 43) is, as an example, from the position where the light I1 is aligned with the back surface 21b of the semiconductor substrate 21 so that the light I1 is aligned with the semiconductor substrate. The amount of movement of the imaging unit 4 when the imaging unit 4 is moved in the Z direction according to a desired position inside 21 . Next, the control unit 8 inputs photographic data from a predetermined storage device (step S4). Then, the control unit 8 judges the formation state of the crack 14k (step S5). Here, as an example, the control unit 8 automatically judges an internal image GD (marker image) in which the image of the crack 14k is relatively clear among a plurality of internal images GD by image recognition (performs AI judgment). Here, an example of an algorithm for detecting cracks and modified regions by AI judgment will be described. 29 and 30 are diagrams illustrating crack detection. FIG. 29 illustrates internal observation results (internal images of the semiconductor substrate 21 ). The control unit 8 first detects the straight line group 140 with respect to the internal image of the semiconductor substrate 21 shown in (a) of FIG. 29 . For the detection of the straight line group 140 , for example, algorithms such as Hough transform and LSD (Line Segment Detector: straight line segment detection algorithm) can be used. The Hough transform is a method of detecting all straight lines passing through all points on an image, and assigning weights to straight lines passing more feature points to detect straight lines. LSD is a method of detecting a straight line by estimating an area to be a line segment by calculating the slope and angle of luminance values in an image, and approximating the area to a rectangle. Next, the control unit 8 detects the crack 14 from the straight line group 140 by calculating the similarity with the crack line for the straight line group 140 as shown in FIG. 30 . The crack line, as shown in the upper graph of Fig. 30, has the characteristic of being very bright front and back in the Y direction relative to the brightness value on the line. Therefore, the control unit 8 compares, for example, the brightness values of all the pixels in the detected straight line group 140 with those before and after in the Y direction, and uses the number of pixels whose difference is equal to or greater than the threshold value as a similarity score. Then, among the plurality of detected straight line groups 140 , the one with the highest similarity score to the crack line is taken as the representative value of the image. The higher the representative value, the higher the probability that the crack 14 exists is an index. The control unit 8 compares representative values of a plurality of images, and selects an image with a relatively high score as a crack image candidate. 31 to 33 are diagrams illustrating flaw detection. FIG. 31 illustrates internal observation results (internal images of the semiconductor substrate 21 ). For the image inside the semiconductor substrate 21 shown in (a) of FIG. 31 , the control unit 8 detects corners (gathering of edges) in the image as key points, detects the position, size, and direction of the key points, and detects features. 250 points. The method of detecting feature points in this way is known as Eigen, Harris, Fast, SIFT, SURF, STAR, MSER, ORB, AKAZE, etc. Here, as shown in FIG. 32 , since the scars 280 are arranged at regular intervals in shapes such as circles and rectangles, they have strong characteristics as corners. Therefore, the flaw 280 can be detected with high precision by counting the feature quantities of the feature points 250 in the image. As shown in FIG. 33 , by comparing the sum of the feature values for each image taken by shifting in the depth direction, it is possible to confirm the change in the value indicating the crack displacement for each modified layer. The control unit 8 estimates the peak value of the change as the position of the flaw 280 . By statistical feature quantities in this way, not only the position of the scar but also the pulse interval can be estimated. The description of the above AI judgment is about the crack 14 and the flaw 280 extending in the X direction, but the crack 14k extending in the direction intersecting the Z direction and the X direction can also use the same algorithm, by comparing more The representative value of the internal image ID, and the relatively higher score is judged as the internal image ID with a relatively clear image of the crack 14k. As an example, FIG. 21 shows a plurality of internal images GD taken at positions different from each other in the Z direction. In FIG. 21 , centering on the shooting position of the internal image GDd shown in (d), (c) is the internal image GDc at the shooting position 1 μm away from the back 60 b side, and (b) is the inside image GDc cut away from the back 60 b side. The internal image GDb of the shooting position of 3 μm, (a) is the internal image GDa of the shooting position 5 μm away from the back 60 b side, (e) is the internal image GDe of the shooting position 1 μm away from the front 60 a side, (f ) is an internal image GDf of an imaging position offset by 3 μm from the surface 60 a side, and (g) is an internal image GDg of an imaging position offset by 5 μm from the surface 60 a side. The imaging position here is a value inside the object 60 . In the example shown in FIG. 21 , the image of the crack 14k is the clearest in the internal image GDd, so the control unit 8 judges that the internal image GDd has a relatively high score and the image of the crack 14k is relatively clear. That is, here, it is judged that the crack 14k is detected in the internal image GDd (let the internal image GDd be a marker image). The control unit 8 can acquire the movement amount Fz when the internal image GDd is captured. By performing the same steps and processing on the multiple rows of modified regions 12 with different positions in the Z direction and the cracks 14k extending from each modified region 12, the control unit 8 can acquire and detect each row of modified regions 12 from multiple rows. Or the amount of movement Fz when the crack 14k is extended. That is, in the present embodiment, a plurality of modified regions 12 and cracks 14k having different positions in the Z direction and known as measured values indicating the positions are formed in the object 60, and in the imaging process, The control unit 8 moves the photographing unit 4 in the Z direction so that the converging point of the light I1 is located at a plurality of positions inside the object 60 in the Z direction to photograph the object 60 , thereby acquiring images including multiple A plurality of internal images GD (an image corresponding to the above-described internal image GDd, that is, a marker image) of each crack 14k extending in each modified region 12 is a clear image. Then, as shown in the second column Q2 of FIG. 22 , the control unit 8 acquires the movement amount Fz when each marker image is captured. As shown in the first column Q1 of FIG. 22 , the control unit 8 uses the actual measurement value of the position of the multi-row modified region 12 (crack 14k) in the Z direction as the position relative to the back surface of the object 60 (with respect to the light I1). The back side of the incident surface, here the distance from the surface 60a) is obtained. Next, the control unit 8 exports the correction coefficient (step S6: export step). As shown in FIG. 22 , the control unit 8 acquires and detects cracks extending from each modified region 12 as information for calculating the measured values of the respective positions of the plurality of modified regions 12 arranged in the Z direction. The amount of movement Fz at 14k (column 2 Q2). On the other hand, the control unit 8 can acquire actual measurement values of the respective positions of the plurality of modified regions 12 arranged in the Z direction (first column Q1 ). Therefore, the control unit 8 can correspond to each of the plurality of modified regions 12 arranged in the Z direction by multiplying the movement amount Fz by the correction coefficient, that is, the estimated value of the position of the modified region 12. Export the correction coefficient in the form of actual measured value. In other words, the control unit 8 outputs the correction coefficient as correction coefficient=actually measured value/movement amount Fz. That is, the control unit 8 executes the export process of multiplying the movement amount Fz when the internal image GD of a clear image including the crack 14k, that is, the marker image is taken, that is, the estimated value, multiplied by the correction coefficient, The correction coefficients are exported in such a manner that the actual measured value of the position of the modified region 12 in the Z direction is obtained. The third column Q3 in FIG. 22 shows the correction coefficients thus exported. Thereafter, the control unit 8 saves the data indicating the exported correction coefficient (step S7), and ends the process. The correction coefficients derived in the above manner are derived based on the internal image GD captured by the light I1 from the imaging unit 4 . Therefore, the correction coefficient reflects the device state of the imaging unit 4 when the internal image GD is captured. In addition, the correction coefficients are exported based on internal images GD captured at a plurality of positions of the object 60 in the Z direction. Therefore, the correction coefficient is obtained by taking into account the position in the Z direction where the converging point of the light I1 is aligned within the object 60 , and the aberration correction amount corresponding to the position. Next, in the observation method of this embodiment, by observing an object including the modified region 12 whose position in the Z direction is not actually measured, the method for obtaining the position of the modified region 12 in the Z direction is implemented. A series of steps for information. Fig. 23 is a flow chart showing the procedure for acquiring information on the position of the modified region in the Z direction in the observation method of the present embodiment. As shown in FIG. 23 , here, an object in which a modified region is formed is prepared. Here, laser processing is performed (step S11: preparation step). However, as a step of the observation method, the step of laser processing is not essential, and for example, an object in which the modified region 12 is formed using another laser processing device (or at another time using the laser processing device 1) may also be prepared. . In this step S11 , as shown in FIG. 24 , an object including a semiconductor substrate 21 is prepared. The semiconductor substrate 21 includes a rear surface (first surface) 21b and a surface (second surface) 21a opposite to the rear surface 21b. In the semiconductor substrate 21, the lines 15 extending in the X direction along the back surface 21b and the front surface 21a are set. The semiconductor substrate 21 is supported by the stage 2 so that the back surface 21 b faces the laser irradiation unit 3 so that the back surface 21 b becomes the incident surface of the laser light L. In this state, while controlling the laser irradiation unit 3, the control unit 8 controls the driving unit 7 and/or the moving mechanism of the mounting table 2 to relatively move the semiconductor substrate 21 along the X direction so that the converging point C of the laser light L is along the line 15 moves relative to the semiconductor substrate 21 . At this time, the control unit 8 displays patterns for causing the spatial light modulator 32 to divide the laser light L into a plurality of (here, two) laser beams L1 and L2. As a result, the converging points C1 and C2 of the laser beams L1 and L2 are formed at a distance Dz in the Z direction and a distance Dx in the X direction inside the semiconductor substrate 21 . As a result, in the semiconductor substrate 21 , a plurality of (here, two rows) modified regions 12 a and 12 b are formed along the line 15 . Therefore, here, the X direction is the process advancing direction in which the light-converging points C1 and C2 advance. Thus, here, the control section 8 executes the following laser processing: by controlling the laser irradiation unit 3 (irradiation section), the semiconductor substrate 21 is irradiated with the laser light L along the X direction which is the extending direction of the line 15. In the semiconductor substrate 21, a plurality of modified regions 12 arranged in the X direction and cracks (cracks 14, 14k) extending from the modified regions 12 are formed. In FIG. 24 and subsequent figures, the functional element layer 22 formed on the surface 21 a of the semiconductor substrate 21 is omitted. Next, look inside. That is, in the next step, the semiconductor substrate 21 is moved to the observation position (step S12). More specifically, the control unit 8 relatively moves the semiconductor substrate 21 to directly below the objective lens 43 of the imaging unit 4 by controlling the driving unit 7 and/or the moving mechanism of the stage 2 . In addition, when the semiconductor substrate 21 on which the modified region 12 is formed is separately prepared, for example, the user may place the semiconductor substrate 21 at the observation position. Next, as shown in FIG. 25 , the semiconductor substrate 21 is photographed using light (transmitted light) I1 having transmittance to the semiconductor substrate 21 (step S13: photographing step). In this step S13, the following photographing process is executed: by controlling the photographing unit 4 (photographing section), the light I1 is made to enter the inside of the semiconductor substrate 21 from the back surface 21b of the semiconductor substrate 21, and the light I1 is used to Among the cracks extending from the modified region 12 , a target crack 14 k extending in a direction intersecting the Z direction and the X direction is photographed. The Y direction is an example of the direction intersecting the X direction which is a process advancing direction, and the Z direction which intersects the back surface 21b and the surface 21a. More specifically, in step S13, the control unit 8 moves the imaging unit 4 in the Z direction by controlling the driving unit 7 (moving unit) and the imaging unit 4, so that the converging point of the light I1 is positioned inside the semiconductor substrate 21. The semiconductor substrate 21 is photographed at a plurality of positions, thereby acquiring a plurality of internal image IDs. As mentioned above, in this embodiment, the objective lens 43 moves integrally with the imaging unit 4 . Therefore, to move the imaging unit 4 is to move the objective lens 43 , and the moving amount of the imaging unit 4 is equal to the moving amount of the objective lens 43 . At this time, the control unit 8 moves the imaging unit 4 in the Z direction by controlling the driving unit 7, and moves the focus point (focus F, virtual focus Fv) of the light I1 in the Z direction while performing multiple operations. Photograph of the sub-semiconductor substrate 21 . The range in which the converging point of the light I1 is moved may be the entire range of the thickness of the semiconductor substrate 21, but here, during the laser processing in step S11, a part of the range RA can be selected, and this part of the range RA includes the range for forming the modified laser beam. The positions in the Z direction where the focal points C1 and C2 of the laser beams L1 and L2 are aligned according to the regions 12a and 12b. The moving interval of the imaging unit 4 in the Z direction when performing multiple imaging, that is, the imaging interval of the semiconductor substrate 21 is arbitrary, but it is preferable to set it more finely from the viewpoint of detecting the crack 14k more accurately. The imaging interval is within 1 μm as an example, and it is 0.2 μm here. Further, here, the control section 8 controls the imaging unit 4 and the drive unit 7 to perform direct observation and rear reflection observation of the semiconductor substrate 21 . More specifically, the control unit 8 first executes the following first photographing process, making the light I1 incident on the semiconductor substrate 21 from the back surface 21b, and moving the photographing unit 4 in the Z direction, thereby making the light I1 not pass through the surface 21a The converging point (focal point F) of the reflected light I1 is moved from the back surface 21b side to the front surface 21a side, and the semiconductor substrate 21 is photographed at a plurality of positions in the Z direction, thereby acquiring a plurality of first images as internal image IDs. 1 internal image ID1. This first photographing process is direct observation. In addition, the control unit 8 executes the second photographing process of causing the light I1 to enter the object from the back surface 21b, moving the photographing unit 4 in the Z direction, thereby concentrating the light I1 reflected on the surface 21a. The light spot (virtual focal point Fv) moves from the front surface 21a side to the rear surface 21b side, and the semiconductor substrate 21 is imaged at a plurality of positions, thereby acquiring a plurality of second internal images ID2 as internal image IDs. This second photographing process is observation performed from the back side (here, in the structure of the semiconductor substrate 21 , referred to as the surface 21 a ) side with respect to the incident surface of the light I1 , and thus is back reflection observation. In the next step, photographing data on the internal image ID acquired by photographing in step S13 is saved (step S14). As described above, in step S13 , the control unit 8 performs imaging while moving the imaging unit 4 (that is, the converging point of the light I1 ) in the Z direction by controlling the driving unit 7 . Therefore, the control unit 8 can acquire the movement amount Fz of the imaging unit 4 when each internal image is captured. Here, the information on the amount of movement Fz of each internal image ID can be associated with each internal image ID and stored as photographic data. The photographic data can be stored in any storage device that can be communicated with the control unit 8 regardless of inside or outside the control unit 8 and the laser processing apparatus 1 . The amount of movement of the imaging unit 4 (objective lens 43), as an example, can be selected from the position where the light I1 is aligned with the back surface 21b of the semiconductor substrate 21, so that the light I1 is aligned with the semiconductor substrate. The movement amount of the imaging unit 4 when the imaging unit 4 is moved in the Z direction according to a desired position inside the substrate 21 . Next, the control unit 8 inputs photographic data from a predetermined storage device (step S15). Then, the control unit 8 judges the formation state of the crack 14k (step S16). Here, as an example, the control unit 8 automatically judges an internal image ID whose image of the crack 14k is relatively clear among a plurality of internal image IDs by image recognition (performs AI judgment). An example of AI judgment is as described above. FIG. 26 shows a plurality of internal image IDs taken at positions different from each other in the Z direction. In FIG. 26 , centering on the shooting position of the internal image IDd shown in (d), (c) is the internal image IDc at the shooting position 1 μm away from the rear surface 21b side, and (b) is the internal image IDc cut away from the rear surface 21b side. The internal image IDb of the photographing position of 3 μm, (a) is the internal image IDa of the photographing position 5 μm removed from the back 21 b side, (e) is the internal image IDe of the photographing position 1 μm removed from the front 21 a side, (f ) is an internal image IDf of an imaging position offset by 3 μm from the surface 21 a side, and (g) is an internal image IDg of an imaging position offset by 5 μm from the surface 21 a side. However, the imaging position here is a value inside the semiconductor substrate 21 . In the example shown in FIG. 26 , according to the fact that the image of the crack 14k is the clearest in the internal image IDd, the control unit 8 judges that the internal image IDd has a relatively high score and the image of the crack 14k is relatively clear. (that is, it is determined that the crack 14k is detected in the internal image IDd). The control unit 8 can acquire the amount of movement when the internal image IDd is captured. Therefore, the control unit 8 can calculate the crack position of the crack 14k based on the movement amount when the internal image IDd is captured. In this way, the control unit 8 executes the following imaging process: by controlling the imaging unit 4 and the driving unit 7, the light I1 is incident on the semiconductor substrate 21 from the back surface 21b, and the imaging unit 4 (objective lens 43) is aligned in the Z direction. The semiconductor substrate 21 is photographed with the light I1 while moving, thereby acquiring a detection image which is an internal image ID including a clear image of the crack 14k. In addition, the control unit 8 executes a calculation process of calculating a moving distance Fz extending in a direction intersecting the Z direction and the X direction based on a plurality of internal image IDs and the movement amount Fz of the imaging unit 4 when each internal image ID is captured. The crack 14k is the position of the target crack in the Z direction, that is, the crack position. More specifically, in the calculation process, the control unit 8 determines an internal image ID with a clear image of the crack 14k among a plurality of internal image IDs, and calculates based on the movement amount Fz when the determined internal image ID is captured. Crack location. The crack position can be calculated by multiplying the movement amount Fz by a predetermined correction coefficient, for example. The calibration coefficients have been exported through the above steps S1-7. That is, the control unit 8 stores a plurality of correction coefficients corresponding to the movement amount Fz, and calculates the correction coefficient corresponding to the crack position of the crack 14k using the correction coefficients corresponding to the movement amount Fz when the detected image is captured in the calculation process. The estimated value of the position of the mass region 12. The control part 8 can perform the above-mentioned calculation of the crack position of the crack 14k for both the 1st internal image ID1 acquired by direct observation, and the 2nd internal image ID2 acquired by back reflection observation. Thus, the control unit 8 can calculate the crack position of the crack 14k corresponding to the first internal image ID1 and located relatively on the back surface 21b side, and the crack position of the crack 14k corresponding to the second internal image ID2 and located relatively on the front surface 21a side. Crack location. That is, in this case, the control unit 8 executes the first arithmetic processing and the second arithmetic processing. In the first arithmetic processing, the first internal image in which the crack 14k is clear among the plurality of first internal images ID1 is judged, The first crack position Z1 as the crack position is calculated based on the movement amount of the imaging unit 4 when the determined first internal image is captured, and in the second calculation process, among the plurality of second internal images ID2 is determined The second internal image of the crack 14k is clear, and the second crack position Z2 (for the first crack position Z1 and For an example of the second crack position Z2, refer to FIG. 15). The distance between the first crack position Z1 relatively located on the back surface 21b side and the second crack position Z2 relatively located on the surface 21a side defines the portion (crack initiation portion) in the modified region 12 where the crack 14k is formed. width. Next, in step S16 , the control unit 8 estimates the position of the modified region 12 and the like based on the acquired crack position and the like. That is, here, the control unit 8 executes an estimation process of estimating the end (defect) on the rear surface 21b side of the modified region 12 based on the formation conditions of the modified region 12 (here, the processing conditions of laser processing) and the position of the crack. upper end of the upper region) in the Z direction, the position of the end of the modified region 12 on the surface 21a side (defect lower end) in the Z direction, and the width of the modified region 12 in the Z direction (the upper end of the defect upper region and At least one of the spacing at the lower end of the defect). Here, the control unit 8 calculates the first crack position Z1 of the crack 14k (upper crack) on the back surface 21b side based on direct observation, and calculates the second crack position Z2 of the crack 14k (lower crack) on the surface 21a side based on back reflection observation. Therefore, the control unit 8 can calculate the width of the crack initiation portion inside the semiconductor substrate 21 as the distance between the first crack position Z1 of the upper crack and the second crack position Z2 of the lower crack. Then, the control unit 8 can calculate the width of the modified region 12 inside the semiconductor substrate 21 in the Z direction by, for example, multiplying the calculated width of the crack initiation portion by a coefficient related to the processing conditions of laser processing. The coefficient here is determined based on various conditions that affect the formation of the modified region 12 , such as the wavelength of the laser light L during laser processing, the amount of aberration correction, the pulse width, and the pulse energy. The coefficient here is about 3.0 in one example. In this way, in the estimation process, the control unit 8 can estimate the position of the modified region 12 in the Z direction based on the formation conditions of the modified region 12 (processing conditions of laser processing) and the distance between the first crack position Z1 and the second crack position Z2. width. On the other hand, the control unit 8 can calculate the lower end of the modified region 12 on the surface 21a side by subtracting the assumed modified region width, that is, the assumed modified region width, from the first crack position Z1 of the upper crack. s position. The modified region width can be determined based on various conditions affecting the formation of the modified region 12 such as the wavelength of the laser light L during laser processing, the amount of aberration correction, the pulse width, and the pulse energy, for example. Assume that the width of the modified region is about 20 μm as an example. Furthermore, the control unit 8 can calculate the position of the lower end of the surface 21a of the modified region 12 by subtracting the width of the assumed defect region 12m, that is, the assumed defect region width, from the second crack position Z2 of the lower crack. The hypothetical defect region width can be determined based on, for example, various conditions affecting the formation of the modified region 12 such as the wavelength of the laser L during laser processing, the amount of aberration correction, the pulse width, and the pulse energy. Assume that the defect region width is about 10 μm as an example. Further, the control unit 8 can calculate the position of the upper end of the modified region 12 on the rear surface 21b side by adding the width of the assumed defect upper region 12n, that is, the assumed defect upper region width, to the second crack position Z2 of the lower crack. The width of the region above the hypothetical defect can be determined based on various conditions that affect the formation of the modified region 12 such as the wavelength of the laser light L during laser processing, the amount of aberration correction, the pulse width, and the pulse energy. Assume that the width of the region above the defect is about 10 μm as an example. As described above, the control unit 8 estimates and acquires various information on the position of the modified region 12 in step S16. In the next step, the control unit 8 outputs the information of the judgment result of step S16 to an arbitrary storage device (step S17), and stores it in the storage device (step S18). Thereafter, as necessary, various information is displayed on the display 150 in a state where the user's input can be accepted (step S19), and the processing is completed. The information displayed on the display 150 is, for example, the first crack position Z1, the second crack position Z2, the width of the initial part, the position of the end of the modified region 12, and the width of the modified region 12 in the Z direction. In this way, in step S19 , the control unit 8 controls the display 150 to execute display processing for displaying information on the crack position on the display 150 . As described above, the observation method using the laser processing apparatus 1 is completed. In this embodiment, the observation method is performed by the imaging unit 4 , the driving unit 7 and the control unit 8 in the laser processing device 1 . In other words, in the laser processing apparatus 1, the observation device 1A (refer to FIG. 1 ) is constituted by the imaging unit 4, the drive unit 7, and the control unit 8. The object 60 and the semiconductor substrate 21 are photographed by the permanent light I1, the driving unit 7 is used to move the imaging unit 4 relative to the object 60 and the semiconductor substrate 21, and the control unit 8 is used to control at least the imaging unit 4 and the driving unit 7. . FIG. 27 is a graph showing the error between the measured value and the actual measured value of the position of the modified region. As shown in FIG. 27 , in the observation method of this embodiment, the correction coefficients derived in consideration of the state of the device, the position in the Z direction (the depth position in FIG. 27 ), and the amount of aberration correction in steps S1 to S7 are used, In steps S11 to S19, the estimated value of the modified region 12 is calculated. Therefore, the error between the measured value and the actual measured value in this embodiment is approximately within 6 μm. However, the error of the comparative example using a certain (fixed) correction coefficient is as large as about 19 μm compared with the present embodiment. As described above, in the observation device 1A of the present embodiment and the object 60 of the observation method, a mark (here, a modified mark) whose measured value is known at a position in the Z direction intersecting the back surface 60b and the front surface 60a is provided. quality region 12 and crack 14k). In the observation device 1A and the observation method of the present embodiment, by imaging such an object 60 while moving the imaging unit 4, an image including the crack 14k as an internal image GD of the object 60 can be obtained. Tag the image. Then, the correction coefficient is derived so that the value (measured value) obtained by multiplying the movement amount Fz when the marker image is obtained by the correction coefficient is an actual measurement value of the known position of the modified region 12 . That is, according to the observation device 1A and the observation method, correction coefficients corresponding to the state of the device when the marker image is captured and the movement amount of the imaging unit 4 (that is, the observation depth) when the marker image is captured can be derived. Therefore, when the modified region 12 is observed with the light I1 to calculate the estimated value of the modified region 12 position, if this correction coefficient is used, more accurate information on the modified region 12 position can be obtained. In addition, in the observation device 1A of the present embodiment, a plurality of modified regions 12 and cracks 14k whose positions in the Z direction are different from each other and whose actual measurement values are known are formed on the object 60 . In this case, the control unit 8 photographs the object 60 by positioning the focusing point at a plurality of positions inside the object 60 in the Z direction, and obtains the image including the area extending from each modified region 12 of the plurality of modified regions 12. Multiple marker images of images of individual cracks 14k. Then, in the export process, the control unit 8 converts each value obtained by multiplying each movement amount Fz of the imaging unit 4 by a correction coefficient, that is, an estimated value, into a plurality of modified cracks 14k. A plurality of correction coefficients are exported in the form of the actual measured values of the respective positions of the regions 12 . Therefore, when the modified region 12 is observed with the light I1 and the estimated value of the modified region 12 position is calculated, information on the modified region 12 position can be acquired more accurately over a wider range in the Z direction. In addition, in the observation device 1A, the modified region 12 aligned in the Z direction and the cracks 14, 14k extending from the modified region 12 are formed in the object 60, and in the imaging process, by moving the imaging unit 4 along the Moving in the Z direction, while moving the focus point of the light I1, the object 60 is photographed with the light I1, and cracks extending in a direction intersecting the X direction and the Z direction, including the cracks 14 and 14k, are acquired as marker images. 14k like internal image GD. According to the findings of the inventors of the present invention, when the modified region 12 is formed by, for example, laser processing inside the object 60 , cracks extending in various directions from the modified region 12 may be formed. Furthermore, among the cracks, the crack 14k extending in the direction intersecting the Z direction and the X direction is accurately detected by the light I1 transmitted from the object 60 compared with the modified region 12, wherein the Z direction It is a direction intersecting the back surface 60b which is the laser beam incident surface of the object 60, and the X direction is the advancing direction of laser processing. Therefore, if the internal image GD including the image of the crack 14k is used as the marker image as described above, it is possible to reduce the variation in the movement amount Fz of the imaging unit 4 when the marker image is captured. As a result, more accurate correction coefficients can be exported. In the observation device 1A of the present embodiment, the imaging unit 4 has a correction ring lens including an objective lens 43 and a correction ring 43 a provided on the objective lens 43 for correcting aberrations generated in the object 60 . In this way, when the objective lens 43 for condensing the light I1 toward the object 60 is provided with the correction ring 43a, the state of the device may change before and after the operation of the correction ring 43a. Therefore, it is more effective to export the correction coefficient according to the state of the device as described above. Here, the observation device 1A of the present embodiment includes: an imaging unit 4 having an objective lens 43 for converging light I1 having transmittance with respect to the semiconductor substrate 21 toward the semiconductor substrate 21, and for imaging the semiconductor substrate 21 with the light I1. The substrate 21 performs imaging; the driving unit 7 for moving the objective lens 43 relative to the semiconductor substrate 21 ; and the control unit 8 for controlling at least the imaging unit 4 and the driving unit 7 . The semiconductor substrate 21 has a back surface 21b and a surface 21a opposite to the back surface 21b, and the semiconductor substrate 21 is provided with a modified region 12 and cracks 14, 14k extending from the modified region 12. The control unit 8 executes imaging processing and arithmetic processing. In the aforementioned imaging processing, by controlling the imaging unit 4 and the driving unit 7, the light I1 is incident on the semiconductor substrate 21 from the back surface 21b, and the imaging unit 4 is moved along the Z direction. While moving, the semiconductor substrate 21 is photographed with the light I1, thereby acquiring the internal image ID of the image including the crack 14k, that is, the inspection image. The movement amount Fz of the imaging unit 4 is multiplied by the correction coefficient to calculate the position of the crack 14k in the Z direction, that is, the crack position. The control unit 8 stores a plurality of correction coefficients corresponding to the movement amount Fz. This observation device 1A stores a correction coefficient corresponding to the movement amount Fz of the imaging unit 4 as described above. Therefore, by calculating the estimated value of the position of the crack 14k using this correction coefficient, more accurate information on the position of the modified region 12 can be obtained. The object 60 of this embodiment has a back surface 60b and a surface 60a on the opposite side of the back surface 60b, and a modified region 12 and a crack 14k extending from the modified region 12 are provided as marks, and the object 60 is used for export correction. This correction coefficient is used to calculate the measured values of the positions of the modified regions 12 and the cracks 14 k in the Z direction intersecting the back surface 60 b and the surface 60 a from the measured values of the modified regions 12 positions. Using this object 60, the correction coefficients can be exported as described above. The above-mentioned embodiment is for explaining one aspect of this invention. Therefore, the present invention is not limited to the above-mentioned embodiments, and can be arbitrarily modified. For example, in the above-described embodiment, the drive unit 7 for moving the objective lens 43 together with the imaging unit 4 was illustrated as means for moving the objective lens 43 in the Z direction relative to the semiconductor substrate 21 . However, for example, an actuator may be used to move only the objective lens 43 in the Z direction. In addition, in the above-mentioned embodiment, the example in which the control unit 8 automatically judges the image in steps S5 and S16 is described, but the control unit 8 may acquire the crack position of the crack 14k based on the judgment result of the user. In this case, for example, the control unit 8 causes the display 150 to display a plurality of internal images GD, ID, and displays one internal image prompting the determination (selection) of the image of the crack 14k from the plurality of internal images GD, ID. like information. Then, the control unit 8 can receive the input of the judgment result through the display 150, and calculate the crack position of the crack 14k based on the internal image GD and the movement amount Fz of the ID corresponding to the judgment result. In this case, the display 150 is a display unit for displaying information, and is also an input receiving unit for receiving input. In this case, the processing load on the control unit 8 for performing image recognition and the like is reduced. In addition, in the above-mentioned embodiment, in steps S2 and S13, the observation of one modified region 12 was performed both of direct observation and rear reflection observation, and the first internal images as internal images GD and ID were acquired. GD1, ID1 and the second internal image GD2, ID2. However, in steps S2 and S13, only one of direct observation and back reflection observation may be performed. In this case, since one of the first internal image GD1, ID1 and the second internal image GD2, ID2 can be obtained, the export of the correction coefficient and the modification of the modified region 12 can also be performed based on this one. Estimation of the position and width of the end. Here, in the observation apparatus 1A, the object 60 for exporting the correction coefficient may always be provided. That is, the observation apparatus 1A may include a setting unit (for example, the mounting table 2A in FIG. 1 ) for setting the object 60 and the object 60 set on the setting unit. In this way, by always setting the object 60 provided with the modified region 12 and the crack 14k as a mark, it is possible to export the correction coefficient at any timing. Further, in the above-mentioned embodiment, when exporting the correction coefficient, the internal image including the crack 14k is used as the marker image, and the movement amount Fz when the marker image is obtained by photographing is used. That is, by detecting the crack 14k, the correction coefficient is exported based on the position where the crack 14k is detected. However, in the observation device 1A and the observation method, when exporting the correction coefficients, the modified region 12 itself may be detected. In this case, the processing of the control unit 8 is as follows. That is, the control unit 8 executes the imaging process and the exporting process. In the imaging process, by controlling the imaging unit 4 and the drive unit 7, the light I1 is made to enter the inside of the object 60 from the back surface 60b, and the imaging unit 4 Move in the Z direction to move the focus point of the light I1 in the Z direction, and image the object 60 with the light I1, thereby acquiring an image including the modified region 12 as an internal image GD of the object 60 In the export process, after the photographing process, the value obtained by multiplying the movement amount Fz of the imaging unit 4 when the marker image is obtained by multiplying the correction coefficient, that is, the estimated value is the value of the modified region 12 in the export process. Export the correction coefficient in the form of the measured value of the position. In this way, when the modified region 12 itself is the detection object, in order to export the correction coefficients at more positions in the Z direction, it is necessary to use smaller intervals in the Z direction (for example, the movement amount Fz is 5 μm, and the object 60 inside is about 20 μm) to capture an internal image, there is a problem that an image of one modified region 12 is captured across a plurality of internal images. In this case, the image of the modified region 12 at a certain position in the Z direction and the image of the modified region 12 at another position in the Z direction overlap in one internal image. In order to solve such a problem, when the modified regions 12 are the detection targets, it is conceivable to shift rows of modified regions 12 arranged in the Z direction from each other in the Y direction as shown in FIG. 28 . In this way, it can be avoided that the image of the modified region 12 at a certain position in the Z direction and the image of the modified region 12 at another position in the Z direction overlap in one internal image, so that internal images can be imaged at a smaller interval. The image is photographed, and the correction coefficients of more positions can be exported. In addition, when the detection target for exporting the correction coefficient is the modified region 12, in the subsequent step of acquiring information on the position of the modified region 12 of the semiconductor substrate 21, the crack 14k may be replaced by The detection is performed to detect the modified region 12 itself. Here, in the above-mentioned embodiment, when exporting the correction coefficients, an example was given of using the object 60 in which the modified region 12 and the crack 14k whose actual measurement values are known are set as marks. However, the object for exporting the correction coefficient is not limited to this. For example, one surface of a wafer with a known thickness may be affixed with a test icon as a mark, or a predetermined pattern may be formed as a mark on one surface of a wafer with a known thickness. Get something as an object. In these cases, by preparing a plurality of objects having different thicknesses, it is possible to export correction coefficients for a plurality of positions in the Z direction.

1:鐳射加工裝置 1A:觀察裝置 2:載置台 2A:載置台 3:鐳射照射單元 4:攝影單元 5:攝影單元 6:攝影單元 7:驅動單元 8:控制部 11:對象物 12:改質區域 12a:改質區域 12b:改質區域 12s:改質點 12m:缺陷區域 12n:缺陷上方區域 14:裂紋 14a:裂紋 14b:裂紋 14c:裂紋 14d:裂紋 14e:前端 14k:裂紋 15:線 20:晶圓 21:半導體基板 21a:表面 21b:背面 21c:缺口 22:功能元件層 22a:功能元件 23:格線區域 31:光源 32:空間光調變器 33:聚光透鏡 41:光源 42:反射鏡 43:物鏡 43a:校正環 44:光檢測部 51:光源 52:反射鏡 53:透鏡 54:光檢測部 60:對象物 60a:表面 60b:背面 140:檢測直線組 150:顯示器 250:特徵點 280:傷痕 A1:區域 A2:區域 A3:區域 C:聚光點 C1:聚光點 C2:聚光點 Dx:間隔距離 Dz:間隔距離 RA:範圍 ID:內部圖像 ID1:內部圖像 ID2:內部圖像 IDa:內部圖像 IDb:內部圖像 IDc:內部圖像 IDd:內部圖像 IDe:內部圖像 IDf:內部圖像 IDg:內部圖像 L:鐳射光 L1:鐳射光 L2:鐳射光 I1:光 I2:光 F:焦點 Fv:虛擬焦點 Z1:第1裂紋位置 Z2:第2裂紋位置 S1:匯出用對象物移動 S2:攝影 S3:保存攝影資料 S4:輸入資料 S5:判定狀態 S6:匯出校正係數 S7:保存資料 S11:鐳射加工 S12:對象物移動 S13:攝影 S14:保存攝影資料 S15:輸入資料 S16:判定狀態 S17:輸出判定結果 S18:保存資料 S19:顯示 GD:內部圖像 GD1:內部圖像 GD2:內部圖像 GDa:內部圖像 GDb:內部圖像 GDc:內部圖像 GDd:內部圖像 GDe:內部圖像 GDf:內部圖像 GDg:內部圖像 Hz:移動量 1: Laser processing device 1A: Observation device 2: Carrying table 2A: Carrying table 3: Laser irradiation unit 4: Photography unit 5: Photography unit 6: Photography unit 7: Drive unit 8: Control Department 11: object 12:Modified area 12a: Modified area 12b:Modified area 12s: modification point 12m: defect area 12n: area above the defect 14: Crack 14a: Crack 14b: Crack 14c: crack 14d: Crack 14e: front end 14k: crack 15: line 20: Wafer 21: Semiconductor substrate 21a: Surface 21b: back 21c: Gap 22: Functional component layer 22a: Functional elements 23: grid area 31: light source 32: Spatial light modulator 33: Concentrating lens 41: light source 42: Mirror 43: objective lens 43a: Calibration ring 44: Light detection unit 51: light source 52: Mirror 53: lens 54: Light detection unit 60: object 60a: surface 60b: back 140:Detect line group 150: display 250: Feature points 280: scars A1: area A2: area A3: area C: focus point C1: focus point C2: focus point Dx: separation distance Dz: separation distance RA: range ID: internal image ID1: internal image ID2: internal image IDa: internal image IDb: internal image IDc: internal image IDd: internal image IDe: internal image IDf: internal image IDg: internal image L: laser light L1: laser light L2: laser light I1: light I2: light F: focus Fv: virtual focus Z1: The first crack position Z2: The second crack position S1: Move object for export S2: Photography S3: save photographic data S4: input data S5: Judgment status S6: Export correction coefficient S7: save data S11:Laser processing S12: Object movement S13: Photography S14: Save photographic data S15: input data S16: Judgment status S17: output judgment result S18: save data S19: display GD: internal image GD1: Internal image GD2: Internal Image GDa: Internal Image GDb: internal image GDc: internal image GDd: internal image GDe: internal image GDf: internal image GDg: internal image Hz: amount of movement

[圖1]是一個實施型態的鐳射加工裝置的結構圖。 [圖2]是一個實施型態的晶圓的俯視圖。 [圖3]是圖2所示的晶圓的一部分的截面圖。 [圖4]是圖1所示的鐳射照射單元的結構圖。 [圖5]是圖1所示的檢查用攝影單元的結構圖。 [圖6]是圖1所示的對準校正用攝影單元的結構圖。 [圖7]是用於說明圖5所示的檢查用攝影單元的攝影原理的晶圓的截面圖、和該檢查用攝影單元所獲得的各部位處的圖像。 [圖8]是用於說明圖5所示的檢查用攝影單元的攝影原理的晶圓的截面圖、和該檢查用攝影單元所獲得的各部位處的圖像。 [圖9]是形成在半導體基板的內部的改質區域和裂紋的SEM圖像。 [圖10]是形成在半導體基板的內部的改質區域和裂紋的SEM圖像。 [圖11]是用於說明圖5所示的檢查用攝影單元的攝影原理的示意圖。 [圖12]是用於說明圖5所示的檢查用攝影單元的攝影原理的示意圖。 [圖13]是表示形成了改質區域的對象物的圖。 [圖14]是關於Z方向上的改質區域和裂紋的位置的圖表。 [圖15]是在對象物的截面照片對檢測結果進行繪製而得到的圖。 [圖16]是用於說明校正係數的示意圖。 [圖17]是表示檢測對象在Z方向上的位置與檢測出檢測對象時的移動量的關係的圖表。 [圖18]是表示用於匯出本實施型態的觀察方法中的校正係數的步驟的流程圖。 [圖19]是表示用於匯出校正係數的對象物的側視圖。 [圖20]是表示圖18所示的觀察方法的一個步驟的圖。 [圖21]是在Z方向上彼此不同的位置攝影得到的多個內部圖像。 [圖22]是表示改質區域的位置的實測值與移動量與校正係數的關係的表。 [圖23]是表示本實施型態的觀察方法中,用於獲取關於改質區域在Z方向上的位置的資訊的步驟的流程圖。 [圖24]是表示圖23所示的觀察方法的一個步驟的圖。 [圖25]是表示圖23所示的觀察方法的一個步驟的圖。 [圖26]是表示改質區域的位置的實測值與移動量與校正係數的關係的表。 [圖27]是表示改質區域的位置的測算值與實測值的誤差的圖表。 [圖28]是表示變形例的對象物的示意的截面圖。 [圖29]是對裂紋檢測進行說明的圖。 [圖30]是對裂紋檢測進行說明的圖。 [圖31]是對傷痕檢測進行說明的圖。 [圖32]是對傷痕檢測進行說明的圖。 [圖33]是對傷痕檢測進行說明的圖。 [FIG. 1] It is a block diagram of the laser processing apparatus of one embodiment. [ Fig. 2 ] is a plan view of a wafer of an embodiment. [ Fig. 3 ] is a cross-sectional view of a part of the wafer shown in Fig. 2 . [ Fig. 4 ] is a configuration diagram of the laser irradiation unit shown in Fig. 1 . [ Fig. 5 ] is a configuration diagram of the imaging unit for inspection shown in Fig. 1 . [FIG. 6] It is a block diagram of the imaging unit for alignment correction shown in FIG. 1. [FIG. [ Fig. 7] Fig. 7 is a cross-sectional view of a wafer for explaining the imaging principle of the inspection imaging unit shown in Fig. 5 , and images at various locations obtained by the inspection imaging unit. [ Fig. 8] Fig. 8 is a cross-sectional view of a wafer for explaining the imaging principle of the inspection imaging unit shown in Fig. 5 , and images at various locations obtained by the inspection imaging unit. [ Fig. 9 ] is an SEM image of a modified region and cracks formed inside a semiconductor substrate. [ Fig. 10 ] is an SEM image of a modified region and cracks formed inside a semiconductor substrate. [ Fig. 11 ] is a schematic diagram for explaining the imaging principle of the inspection imaging unit shown in Fig. 5 . [ Fig. 12 ] is a schematic diagram for explaining the imaging principle of the inspection imaging unit shown in Fig. 5 . [ Fig. 13 ] is a diagram showing an object on which a modified region is formed. [ Fig. 14 ] is a graph regarding the positions of modified regions and cracks in the Z direction. [ Fig. 15 ] is a diagram obtained by plotting detection results on cross-sectional photographs of objects. [ Fig. 16 ] is a schematic diagram for explaining correction coefficients. [ Fig. 17 ] is a graph showing the relationship between the position of the detection target in the Z direction and the movement amount when the detection target is detected. [FIG. 18] It is a flowchart which shows the procedure for exporting the correction coefficient in the observation method of this embodiment. [ Fig. 19 ] is a side view showing an object for exporting correction coefficients. [ Fig. 20 ] is a diagram showing one step of the observation method shown in Fig. 18 . [ Fig. 21 ] are a plurality of internal images taken at positions different from each other in the Z direction. [ Fig. 22 ] is a table showing the relationship between the actually measured value of the position of the modified region, the amount of movement, and the correction coefficient. [ Fig. 23 ] is a flowchart showing a procedure for acquiring information on the position of the modified region in the Z direction in the observation method of the present embodiment. [ Fig. 24 ] is a diagram showing one step of the observation method shown in Fig. 23 . [ Fig. 25 ] is a diagram showing one step of the observation method shown in Fig. 23 . [ Fig. 26 ] is a table showing the relationship between the actually measured value of the position of the modified region, the amount of movement, and the correction coefficient. [ Fig. 27 ] is a graph showing the error between the measured value and the actual measured value of the position of the modified region. [ Fig. 28 ] is a schematic cross-sectional view showing an object of a modified example. [ Fig. 29 ] is a diagram illustrating crack detection. [ Fig. 30 ] is a diagram illustrating crack detection. [ Fig. 31 ] is a diagram illustrating flaw detection. [ Fig. 32 ] is a diagram illustrating flaw detection. [ Fig. 33 ] is a diagram illustrating flaw detection.

Claims (8)

一種觀察裝置,其特徵在於,包括: 攝影部,其具有用於使相對於對象物具有透射性的透射光向前述對象物聚光的聚光透鏡,用於利用前述透射光對前述對象物進行攝影; 用於使前述聚光透鏡相對於前述對象物相對地移動的移動部;和 用於至少控制前述攝影部和前述移動部的控制部, 前述對象物具有第1面和前述第1面的相反側的第2面, 在前述對象物設置有與前述第1面和前述第2面交叉的Z方向上的位置的實測值已知的標記, 前述控制部執行攝影處理和匯出處理,其中, 在前述攝影處理中,控制前述攝影部和前述移動部,使前述透射光從前述第1面射入至前述對象物的內部,使前述聚光透鏡沿前述Z方向移動並利用前述透射光對前述對象物進行攝影,來獲取作為前述對象物的內部圖像的包含前述標記的像的標記圖像, 前述匯出處理在前述攝影處理之後匯出校正係數,前述校正係數使得該校正係數與拍攝得到前述標記圖像時的前述聚光透鏡的移動量的乘積的值即測算值成為前述實測值。 An observation device, characterized in that it comprises: an imaging unit having a condensing lens for condensing transmitted light having transmittance with respect to the object toward the object, and for photographing the object with the transmitted light; a moving unit for relatively moving the condensing lens relative to the object; and a control unit for controlling at least the aforementioned photographing unit and the aforementioned moving unit, The aforementioned object has a first surface and a second surface opposite to the first surface, A mark whose measured value is known at a position in the Z direction intersecting the first surface and the second surface is provided on the object, The aforementioned control unit executes the photographing process and the exporting process, wherein, In the photographing process, the photographing unit and the moving unit are controlled so that the transmitted light enters the inside of the object from the first surface, the condenser lens is moved in the Z direction, and the transmitted light is used to control the The object is photographed to obtain a marker image including the image of the marker as an internal image of the object, The export process exports a correction coefficient after the photographing process so that the measured value obtained by multiplying the correction coefficient by the moving amount of the condenser lens when the marker image is captured becomes the actual measured value. 如請求項1所述的觀察裝置,其中: 在前述對象物中,形成有前述Z方向上的位置彼此不同且該位置的前述實測值已知的多個前述標記, 在前述攝影處理中,前述控制部使前述聚光透鏡沿前述Z方向相對移動,使得前述透射光的聚光點位於前述Z方向上的前述對象物的內部的多個位置來對前述對象物進行攝影,藉此獲取包含多個前述標記各自的像的多個前述標記圖像, 在前述匯出處理中,前述控制部匯出多個前述校正係數,多個前述校正係數使得前述校正係數與拍攝得到多個前述標記圖像的各個標記圖像時的前述聚光透鏡的各個前述移動量的乘積的值即測算值各自分別成為多個前述標記的前述實測值的各者。 The observation device as claimed in claim 1, wherein: In the aforementioned object, a plurality of the aforementioned marks having different positions in the Z direction and whose aforementioned actually measured values of the positions are known are formed, In the photographing process, the control unit relatively moves the condensing lens in the Z direction so that the condensing points of the transmitted light are located at a plurality of positions inside the target object in the Z direction, so that the target object is scanned. photography, whereby a plurality of images of the aforementioned markers are obtained comprising images of respective images of the plurality of aforementioned markers, In the export process, the control unit exports a plurality of the correction coefficients such that the correction coefficients are consistent with each of the condenser lenses when each of the plurality of marker images is captured. The value of the product of the movement amount, that is, the measured value is each of the above-mentioned actual measurement values of the plurality of the above-mentioned marks. 如請求項1或2所述的觀察裝置,其中: 在前述對象物中,作為前述標記形成有在沿前述第1面和前述第2面的X方向上排列的改質區域和從前述改質區域延伸的裂紋, 在前述攝影處理中,使聚光透鏡沿前述Z方向移動,來使前述透射光的聚光點移動而利用前述透射光對前述對象物進行攝影,藉此獲取包含前述裂紋中的沿與前述X方向和前述Z方向交叉的方向延伸的裂紋的像的前述內部圖像作為前述標記圖像。 The observation device as claimed in claim 1 or 2, wherein: In the object, modified regions aligned in the X direction along the first surface and the second surface and cracks extending from the modified regions are formed as the marks, In the photographing process, the condensing lens is moved in the Z direction to move the converging point of the transmitted light, and the object is photographed with the transmitted light, thereby obtaining the edge including the crack and the X The aforementioned internal image of an image of a crack extending in a direction intersecting with the aforementioned Z direction is used as the aforementioned marker image. 如請求項1~3中任一項所述的觀察裝置,其中: 前述攝影部具有校正環透鏡,該校正環透鏡包含前述聚光透鏡和設置於前述聚光透鏡且用於對在前述對象物產生的像差進行校正的校正環。 The observation device as described in any one of claim items 1 to 3, wherein: The imaging unit has a correction ring lens including the condenser lens and a correction ring provided on the condenser lens for correcting aberrations occurring in the object. 如請求項1~4中任一項所述的觀察裝置,其中,具備: 用於設置前述對象物的設置部;和 設置於前述設置部的前述對象物。 The observation device as described in any one of claims 1 to 4, wherein: a setting part for setting the aforementioned object; and The aforementioned object installed in the aforementioned setting portion. 一種觀察方法,其特徵在於,具備: 準備對象物的準備步驟,前述對象物具有第1面和前述第1面的相反側的第2面,在前述對象物形成有與前述第1面和前述第2面交叉的Z方向上的位置的實測值已知的標記; 攝影步驟,使相對於前述對象物具有透射性的透射光從前述第1面射入至前述對象物的內部,使用來將前述透射光聚光的聚光透鏡沿前述Z方向移動並利用前述透射光對前述對象物進行攝影,來獲取作為前述對象物的內部圖像的包含前述標記的像;和 匯出步驟,在前述攝影步驟之後匯出校正係數,前述校正係數使得該校正係數與拍攝得到前述標記圖像時的前述聚光透鏡的移動量的乘積的值即測算值成為前述實測值。 A kind of observation method is characterized in that, has: A preparation step for preparing an object, wherein the object has a first surface and a second surface opposite to the first surface, and a position in the Z direction intersecting the first surface and the second surface is formed on the object The marker whose measured value is known; In the photographing step, the transmitted light having transmittance with respect to the object is made to enter the inside of the object from the first surface, and the condensing lens for condensing the transmitted light is moved in the Z direction to utilize the transmitted light. photographing the object with light to acquire an image including the mark as an internal image of the object; and In the exporting step, after the photographing step, a correction coefficient is exported so that the measured value obtained by multiplying the correction coefficient and the moving amount of the condenser lens when the marker image is captured becomes the actual measured value. 一種觀察裝置,其特徵在於,具備: 攝影部,其具有用於使相對於對象物具有透射性的透射光向前述對象物聚光的聚光透鏡,用於利用前述透射光對前述對象物進行攝影; 用於使前述聚光透鏡相對於前述對象物相對地移動的移動部;和 用於至少控制前述攝影部和前述移動部的控制部, 前述對象物具有第1面和前述第1面的相反側的第2面, 在前述對象物中設有改質區域和從前述改質區域延伸的裂紋, 前述控制部執行攝影處理和運算處理,其中, 在前述攝影處理中,控制前述攝影部和前述移動部,使前述透射光從前述第1面射入至前述對象物,使前述聚光透鏡沿與前述第1面和前述第2面交叉的Z方向移動並利用前述透射光對前述對象物進行攝影,來獲取作為內部圖像的包含前述改質區域和/或前述裂紋的像的檢測圖像, 前述運算處理在前述攝影處理之後,將校正係數乘以拍攝得到前述檢測圖像時的前述聚光透鏡的移動量,來運算前述改質區域和/或前述裂紋在前述Z方向上的位置的測算值, 前述控制部保存有與前述移動量相應的多個校正係數。 An observation device, characterized in that it has: an imaging unit having a condensing lens for condensing transmitted light having transmittance with respect to the object toward the object, and for photographing the object with the transmitted light; a moving unit for relatively moving the condensing lens relative to the object; and a control unit for controlling at least the aforementioned photographing unit and the aforementioned moving unit, The aforementioned object has a first surface and a second surface opposite to the first surface, A modified region and a crack extending from the modified region are provided in the object, The aforementioned control unit executes imaging processing and calculation processing, wherein, In the photographing process, the photographing unit and the moving unit are controlled so that the transmitted light enters the object from the first surface, and the condenser lens is moved along a Z direction intersecting the first plane and the second plane. moving in the direction and using the transmitted light to photograph the aforementioned object to obtain a detection image including the image of the aforementioned modified region and/or the aforementioned crack as an internal image, After the aforementioned photographing process, the calculation process multiplies the correction coefficient by the moving amount of the condenser lens when the detection image is captured, to calculate the position of the modified region and/or the crack in the Z direction. value, The control unit stores a plurality of correction coefficients corresponding to the movement amount. 一種觀察對象物,具有第1面和前述第1面的相反側的第2面並設置有標記,其特徵在於: 前述觀察對象物用於匯出校正係數,該匯出校正係數用於從前述標記的位置的實測值運算與前述第1面和前述第2面交叉的Z方向上的前述標記的位置的測算值。 An observation object having a first surface and a second surface opposite to the first surface and provided with a mark, characterized in that: The observation object is used to export a correction coefficient for calculating a measured value of the position of the mark in the Z direction intersecting the first surface and the second surface from the actual measured value of the position of the mark. .
TW111102335A 2021-01-29 2022-01-20 Observation device, observation method, and object to be observed TW202302263A (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2021-013546 2021-01-29
JP2021013546A JP2022117064A (en) 2021-01-29 2021-01-29 Observation device, method for observation, and observation object

Publications (1)

Publication Number Publication Date
TW202302263A true TW202302263A (en) 2023-01-16

Family

ID=82749996

Family Applications (1)

Application Number Title Priority Date Filing Date
TW111102335A TW202302263A (en) 2021-01-29 2022-01-20 Observation device, observation method, and object to be observed

Country Status (4)

Country Link
JP (1) JP2022117064A (en)
KR (1) KR20220110084A (en)
CN (1) CN114905170A (en)
TW (1) TW202302263A (en)

Also Published As

Publication number Publication date
CN114905170A (en) 2022-08-16
KR20220110084A (en) 2022-08-05
JP2022117064A (en) 2022-08-10

Similar Documents

Publication Publication Date Title
US7127098B2 (en) Image detection method and its apparatus and defect detection method and its apparatus
JP5293782B2 (en) In-focus position adjusting method, in-focus position adjusting apparatus, and laser processing apparatus
US20230154774A1 (en) Inspection device and inspection method
US20230095941A1 (en) Laser machining device and laser machining method
JP2014137229A (en) Inspection system and defect inspection method
JP5508734B2 (en) Pattern drawing apparatus and pattern drawing method
TW202302263A (en) Observation device, observation method, and object to be observed
US12023761B2 (en) Inspection device and inspection method
JP7385504B2 (en) Inspection equipment and processing system
TW202235195A (en) Observation device and observation method performing the moving of the condensing position at high speed even if imaging region is larger
TW202235194A (en) Observation device and observation method capable of more accurately acquiring information on the position of the modified region
TWI681165B (en) Line width measuring system and line width measuring device
KR20230086588A (en) Inspection device and inspection method
US20220331909A1 (en) Inspection device and inspection method
JP7493967B2 (en) Inspection device and inspection method
CN116246968A (en) Inspection method
JP3994223B2 (en) Overlay measuring device and overlay measuring method
CN114054985A (en) Inspection apparatus and inspection method
TW202145397A (en) Inspection device and inspection method
JP2009043960A (en) Automatic defect image collection method
JP2005025060A (en) Method for adjusting microscope device
JP2008281719A (en) Optical microscope, focusing position adjusting method, and manufacturing method for pattern substrate