WO2022138097A1 - 固体撮像装置およびその製造方法 - Google Patents
固体撮像装置およびその製造方法 Download PDFInfo
- Publication number
- WO2022138097A1 WO2022138097A1 PCT/JP2021/044625 JP2021044625W WO2022138097A1 WO 2022138097 A1 WO2022138097 A1 WO 2022138097A1 JP 2021044625 W JP2021044625 W JP 2021044625W WO 2022138097 A1 WO2022138097 A1 WO 2022138097A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- semiconductor substrate
- solid
- pixel separation
- substrate
- state image
- Prior art date
Links
- 238000003384 imaging method Methods 0.000 title claims abstract description 41
- 238000000034 method Methods 0.000 title claims abstract description 21
- 238000004519 manufacturing process Methods 0.000 title claims abstract description 15
- 239000000758 substrate Substances 0.000 claims abstract description 360
- 238000000926 separation method Methods 0.000 claims abstract description 147
- 238000006243 chemical reaction Methods 0.000 claims abstract description 64
- 239000004065 semiconductor Substances 0.000 claims description 273
- 230000000149 penetrating effect Effects 0.000 claims description 6
- QVGXLLKOCUKJST-UHFFFAOYSA-N atomic oxygen Chemical compound [O] QVGXLLKOCUKJST-UHFFFAOYSA-N 0.000 claims description 4
- 239000001301 oxygen Substances 0.000 claims description 4
- 229910052760 oxygen Inorganic materials 0.000 claims description 4
- 239000007787 solid Substances 0.000 claims description 2
- 239000010408 film Substances 0.000 description 169
- 239000010410 layer Substances 0.000 description 71
- 238000009792 diffusion process Methods 0.000 description 34
- 238000012545 processing Methods 0.000 description 22
- 238000001514 detection method Methods 0.000 description 21
- 238000004891 communication Methods 0.000 description 17
- 230000006870 function Effects 0.000 description 16
- XUIMIQQOPSSXEZ-UHFFFAOYSA-N Silicon Chemical compound [Si] XUIMIQQOPSSXEZ-UHFFFAOYSA-N 0.000 description 12
- 239000011229 interlayer Substances 0.000 description 12
- 229910052710 silicon Inorganic materials 0.000 description 12
- 239000010703 silicon Substances 0.000 description 12
- 230000003287 optical effect Effects 0.000 description 11
- 238000010586 diagram Methods 0.000 description 9
- 238000012986 modification Methods 0.000 description 8
- 230000004048 modification Effects 0.000 description 8
- 210000001519 tissue Anatomy 0.000 description 8
- 238000002674 endoscopic surgery Methods 0.000 description 7
- VYPSYNLAJGMNEJ-UHFFFAOYSA-N Silicium dioxide Chemical compound O=[Si]=O VYPSYNLAJGMNEJ-UHFFFAOYSA-N 0.000 description 6
- 230000005540 biological transmission Effects 0.000 description 6
- 239000013078 crystal Substances 0.000 description 6
- 229910052814 silicon oxide Inorganic materials 0.000 description 6
- 235000012431 wafers Nutrition 0.000 description 6
- 230000000052 comparative effect Effects 0.000 description 5
- 229910052751 metal Inorganic materials 0.000 description 5
- 239000002184 metal Substances 0.000 description 5
- 238000001356 surgical procedure Methods 0.000 description 5
- 230000008859 change Effects 0.000 description 4
- 239000010949 copper Substances 0.000 description 4
- 230000005284 excitation Effects 0.000 description 4
- 230000007246 mechanism Effects 0.000 description 4
- 230000008569 process Effects 0.000 description 4
- 230000003321 amplification Effects 0.000 description 3
- 238000010336 energy treatment Methods 0.000 description 3
- 238000003199 nucleic acid amplification method Methods 0.000 description 3
- 230000003647 oxidation Effects 0.000 description 3
- 238000007254 oxidation reaction Methods 0.000 description 3
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 2
- 208000005646 Pneumoperitoneum Diseases 0.000 description 2
- 229910052782 aluminium Inorganic materials 0.000 description 2
- XAGFODPZIPBFFR-UHFFFAOYSA-N aluminium Chemical compound [Al] XAGFODPZIPBFFR-UHFFFAOYSA-N 0.000 description 2
- 230000000903 blocking effect Effects 0.000 description 2
- 210000004204 blood vessel Anatomy 0.000 description 2
- 239000003153 chemical reaction reagent Substances 0.000 description 2
- 229910052802 copper Inorganic materials 0.000 description 2
- MOFVSTNWEDAEEK-UHFFFAOYSA-M indocyanine green Chemical compound [Na+].[O-]S(=O)(=O)CCCCN1C2=CC=C3C=CC=CC3=C2C(C)(C)C1=CC=CC=CC=CC1=[N+](CCCCS([O-])(=O)=O)C2=CC=C(C=CC=C3)C3=C2C1(C)C MOFVSTNWEDAEEK-UHFFFAOYSA-M 0.000 description 2
- 229960004657 indocyanine green Drugs 0.000 description 2
- 238000009413 insulation Methods 0.000 description 2
- 230000001678 irradiating effect Effects 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 230000001590 oxidative effect Effects 0.000 description 2
- 230000002194 synthesizing effect Effects 0.000 description 2
- 238000012546 transfer Methods 0.000 description 2
- WFKWXMTUELFFGS-UHFFFAOYSA-N tungsten Chemical compound [W] WFKWXMTUELFFGS-UHFFFAOYSA-N 0.000 description 2
- 229910052721 tungsten Inorganic materials 0.000 description 2
- 239000010937 tungsten Substances 0.000 description 2
- 240000004050 Pentaglottis sempervirens Species 0.000 description 1
- 235000004522 Pentaglottis sempervirens Nutrition 0.000 description 1
- 230000003187 abdominal effect Effects 0.000 description 1
- 230000001133 acceleration Effects 0.000 description 1
- 230000000740 bleeding effect Effects 0.000 description 1
- 238000002485 combustion reaction Methods 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 230000006866 deterioration Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000001312 dry etching Methods 0.000 description 1
- 238000002073 fluorescence micrograph Methods 0.000 description 1
- 239000007789 gas Substances 0.000 description 1
- 230000005484 gravity Effects 0.000 description 1
- 230000031700 light absorption Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000012423 maintenance Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 239000003595 mist Substances 0.000 description 1
- 230000000116 mitigating effect Effects 0.000 description 1
- 210000004400 mucous membrane Anatomy 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 239000011347 resin Substances 0.000 description 1
- 229920005989 resin Polymers 0.000 description 1
- 238000007789 sealing Methods 0.000 description 1
- 239000002344 surface layer Substances 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 239000010409 thin film Substances 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/1463—Pixel isolation structures
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/14609—Pixel-elements with integrated switching, control, storage or amplification elements
- H01L27/14612—Pixel-elements with integrated switching, control, storage or amplification elements involving a transistor
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/14609—Pixel-elements with integrated switching, control, storage or amplification elements
- H01L27/14612—Pixel-elements with integrated switching, control, storage or amplification elements involving a transistor
- H01L27/14616—Pixel-elements with integrated switching, control, storage or amplification elements involving a transistor characterised by the channel of the transistor, e.g. channel having a doping gradient
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/1462—Coatings
- H01L27/14623—Optical shielding
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/14634—Assemblies, i.e. Hybrid structures
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/14638—Structures specially adapted for transferring the charges across the imager perpendicular to the imaging plane
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14683—Processes or apparatus peculiar to the manufacture or treatment of these devices or parts thereof
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/1464—Back illuminated imager structures
Definitions
- This disclosure relates to a solid-state image sensor and a method for manufacturing the same.
- a pixel separation groove that surrounds these photoelectric conversion units in a ring shape may be provided in the substrate for each photoelectric conversion unit.
- an insulating film such as an oxide film and a light-shielding film such as a metal film are often embedded in order as a pixel separation portion.
- the ratio of the size of the insulating film to the size of the photoelectric conversion unit becomes large, the size of the photoelectric conversion unit is too small, and the size of the pixel separation groove is too large. Is a problem. For example, if the size of the photoelectric conversion unit is too small, the performance such as the dark current characteristics of the photoelectric conversion unit deteriorates.
- the present disclosure provides a solid-state image sensor capable of preferably forming a pixel separation portion in the pixel separation groove, and a method for manufacturing the same.
- the solid-state image sensor on the first aspect of the present disclosure includes a first substrate including a first semiconductor substrate, a plurality of photoelectric conversion units provided in the first semiconductor substrate, and the photoelectric in the first semiconductor substrate.
- a pixel separation unit provided between the conversion units is provided, and the interface between the side surface of the pixel separation unit and the first semiconductor substrate has a ⁇ 100 ⁇ surface.
- the pixel separating portion may include an insulating film. This makes it possible to form, for example, a thin insulating film for the pixel separation portion, and as a result, the size of the pixel separation portion can be reduced.
- the pixel separation portion may further include a light-shielding film. This makes it possible to form a thick light-shielding film for the pixel separation portion, for example, by forming a thin insulating film for the pixel separation portion.
- the insulating film may contain an element contained in the first semiconductor substrate and oxygen. This makes it possible to form an insulating film by, for example, oxidizing the side surface of the first semiconductor substrate.
- the insulating film is provided at the first portion having the first film thickness in a plan view and the corner portion of the pixel separation portion, and is thicker than the first film thickness. It may include a second portion having a film thickness of 2. This makes it possible, for example, to limit the thick portion of the insulating film to the corner portion of the pixel separation portion and reduce the overall film thickness of the insulating film.
- the pixel separation portion is parallel to a plurality of first portions extending in a first direction parallel to the surface of the first semiconductor substrate in a plan view and parallel to the surface of the first semiconductor substrate. It may include a plurality of second portions extending in the second direction. This makes it possible to realize, for example, a pixel separation portion having a mesh-like planar shape.
- the plan view may correspond to a state in which the light incident surface of the first semiconductor substrate is viewed.
- the thick portion of the insulating film is limited to the corner portion of the pixel separation portion, and the overall film thickness of the insulating film can be reduced. It becomes.
- the first or second direction may be parallel to the ⁇ 100> direction of the first semiconductor substrate.
- the side surface of the first semiconductor substrate parallel to the first or second direction, it is possible to make the side surface of the first semiconductor substrate a ⁇ 100 ⁇ surface.
- the pixel separation portion may be provided in the pixel separation groove penetrating the first semiconductor substrate. This makes it possible to suitably form a pixel separation portion in the pixel separation groove penetrating the first semiconductor substrate, for example.
- the pixel separation portion may be provided in a pixel separation groove that does not penetrate the first semiconductor substrate. This makes it possible to suitably form a pixel separation portion in the pixel separation groove that does not penetrate the first semiconductor substrate, for example.
- the solid-state image sensor on the first side surface is provided with a first insulating layer provided on the side opposite to the light incident surface of the first substrate and a second insulating layer provided so as to face the first insulating layer.
- a second substrate including a semiconductor substrate is further provided, and the second substrate may include a transistor. This makes it possible to use, for example, a second semiconductor substrate suitable for a transistor while using a first semiconductor substrate suitable for a pixel separation portion.
- the pixel separation portion is parallel to a plurality of first portions extending in a first direction parallel to the surface of the first semiconductor substrate in a plan view and parallel to the surface of the first semiconductor substrate. It may include a plurality of second portions extending in the second direction. This makes it possible to realize, for example, a pixel separation portion having a mesh-like planar shape.
- the first or second direction is parallel to the ⁇ 110> direction of the second semiconductor substrate, and the transistor is an n-type having a channel direction parallel to the ⁇ 110> direction. It may be a flat transistor. This makes it possible to use, for example, a second semiconductor substrate suitable for an n-type planar transistor.
- the first or second direction is parallel to the ⁇ 100> direction of the second semiconductor substrate
- the transistor is a fin which is a ⁇ 100 ⁇ plane of the second semiconductor substrate.
- a fin-type transistor having a side wall and having a channel direction parallel to the first or second direction may be used. This makes it possible to suitably form a fin-type transistor in a second substrate whose first or second direction is parallel to the ⁇ 100> direction, for example.
- the first or second direction is parallel to the ⁇ 100> direction of the second semiconductor substrate, and the transistor is a p-type having a channel direction parallel to the ⁇ 100> direction. It may be a flat transistor. This makes it possible to use, for example, a second semiconductor substrate suitable for a p-type planar transistor.
- the first or second direction is parallel to the ⁇ 110> direction of the second semiconductor substrate
- the transistor is a fin which is a ⁇ 100 ⁇ plane of the second semiconductor substrate.
- a fin-type transistor having a side wall and having channel directions non-parallel to the first and second directions may be used. This makes it possible to suitably form a fin-type transistor in a second semiconductor substrate whose first or second direction is parallel to the ⁇ 110> direction, for example.
- the solid-state image sensor on the second side of the present disclosure includes a first substrate including a first semiconductor substrate, a plurality of photoelectric conversion units provided in the first semiconductor substrate, and the photoelectric in the first semiconductor substrate.
- a pixel separation unit provided between the conversion units is provided, the pixel separation unit includes an insulating film, and the insulating film has a first portion having a first film thickness in a plan view and a pixel separation unit. It includes a second portion provided at a corner portion and having a second film thickness thicker than the first film thickness.
- the size of the pixel separation portion can be reduced, and the pixel separation portion can be suitably formed in the pixel separation groove.
- the thick portion of the insulating film for the pixel separation portion can be limited to the corner portion of the pixel separation portion, and the overall film thickness of the insulating film for the inside of the pixel separation portion can be reduced.
- a plurality of photoelectric conversion units are formed in a first semiconductor substrate of a first substrate, and pixels are separated between the photoelectric conversion units in the first semiconductor substrate.
- the pixel separation portion includes forming a portion, and the pixel separation portion is formed so that the interface between the side surface of the pixel separation portion and the first semiconductor substrate has a ⁇ 100 ⁇ plane.
- the size of the pixel separation portion can be reduced, and the pixel separation portion can be suitably formed in the pixel separation groove.
- the pixel separation portion may be formed so as to include an insulating film.
- an insulating film for example, a thin insulating film can be formed on the side surface of the first substrate, and as a result, the size of the pixel separation portion can be reduced.
- the insulating film is provided at the first portion having the first film thickness in a plan view and the corner portion of the pixel separation portion, and is thicker than the first film thickness. It may be formed so as to include a second portion having a film thickness of 2. This makes it possible, for example, to limit the thick portion of the insulating film to the corner portion of the pixel separation portion and reduce the overall film thickness of the insulating film.
- FIG. 1 is a block diagram showing a configuration of a solid-state image sensor according to the first embodiment.
- the solid-state image sensor of FIG. 1 is a CMOS (Complementary Metal Oxide Semiconductor) type image sensor, which includes a pixel array region 2 having a plurality of pixels 1, a control circuit 3, a vertical drive circuit 4, and a plurality of column signal processes. It includes a circuit 5, a horizontal drive circuit 6, an output circuit 7, a plurality of vertical signal lines 8, and a horizontal signal line 9.
- CMOS Complementary Metal Oxide Semiconductor
- Each pixel 1 includes a photodiode that functions as a photoelectric conversion unit and a MOS transistor that functions as a pixel transistor.
- Examples of pixel transistors are transfer transistors, reset transistors, amplification transistors, selection transistors, and the like. These pixel transistors may be shared by some pixels 1.
- the pixel array area 2 has a plurality of pixels 1 arranged in a two-dimensional array.
- the pixel array region 2 is an effective pixel region that receives light and performs photoelectric conversion to amplify and output the signal charge generated by the photoelectric conversion, and a black reference pixel that outputs optical black as a reference for the black level. Includes areas and.
- the black reference pixel region is arranged on the outer peripheral portion of the effective pixel region.
- the control circuit 3 generates various signals that serve as reference for the operation of the vertical drive circuit 4, the column signal processing circuit 5, the horizontal drive circuit 6, etc., based on the vertical sync signal, the horizontal sync signal, the master clock, and the like.
- the signal generated by the control circuit 3 is, for example, a clock signal or a control signal, and is input to the vertical drive circuit 4, the column signal processing circuit 5, the horizontal drive circuit 6, and the like.
- the vertical drive circuit 4 includes, for example, a shift register, and scans each pixel 1 in the pixel array area 2 in a row unit in the vertical direction.
- the vertical drive circuit 4 further supplies a pixel signal based on the signal charge generated by each pixel 1 to the column signal processing circuit 5 through the vertical signal line 8.
- the column signal processing circuit 5 is arranged, for example, for each column of the pixel 1 in the pixel array area 2, and the signal processing of the signal output from the pixel 1 for one row is based on the signal from the black reference pixel area. Do it for each row. Examples of this signal processing are denoising and signal amplification.
- the horizontal drive circuit 6 includes, for example, a shift register, and supplies pixel signals from each column signal processing circuit 5 to the horizontal signal line 9.
- the output circuit 7 performs signal processing on the signal supplied from each column signal processing circuit 5 through the horizontal signal line 9, and outputs the signal to which this signal processing has been performed.
- FIG. 2 is a cross-sectional view and a plan view showing the structure of the solid-state image sensor of the first embodiment.
- the solid-state image sensor of the present embodiment includes a semiconductor substrate 11, a photoelectric conversion unit 12, an n-type semiconductor region 13, a p-type semiconductor region 14, a pixel separation groove 21, and pixels. It includes a separation portion 22, an insulating film 23, a light-shielding film 24, a light-shielding film 25, a flattening film 26, a color filter 27, an on-chip lens 28, a substrate 31, and an insulating layer 32.
- the semiconductor substrate 11 is an example of the first semiconductor substrate of the present disclosure.
- the solid-state image sensor of the present embodiment further includes a substrate 11'including a semiconductor substrate 11 and an insulating film 23. The substrate 11'is an example of the first substrate of the present disclosure.
- a in FIG. 2 shows an X-axis, a Y-axis, and a Z-axis that are perpendicular to each other.
- the X and Y directions correspond to the horizontal direction (horizontal direction), and the Z direction corresponds to the vertical direction (vertical direction). Further, the + Z direction corresponds to the upward direction, and the ⁇ Z direction corresponds to the downward direction.
- the ⁇ Z direction may or may not exactly coincide with the direction of gravity.
- One of the X and Y directions is an example of the first direction of the present disclosure, and the other of the X and Y directions is an example of the second direction of the present disclosure.
- FIG. 2B is a plan view showing the structure of the substrate (wafer) 11 before dicing.
- FIG. 2C is a cross-sectional view showing the structure of the pixel separation groove 21 and the pixel separation portion 22.
- the semiconductor substrate 11 is, for example, a silicon substrate.
- the surface (lower surface) in the ⁇ Z direction of the semiconductor substrate 11 is the surface of the semiconductor substrate 11
- the surface (upper surface) in the + Z direction of the semiconductor substrate 11 is the back surface of the semiconductor substrate 11. Since the solid-state image sensor of this embodiment is a back-illuminated type, the back surface of the semiconductor substrate 11 is the light incident surface (light receiving surface) of the semiconductor substrate 11.
- the back surface of the semiconductor substrate 11 is an example of the first surface of the present disclosure
- the front surface of the semiconductor substrate 11 is an example of the second surface of the present disclosure.
- a and C in FIG. 2 show a solid-state image sensor manufactured by dicing a semiconductor substrate 11, while B in FIG. 2 shows a semiconductor substrate 11 before dicing.
- the semiconductor substrate 11 shown in FIG. 2B includes a plurality of chip regions 11a and a dicing region 11b.
- the chip region 11a has a square or rectangular planar shape.
- the dicing region 11b has a planar shape that encloses these chip regions 11a in an annular shape for each chip region 11a.
- the semiconductor substrate 11 is divided into these chip regions 11a, and one solid-state imaging device is manufactured from each chip region 11a.
- B in FIG. 2 further shows the notch N of the semiconductor substrate 11.
- a notch N is provided on the end face of the semiconductor substrate 11 in the ⁇ Y direction.
- the four sides of each chip region 11a extend in the X or Y direction.
- the dicing region 11b has a mesh-like planar shape including a plurality of linear portions extending in the X direction and a plurality of linear portions extending in the Y direction.
- the semiconductor substrate 11 of the present embodiment has a front surface and a back surface which are ⁇ 100 ⁇ planes, and is a ⁇ 100> notch substrate (45 ° notch substrate).
- the direction from the notch of the substrate toward the center of the substrate is the ⁇ 100> direction. Therefore, in the semiconductor substrate 11 of the present embodiment, the + Y direction in the semiconductor substrate 11 is the ⁇ 100> direction.
- the arrow A shown in FIG. 2B indicates the ⁇ 110> direction in the semiconductor substrate 11. In B of FIG. 2, the inclination of the arrow A with respect to the + Y direction is 45 °.
- FIG. 3 is a plan view for explaining the structure of the solid-state image sensor of the first embodiment.
- a of FIG. 3 shows a ⁇ 100> notch substrate (45 ° notch substrate) which is the semiconductor substrate 11 of the present embodiment as in B of FIG. 2, and B of FIG. 3 is a comparison of the present embodiment.
- the ⁇ 110> notch substrate (0 ° notch substrate) which is the example semiconductor substrate 11 is shown.
- the direction from the notch of the substrate toward the center of the substrate is the ⁇ 110> direction. Therefore, in the semiconductor substrate 11 of this comparative example, the + Y direction in the semiconductor substrate 11 is the ⁇ 110> direction.
- the arrow A shown in FIG. 3B indicates the ⁇ 110> direction in the semiconductor substrate 11.
- the inclination of the arrow A with respect to the + Y direction is 0 °.
- the semiconductor substrate 11 shown in FIG. 3B also has a front surface and a back surface which are ⁇ 100 ⁇ surfaces.
- the photoelectric conversion unit 12 is provided in the semiconductor substrate 11 for each pixel 1.
- a in FIG. 2 shows one photoelectric conversion unit 12 included in one pixel 1.
- the photoelectric conversion unit 12 includes an n-type semiconductor region 13 provided in the semiconductor substrate 11 and a p-type semiconductor region 14 provided around the n-type semiconductor region 13 in the semiconductor substrate 11.
- a photodiode is realized by a pn junction between the n-type semiconductor region 13 and the p-type semiconductor region 14, and the photodiode converts light into a charge.
- the photoelectric conversion unit 12 receives light from the back surface side of the semiconductor substrate 11, generates a signal charge according to the amount of the received light, and stores the generated signal charge in the n-type semiconductor region 13.
- the pixel separation groove 21 is provided in the semiconductor substrate 11, and specifically, is provided between the photoelectric conversion units 12 of the pixels 1 adjacent to each other.
- the pixel separation groove 21 of the present embodiment penetrates the inside of the semiconductor substrate 11 from the back surface side of the semiconductor substrate 11 to the front surface side of the semiconductor substrate 11.
- the pixel separation unit 22 is provided in the pixel separation groove 21, and includes the insulating film 23 and the light-shielding film 24 in order.
- the insulating film 23 is provided on the side surface and the bottom surface of the pixel separation groove 21, and the light-shielding film 24 is provided on the side surface and the bottom surface of the pixel separation groove 21 via the insulating film 23.
- the insulating film 23 is, for example, a silicon oxide film. Since the insulating film 23 of the present embodiment is formed by oxidizing the side surface of the semiconductor substrate 11, it contains a Si (silicon) element derived from the semiconductor substrate 11 and an O (oxygen) element derived from oxidation. I'm out.
- the light-shielding film 24 is a film containing a metal element such as W (tungsten), Al (aluminum), or Cu (copper), and has a function of blocking light.
- FIG. 2 shows a cross section of the pixel separation groove 21 and the pixel separation portion 22.
- the pixel separation groove 21 includes a plurality of first linear portions 21a extending in the X direction in a plan view and a plurality of second linear portions 21b extending in the Y direction in a plan view. One of these first linear portions 21a and one of these second linear portions 21b are shown.
- the pixel separation unit 22 includes a plurality of first linear portions 22a extending in the X direction in a plan view and a plurality of second linear portions 22b extending in the Y direction in a plan view, and is shown in FIG. C indicates one of these first linear portions 22a and one of these second linear portions 22b.
- first linear portion 22a and the second linear portion 22b is an example of the first portion of the pixel separation portion of the present disclosure, and the other of the first linear portion 22a and the second linear portion 22b is the present. It is an example of the second part of the pixel separation part of the disclosure.
- the above-mentioned plan view of the present embodiment corresponds to a state in which the light incident surface of the semiconductor substrate 11 is viewed.
- C in FIG. 2 further shows a side surface S1 extending in the X direction and a side surface S2 extending in the Y direction as side surfaces of the semiconductor substrate 11 in the pixel separation groove 21.
- C in FIG. 2 further shows a corner portion C between the side surface S1 and the side surface S2 as a corner portion of the semiconductor substrate 11 in the pixel separation groove 21.
- the corner portion C corresponds to the corner portion of the pixel separation portion 22.
- the insulating film 23 of the present embodiment includes a first portion 23a formed on the side surface S1 and the side surface S2, and a second portion 23b formed on the corner portion C, and is a film of the second portion 23b in a plan view.
- the thickness (T2) is thicker than the film thickness (T1) of the first portion 23a.
- the corner portion C is located in the second portion 23b.
- the film thickness of the first portion 23a is an example of the first film thickness of the present disclosure
- the film thickness of the second portion 23b is an example of the second film thickness of the present disclosure.
- the present embodiment is compared with the above comparative example. Since the semiconductor substrate 11 of the above comparative example is a ⁇ 110> notch substrate, the side surface S1 and the side surface S2 are ⁇ 110 ⁇ surfaces. On the other hand, since the semiconductor substrate 11 of the present embodiment is a ⁇ 100> notch substrate, the side surface S1 and the side surface S2 are ⁇ 100 ⁇ surfaces. In general, the ⁇ 110 ⁇ surface of a silicon substrate is more easily oxidized than the ⁇ 100 ⁇ surface of a silicon substrate. Therefore, in the above comparative example, the first portion 23a becomes thicker, and as a result, the size of the photoelectric conversion unit 12 becomes smaller and the size of the pixel separation unit 22 becomes larger.
- the first portion 23a becomes thin, and as a result, the size of the photoelectric conversion unit 12 becomes large and the size of the pixel separation unit 22 becomes small. Therefore, according to the present embodiment, it is possible to suppress the deterioration of the performance of the photoelectric conversion unit 12 due to the reduction in the size of the photoelectric conversion unit 12.
- the side surface S1 and the side surface S2 of the present embodiment are ⁇ 100 ⁇ planes, and the interface between the side surface of the pixel separation portion 22 and the semiconductor substrate 11 of the present embodiment has a ⁇ 100 ⁇ plane.
- the planar shape of the corner portion C is generally not a perfect right angle but a curved shape. Therefore, a small ⁇ 110 ⁇ surface is formed in the corner portion C of the present embodiment, and the corner portion C of the present embodiment is more easily oxidized than the side surface S1 and the side surface S2. As a result, the film thickness of the second portion 23b of the present embodiment is thicker than the film thickness of the first portion 23a. According to the present embodiment, since the thick portion of the insulating film 23 can be limited to the corner portion C, the overall film thickness of the insulating film 23 can be reduced.
- the insulating film 23 of the present embodiment may be formed by, for example, radical oxidation.
- the film thickness of the first portion 23a and the film thickness of the second portion 23b can be made the same, and not only the film thickness of the first portion 23a but also the film thickness of the second portion 23b can be reduced. It will be possible.
- FIG. 4 is another cross-sectional view showing the structure of the solid-state image sensor of the first embodiment.
- FIG. 4 shows a vertical cross section of three pixels 1 in the pixel array region 2 of FIG.
- the solid-state image sensor of the present embodiment includes a plurality of photoelectric conversion units 12, and includes a pixel separation groove 21 and a pixel separation unit 22 between the photoelectric conversion units 12 adjacent to each other.
- FIG. 5 is another cross-sectional view showing the structure of the solid-state image sensor of the first embodiment.
- FIG. 4 shows a cross section of the entire four pixels 1 and a part of the twelve pixels 1 in the pixel array region 2 of FIG.
- the solid-state image sensor of the present embodiment includes a plurality of photoelectric conversion units 12, and the pixel separation unit 22 surrounds these photoelectric conversion units 12 in a ring shape for each photoelectric conversion unit 12. It has a mesh-like planar shape. Therefore, each photoelectric conversion unit 12 is provided between two first linear portions 22a adjacent to each other in the Y direction, and is provided between two second linear portions 22b adjacent to each other in the X direction. ing.
- the light-shielding film 25 is provided on the pixel separation portion 22 outside the semiconductor substrate 11.
- the light-shielding film 25 is a film containing a metal element such as W (tungsten), Al (aluminum), or Cu (copper), and has a function of blocking light.
- the light-shielding film 25 may be formed at the same time as the light-shielding film 24.
- the flattening film 26 is formed on the semiconductor substrate 11 via a light-shielding film 25 so as to cover the back surface (upper surface) of the semiconductor substrate 11, whereby the surface on the back surface of the semiconductor substrate 11 is flat. ..
- the flattening film 26 is, for example, an organic film such as a resin film.
- the color filter 27 has a function of transmitting light having a predetermined wavelength, and is formed on the flattening film 26 for each pixel 1.
- the color filters 27 for red (R), green (G), and blue (B) are arranged above the photoelectric conversion unit 12 of the red, green, and blue pixels 1, respectively.
- the color filter 27 for infrared light may be arranged above the photoelectric conversion unit 12 of the infrared light pixel 1. The light transmitted through the color filter 27 is incident on the photoelectric conversion unit 12 via the flattening film 26.
- the on-chip lens 28 has a function of condensing incident light, and is formed on the color filter 27 for each pixel 1.
- the light collected by the on-chip lens 28 is incident on the photoelectric conversion unit 12 via the color filter 27 and the flattening film 26.
- Each on-chip lens 28 of the present embodiment is made of a material that allows light to pass through, and the on-chip lenses 27 are connected to each other via this material.
- the substrate 31 is provided on the surface (lower surface) of the semiconductor substrate 11 via the insulating layer 32, and is provided, for example, in order to secure the strength of the semiconductor substrate 11.
- the substrate 31 is, for example, a semiconductor substrate such as a silicon substrate.
- the substrate 31 of the present embodiment has a front surface and a back surface which are ⁇ 100 ⁇ planes, and is a ⁇ 110> notch substrate (0 ° notch substrate).
- the insulating layer 32 is, for example, a laminated film including a silicon oxide film and another insulating film.
- the light incident on the on-chip lens 28 is condensed by the on-chip lens 28, transmitted through the color filter 27, and incident on the photoelectric conversion unit 12.
- the photoelectric conversion unit 12 converts this light into an electric charge by photoelectric conversion to generate a signal charge.
- the signal charge is output as a pixel signal via the vertical signal line 8 of FIG.
- 6 to 8 are cross-sectional views showing a method of manufacturing the solid-state image sensor of the first embodiment.
- the n-type semiconductor region 13 and the p-type semiconductor region 14 of each photoelectric conversion unit 12 are formed in the semiconductor substrate 11, and the insulating layer 32 is formed on the semiconductor substrate 11 (A in FIG. 6). In this way, a plurality of photoelectric conversion units 12 are formed in the semiconductor substrate 11.
- the step shown in FIG. 6A is performed with the front surface of the semiconductor substrate 11 facing up and the back surface of the semiconductor substrate 11 facing down.
- the semiconductor substrate 11 is turned upside down (B in FIG. 6). As a result, the front surface of the semiconductor substrate 11 faces downward, and the back surface of the semiconductor substrate 11 faces upward. Next, the semiconductor substrate 11 is adhered to the surface (upper surface) of the substrate 31 via the insulating layer 32 (B in FIG. 6).
- the pixel separation groove 21 is formed in the semiconductor substrate 11 by dry etching (A in FIG. 7).
- the pixel separation groove 21 of the present embodiment is formed so as to penetrate the semiconductor substrate 11 and reach the insulating layer 32. Further, the pixel separation groove 21 of the present embodiment is formed so as to have a mesh-like planar shape that surrounds the plurality of photoelectric conversion units 12 in a ring shape for each photoelectric conversion unit 12, and the photoelectric conversion units 12 are adjacent to each other. Formed between.
- the insulating film 23 and the light-shielding film 24 are sequentially formed in the pixel separation groove 21 (B in FIG. 7).
- the pixel separation portion 22 including the insulating film 23 and the light-shielding film 24 is formed in the pixel separation groove 21.
- the insulating film 23 is formed on the side surface and the bottom surface of the pixel separation groove 21, and the light-shielding film 24 is formed on the side surface and the bottom surface of the pixel separation groove 21 via the insulating film 23.
- the semiconductor substrate 11 of this embodiment is a ⁇ 100> notch substrate
- the side surface of the semiconductor substrate 11 in the pixel separation groove 21 is a ⁇ 100 ⁇ surface. Therefore, according to the present embodiment, the insulating film 23 is formed on the side surface of the semiconductor substrate 11 in the pixel separation groove 21 by oxidation, so that the first portion 23a having a thin film thickness and the second portion 23a having a thick film thickness are formed.
- the insulating film 23 including the portion 23b can be formed (see C in FIG. 2).
- the light-shielding film 25 and the flattening film 26 are sequentially formed on the semiconductor substrate 11 (A in FIG. 8).
- the light-shielding film 25 is formed on the pixel separation portion 22, and the flattening film 26 is formed on the semiconductor substrate 11 so as to cover the light-shielding film 25.
- the color filter 27 and the on-chip lens 28 are sequentially formed on the flattening film 26 above each photoelectric conversion unit 12 (B in FIG. 8). Then, by cutting the semiconductor substrate 11 at the dicing region 11b, the semiconductor substrate 11 is divided into individual chip regions 11a (see B in FIG. 2). In this way, the solid-state image sensor of the present embodiment is manufactured.
- the pixel separation portion 22 of the present embodiment is formed by forming the insulating film 23 on the side surface of the semiconductor substrate 11 which is the ⁇ 100 ⁇ plane. Therefore, according to the present embodiment, the pixel separation portion 22 can be suitably formed in the pixel separation groove 21, for example, the size of the pixel separation portion 22 can be reduced by thinning the insulating film 23. It will be possible.
- the semiconductor substrate 11 of the present embodiment is a Si ⁇ 100 ⁇ substrate whose front surface, back surface, or ⁇ 100 ⁇ surface thereof, and is a ⁇ 110> notch substrate in which the + Y direction is the ⁇ 110> direction.
- the meanings of the above-mentioned reference numerals ⁇ xyz ⁇ and reference numeral ⁇ xyz> will be supplemented by taking the Si ⁇ 111 ⁇ substrate and the ⁇ 110> direction as an example.
- the Si ⁇ 111 ⁇ substrate in the present disclosure is a substrate or wafer made of a silicon single crystal and having a crystal plane represented by ⁇ 111 ⁇ in the Miller index notation.
- the Si ⁇ 111 ⁇ substrate in the present disclosure also includes a substrate or wafer whose crystal orientation is deviated by several degrees, for example, a substrate or wafer deviated by several degrees from the ⁇ 111 ⁇ plane in the nearest [110] direction. Further, it also includes a silicon single crystal grown on a part or the entire surface of these substrates or wafers by an epitaxial method or the like.
- the ⁇ 111 ⁇ plane is a crystal plane equivalent to each other in symmetry, which is a (111) plane, a (-111) plane, a (1-11) plane, a (11-1) plane, and a (-) plane. It is a general term for the 1-11) plane, the (-11-1) plane, the (1-1-1) plane, and the (1-1-1) plane. Therefore, the description of the Si ⁇ 111 ⁇ substrate in the specification of the present disclosure may be read as, for example, a Si (1-11) substrate.
- the bar sign for expressing the negative index of the Miller index is substituted with a negative sign.
- the ⁇ 110> direction in the description of the present disclosure is the [110] direction, the [101] direction, the [011] direction, the [-110] direction, and [1-10], which are crystal plane directions equivalent to each other in symmetry.
- FIG. 9 is a cross-sectional view and a plan view showing the structure of the solid-state image sensor of the second embodiment.
- FIG. 9 shows a vertical cross section of one pixel 1 in the pixel array region 2 in FIG. 1, similar to A in FIG.
- FIG. 9B is a plan view showing the structure of the substrate (wafer) 11 before dicing, similar to FIG. 2B.
- FIG. 9C is a cross-sectional view showing the structure of the pixel separation groove 21 and the pixel separation portion 22 as in FIG. 2C.
- the solid-state image sensor of the present embodiment has the same components as the solid-state image sensor of the first embodiment.
- the pixel separation groove 21 of the present embodiment is provided on the back surface (upper surface) side of the semiconductor substrate 11 so as not to penetrate the semiconductor substrate 11.
- the structure of the present embodiment can be adopted, for example, when the pixel separation groove 21 does not need to penetrate the semiconductor substrate 11 or when it is desirable that the pixel separation groove 21 does not penetrate the semiconductor substrate 11.
- the solid-state image sensor of this embodiment is manufactured, for example, by forming a pixel separation groove 21 that does not penetrate the semiconductor substrate 11 in the step shown in FIG. 7A.
- the pixel separation groove 21 of the present embodiment may include both a portion penetrating the semiconductor substrate 11 and a portion not penetrating the semiconductor substrate 11.
- FIG. 10 is a cross-sectional view and a plan view showing the structure of the solid-state image sensor of the third embodiment.
- a in FIG. 10 shows a vertical cross section of one pixel 1 in the pixel array region 2 in FIG. 1, similar to A in FIG.
- the semiconductor substrate 33, the insulating layer 34, the gate electrode 35 of the transistor Tr1, the gate electrode 36 of the transistor Tr2, and the plug It includes a 41, an insulating film 42, a plug 43, and a wiring layer 44.
- the insulating layer 32 includes an insulating film 32a that functions as a gate insulating film of the transistor Tr1 and an interlayer insulating film 32b, and the insulating layer 34 is an interlayer insulating film 34a that functions as a gate insulating film of the transistor Tr2. It contains a film 34b and the like.
- the insulating layer 32 is an example of the first insulating layer of the present disclosure
- the semiconductor substrate 33 is an example of the second semiconductor substrate of the present disclosure.
- the solid-state imaging device of the present embodiment further includes a substrate 33'including a semiconductor substrate 33, an insulating layer 34, a gate electrode 36, a plug 41, an insulating film 42, a plug 43, and a wiring layer 44.
- the substrate 33' is an example of the second substrate of the present disclosure.
- FIG. 10B is a plan view showing the structure of the semiconductor substrate 33 and the gate electrode 36.
- the insulating layer 32 includes an insulating film 32a and an interlayer insulating film 32b which are sequentially provided on the surface (lower surface) of the semiconductor substrate 11.
- the insulating film 32a is, for example, a silicon oxide film.
- the interlayer insulating film 32b is, for example, a laminated film including a silicon oxide film and another insulating film.
- the semiconductor substrate 33 is provided on the lower surface of the insulating layer 32.
- the semiconductor substrate 33 is, for example, a silicon substrate.
- the insulating layer 34 includes an insulating film 34a and an interlayer insulating film 34b which are sequentially provided on the surface (lower surface) of the semiconductor substrate 33.
- the insulating film 34a is, for example, a silicon oxide film.
- the interlayer insulating film 34b is, for example, a laminated film including a silicon oxide film and another insulating film.
- the substrate 31 is provided on the lower surface of the insulating layer 34.
- the solid-state image sensor of the present embodiment includes the semiconductor substrate 33 in addition to the semiconductor substrate 11 and the substrate 31. Similar to the semiconductor substrate 11 of the first embodiment, the semiconductor substrate 11 of the present embodiment has a front surface and a back surface which are ⁇ 100 ⁇ planes, and has a ⁇ 100> notch substrate (45 ° notch substrate). It has become. On the other hand, the semiconductor substrate 33 of the present embodiment has a front surface and a back surface which are ⁇ 100 ⁇ planes, and is a ⁇ 110> notch substrate (0 ° notch substrate). Therefore, in the semiconductor substrate 33 of the present embodiment, the + Y direction in the semiconductor substrate 33 is the ⁇ 110> direction, as in the semiconductor substrate 11 of the comparative example shown in FIG. 3B.
- the gate electrode 35 of the transistor Tr1 is provided on the surface (lower surface) of the semiconductor substrate 11 via an insulating film 32a, and is covered with the interlayer insulating film 32b.
- the transistor Tr1 is, for example, a pixel transistor such as a transfer transistor.
- the gate electrode 35 is, for example, a semiconductor layer or a metal layer.
- the transistor Tr1 further includes a source diffusion layer and a drain diffusion layer (not shown) provided in the substrate 31.
- the gate electrode 36 of the transistor Tr2 is provided on the surface (lower surface) of the semiconductor substrate 33 via an insulating film 34a, and is covered with the interlayer insulating film 34b.
- the transistor Tr2 is, for example, a pixel transistor such as an amplification transistor.
- the gate electrode 36 is, for example, a semiconductor layer or a metal layer.
- the transistor Tr2 further includes a source diffusion layer 33a and a drain diffusion layer 33b provided in the semiconductor substrate 33.
- the transistor Tr2 of the present embodiment is an n-type planar transistor, and includes a source diffusion layer 33a and a drain diffusion layer 33b arranged in the X direction, and a gate electrode 36 extending in the Y direction (B in FIG. 10). Therefore, the channel direction of the transistor Tr2 of the present embodiment is the + X direction and is parallel to the ⁇ 110> direction.
- the performance of the n-type planar transistor is improved by making the channel direction parallel to the ⁇ 110> direction of the silicon substrate.
- the semiconductor substrate 33 of this embodiment is a ⁇ 110> notch substrate as described above. Therefore, according to the present embodiment, by forming the source diffusion layer 33a and the drain diffusion layer 33b arranged in the X direction in the semiconductor substrate 33, the channel direction can be made parallel to the ⁇ 110> direction. This makes it possible to improve the performance of the transistor Tr2.
- the wiring layer 44 is provided below the gate electrode 36 in the interlayer insulating film 34b.
- the plug 43 is provided in the interlayer insulating film 34b, and electrically connects the wiring layer 44 and the gate electrode 36.
- the plug 41 is provided in the insulating layer 34, the semiconductor substrate 33, and the insulating layer 32, and electrically connects the wiring layer 44 and the semiconductor substrate 11. As a result, the transistor Tr2 is electrically connected to the semiconductor substrate 11.
- the plug 41 is provided in the semiconductor substrate 33 via the insulating film 42.
- the solid-state image sensor of the present embodiment includes the semiconductor substrate 33 which is the ⁇ 110> notch substrate in addition to the semiconductor substrate 11 which is the ⁇ 100> notch substrate. Therefore, according to the present embodiment, it is possible to suitably form the n-type planar transistor (transistor Tr2) on the surface of the semiconductor substrate 33 while appropriately forming the pixel separation portion 32 in the semiconductor substrate 11.
- transistor Tr2 n-type planar transistor
- FIG. 11 is a cross-sectional view and a plan view showing the structure of the solid-state image sensor of the fourth embodiment.
- a in FIG. 11 shows a vertical cross section of one pixel 1 in the pixel array region 2 in FIG. 1, similar to A in FIG. FIG. 11B is a plan view showing the structure of the semiconductor substrate 33 and the gate electrode 36, similarly to FIG. 10B.
- the solid-state image sensor of the present embodiment has the same components as the solid-state image sensor of the third embodiment.
- the semiconductor substrate 33 of the present embodiment has a front surface and a back surface which are ⁇ 100 ⁇ planes, and is a ⁇ 100> notch substrate (45 ° notch substrate). Therefore, in the semiconductor substrate 33 of the present embodiment, the + Y direction in the semiconductor substrate 33 is the ⁇ 100> direction.
- the transistor Tr2 of the present embodiment is a fin type transistor, and the gate electrode 36 of the transistor Tr2 has a flat surface portion 36a provided outside the semiconductor substrate 33 and a plurality of fin portions 36b provided inside the semiconductor substrate 33. And include.
- the transistor Tr2 of the present embodiment includes a plurality of source diffusion layers 33a and a plurality of drain diffusion layers 33b in the semiconductor substrate 33, and the source diffusion layer 33a and the drain diffusion layer 33b. Are lined up in the X direction.
- the gate electrode 36 of the present embodiment includes a plurality of fin portions 36b in the semiconductor substrate 33, and these fin portions 36b extend in the Y direction. Therefore, the channel direction of the transistor Tr2 of the present embodiment is the + X direction, which is parallel to the ⁇ 100> direction, and the fin side wall of the transistor Tr2 of the present embodiment is a semiconductor substrate extending in the Y direction. It is the side surface of 33, and is the ⁇ 100 ⁇ side surface.
- the performance of the fin type transistor is improved by using the fin side wall as the ⁇ 100 ⁇ surface of the silicon substrate.
- the semiconductor substrate 33 of this embodiment is a ⁇ 100> notch substrate as described above. Therefore, according to the present embodiment, by forming the fin portion 36b extending in the Y direction in the semiconductor substrate 33, it is possible to make the fin side wall a ⁇ 100 ⁇ plane, thereby improving the performance of the transistor Tr2. It becomes possible.
- FIG. 12 is a cross-sectional view and a plan view showing the structure of the solid-state image sensor of the modified example of the fourth embodiment.
- a and B in FIG. 12 correspond to A and B in FIG. 11, respectively.
- the solid-state imaging device of this modification has a structure in which the semiconductor substrate 33 and the insulating layer 34 are removed from the solid-state imaging device of the fourth embodiment, and the transistor Tr2 is not the surface of the semiconductor substrate 33 but the semiconductor substrate 11. It is formed on the surface. Therefore, the gate insulating film of the transistor Tr2 is replaced with the insulating film 32a from the insulating film 34a, and the diffusion layer of the transistor Tr2 is from the source diffusion layer 33a and the drain diffusion layer 33b in the semiconductor substrate 33 to the inside of the semiconductor substrate 11. It has been replaced by the source diffusion layer 11c and the drain diffusion layer 11d. According to this modification, the performance of the transistor Tr2 can be improved by using the semiconductor substrate 11 instead of the semiconductor substrate 33.
- the solid-state image sensor of the present embodiment includes the semiconductor substrate 33 which is the ⁇ 100> notch substrate in addition to the semiconductor substrate 11 which is the ⁇ 100> notch substrate. Therefore, according to the present embodiment, it is possible to suitably form the fin type transistor (transistor Tr2) on the surface of the semiconductor substrate 33 while appropriately forming the pixel separation portion 32 in the semiconductor substrate 11.
- the fin-type transistor may be formed on the surface of the semiconductor substrate 11 as in the above modification.
- FIG. 13 is a cross-sectional view and a plan view showing the structure of the solid-state image sensor according to the fifth embodiment.
- FIG. 13 shows a vertical cross section of one pixel 1 in the pixel array region 2 in FIG. 1, similarly to A in FIG.
- FIG. 13B is a plan view showing the structure of the semiconductor substrate 33 and the gate electrode 36, similarly to FIG. 10B.
- the solid-state image sensor of the present embodiment has the same components as the solid-state image sensor of the third embodiment.
- the semiconductor substrate 33 of the present embodiment has a front surface and a back surface which are ⁇ 100 ⁇ planes, and is a ⁇ 100> notch substrate (45 ° notch substrate). Therefore, in the semiconductor substrate 33 of the present embodiment, the + Y direction in the semiconductor substrate 33 is the ⁇ 100> direction.
- the transistor Tr2 of the present embodiment is a p-type planar transistor, and includes a source diffusion layer 33a and a drain diffusion layer 33b arranged in the X direction, and a gate electrode 36 extending in the Y direction (B in FIG. 13). Therefore, the channel direction of the transistor Tr2 of the present embodiment is the + X direction and is parallel to the ⁇ 100> direction.
- the performance of the p-type planar transistor is improved by making the channel direction parallel to the ⁇ 100> direction of the silicon substrate.
- the semiconductor substrate 33 of this embodiment is a ⁇ 100> notch substrate as described above. Therefore, according to the present embodiment, by forming the source diffusion layer 33a and the drain diffusion layer 33b arranged in the X direction in the semiconductor substrate 33, the channel direction can be made parallel to the ⁇ 100> direction. This makes it possible to improve the performance of the transistor Tr2.
- FIG. 14 is a cross-sectional view and a plan view showing the structure of the solid-state image sensor of the modified example of the fifth embodiment.
- a and B in FIG. 14 correspond to A and B in FIG. 13, respectively.
- the solid-state imaging device of this modification has a structure in which the semiconductor substrate 33 and the insulating layer 34 are removed from the solid-state imaging device of the fifth embodiment, and the transistor Tr2 is not the surface of the semiconductor substrate 33 but the semiconductor substrate 11. It is formed on the surface. Therefore, the gate insulating film of the transistor Tr2 is replaced with the insulating film 32a from the insulating film 34a, and the diffusion layer of the transistor Tr2 is from the source diffusion layer 33a and the drain diffusion layer 33b in the semiconductor substrate 33 to the inside of the semiconductor substrate 11. It has been replaced by the source diffusion layer 11c and the drain diffusion layer 11d. According to this modification, the performance of the transistor Tr2 can be improved by using the semiconductor substrate 11 instead of the semiconductor substrate 33.
- the solid-state image sensor of the present embodiment includes the semiconductor substrate 33 which is the ⁇ 100> notch substrate in addition to the semiconductor substrate 11 which is the ⁇ 100> notch substrate. Therefore, according to the present embodiment, it is possible to suitably form the p-type planar transistor (transistor Tr2) on the surface of the semiconductor substrate 33 while appropriately forming the pixel separation portion 32 in the semiconductor substrate 11.
- the p-type planar transistor may be formed on the surface of the semiconductor substrate 11 as in the above modification.
- FIG. 15 is a cross-sectional view and a plan view showing the structure of the solid-state image sensor of the sixth embodiment.
- a in FIG. 15 shows a vertical cross section of one pixel 1 in the pixel array region 2 in FIG. 1, similar to A in FIG. FIG. 15B is a plan view showing the structure of the semiconductor substrate 33 and the gate electrode 36, similarly to FIG. 10B.
- the solid-state image sensor of the present embodiment has the same components as the solid-state image sensor of the third embodiment.
- the semiconductor substrate 33 of the present embodiment has a front surface and a back surface which are ⁇ 100 ⁇ planes, and is a ⁇ 110> notch substrate (0 ° notch substrate). Therefore, in the semiconductor substrate 33 of the present embodiment, the + Y direction in the semiconductor substrate 33 is the ⁇ 110> direction.
- the transistor Tr2 of the present embodiment is a fin type transistor, and the gate electrode 36 of the transistor Tr2 has a flat surface portion 36a provided outside the semiconductor substrate 33 and a plurality of fin portions 36b provided inside the semiconductor substrate 33. And include.
- the transistor Tr2 of the present embodiment includes a plurality of source diffusion layers 33a and a plurality of drain diffusion layers 33b in the semiconductor substrate 33, and the source diffusion layer 33a and the drain diffusion layer 33b. Are lined up in a direction inclined by + 45 ° with respect to the + X direction.
- the gate electrode 36 of the present embodiment includes a plurality of fin portions 36b in the semiconductor substrate 33, and these fin portions 36b are tilted by + 45 ° with respect to the + Y direction. It extends in the same direction.
- the channel direction of the transistor Tr2 of the present embodiment is a direction inclined by + 45 ° with respect to the + X direction, is parallel to the ⁇ 100> direction, and is a fin side wall of the transistor Tr2 of the present embodiment. Is a side surface of the semiconductor substrate 33 extending in a direction inclined by + 45 ° with respect to the + Y direction, and is a ⁇ 100 ⁇ surface.
- the performance of the fin type transistor is improved by using the fin side wall as the ⁇ 100 ⁇ surface of the silicon substrate.
- the semiconductor substrate 33 of this embodiment is a ⁇ 110> notch substrate as described above. Therefore, according to the present embodiment, by forming the fin portion 36b extending in the above direction in the semiconductor substrate 33, the fin side wall can be made into a ⁇ 100 ⁇ plane, thereby improving the performance of the transistor Tr2. It is possible to make it.
- the planar shape of each pixel 1 of the present embodiment is a square (or rectangle) having two sides extending in the X direction and two sides extending in the Y direction.
- the length of the fin portion 36b is the maximum, and is about the length of one side of the planar shape of each pixel 1.
- the length of the fin portion 36b is the maximum, which is about ⁇ 2 times the length of one side of the planar shape of each pixel 1.
- the solid-state image sensor of the present embodiment includes the semiconductor substrate 33 which is the ⁇ 110> notch substrate in addition to the semiconductor substrate 11 which is the ⁇ 100> notch substrate. Therefore, according to the present embodiment, it is possible to suitably form the fin type transistor (transistor Tr2) on the surface of the semiconductor substrate 33 while appropriately forming the pixel separation portion 32 in the semiconductor substrate 11.
- transistor Tr2 the fin type transistor
- the channel direction of the transistor Tr2 of the present embodiment may be tilted by + ⁇ with respect to the + X direction (0 ° ⁇ ⁇ 90 °). Further, the fin portion 36b of the present embodiment may extend in a direction inclined by + ⁇ with respect to the + Y direction. The value of ⁇ may be an angle other than 45 °.
- FIG. 16 is a block diagram showing a configuration example of an electronic device.
- the electrical device shown in FIG. 16 is a camera 100.
- the camera 100 includes an optical unit 101 including a lens group and the like, an image pickup device 102 which is a solid-state image pickup device according to any one of the first to sixth embodiments, and a DSP (Digital Signal Processor) circuit 103 which is a camera signal processing circuit.
- the DSP circuit 103, the frame memory 104, the display unit 105, the recording unit 106, the operation unit 107, and the power supply unit 108 are connected to each other via the bus line 109.
- the optical unit 101 captures incident light (image light) from the subject and forms an image on the image pickup surface of the image pickup device 102.
- the image pickup apparatus 102 converts the amount of incident light imaged on the image pickup surface by the optical unit 101 into an electric signal in pixel units, and outputs the light amount as a pixel signal.
- the DSP circuit 103 performs signal processing on the pixel signal output by the image pickup device 102.
- the frame memory 104 is a memory for storing one screen of a moving image or a still image captured by the image pickup apparatus 102.
- the display unit 105 includes a panel-type display device such as a liquid crystal panel or an organic EL panel, and displays a moving image or a still image captured by the image pickup device 102.
- the recording unit 106 records a moving image or a still image captured by the image pickup apparatus 102 on a recording medium such as a hard disk or a semiconductor memory.
- the operation unit 107 issues operation commands for various functions of the camera 100 under the operation of the user.
- the power supply unit 108 appropriately supplies various power sources that serve as operating power sources for the DSP circuit 103, the frame memory 104, the display unit 105, the recording unit 106, and the operation unit 107 to these supply targets.
- the solid-state image sensor can be applied to various other products.
- the solid-state imaging device may be mounted on various moving objects such as automobiles, electric vehicles, hybrid electric vehicles, motorcycles, bicycles, personal mobility, airplanes, drones, ships, and robots.
- FIG. 17 is a block diagram showing a configuration example of a mobile control system.
- the mobile control system shown in FIG. 17 is a vehicle control system 200.
- the vehicle control system 200 includes a plurality of electronic control units connected via the communication network 201.
- the vehicle control system 200 includes a drive system control unit 210, a body system control unit 220, an outside information detection unit 230, an in-vehicle information detection unit 240, and an integrated control unit 250.
- FIG. 17 further shows a microcomputer 251, an audio image output unit 252, and an in-vehicle network I / F (Interface) 253 as components of the integrated control unit 250.
- the drive system control unit 210 controls the operation of the device related to the drive system of the vehicle according to various programs.
- the drive system control unit 210 includes a driving force generating device for generating a driving force of a vehicle such as an internal combustion engine and a driving motor, a driving force transmission mechanism for transmitting the driving force to wheels, and a steering wheel of the vehicle. It functions as a control device such as a steering mechanism that adjusts the angle and a braking device that generates braking force for the vehicle.
- the body system control unit 220 controls the operation of various devices mounted on the vehicle body according to various programs.
- the body system control unit 220 functions as a control device such as a smart key system, a keyless entry system, a power window device, and various lamps (for example, a headlamp, a back lamp, a brake lamp, a winker, and a fog lamp).
- a radio wave transmitted from a portable device that substitutes for a key or a signal of various switches may be input to the body system control unit 220.
- the body system control unit 220 receives such radio wave or signal input and controls a vehicle door lock device, a power window device, a lamp, and the like.
- the outside information detection unit 230 detects information outside the vehicle equipped with the vehicle control system 200.
- an image pickup unit 231 is connected to the vehicle outside information detection unit 230.
- the vehicle outside information detection unit 230 causes the image pickup unit 231 to capture an image of the outside of the vehicle, and receives the captured image from the image pickup unit 231.
- the vehicle outside information detection unit 230 may perform object detection processing or distance detection processing such as a person, a vehicle, an obstacle, a sign, or a character on a road surface based on the received image.
- the image pickup unit 231 is an optical sensor that receives light and outputs an electric signal according to the amount of the light received.
- the image pickup unit 231 can output an electric signal as an image or can output it as distance measurement information.
- the light received by the image pickup unit 231 may be visible light or invisible light such as infrared light.
- the image pickup unit 231 includes the solid-state image pickup device according to any one of the first to sixth embodiments.
- the in-vehicle information detection unit 240 detects information inside the vehicle equipped with the vehicle control system 200.
- a driver state detection unit 241 that detects the state of the driver is connected to the in-vehicle information detection unit 240.
- the driver state detection unit 241 includes a camera that images the driver, and the in-vehicle information detection unit 240 has a degree of fatigue or concentration of the driver based on the detection information input from the driver state detection unit 241. May be calculated, or it may be determined whether or not the driver has fallen asleep.
- This camera may include the solid-state image sensor according to any one of the first to sixth embodiments, and may be, for example, the camera 100 shown in FIG.
- the microcomputer 251 calculates a control target value of the driving force generator, the steering mechanism, or the braking device based on the information inside and outside the vehicle acquired by the information detection unit 230 outside the vehicle or the information inside the vehicle 240, and controls the drive system.
- a control command can be output to the unit 210.
- the microcomputer 251 is a coordinated control for the purpose of realizing ADAS (Advanced Driver Assistance System) functions such as vehicle collision avoidance, impact mitigation, follow-up driving based on inter-vehicle distance, vehicle speed maintenance driving, collision warning, and lane deviation warning. It can be performed.
- ADAS Advanced Driver Assistance System
- the microcomputer 251 controls the driving force generator, the steering mechanism, or the braking device based on the information around the vehicle acquired by the vehicle exterior information detection unit 230 or the vehicle interior information detection unit 240, thereby controlling the driver. It is possible to perform coordinated control for the purpose of automatic driving that runs autonomously without depending on the operation.
- the microcomputer 251 can output a control command to the body system control unit 220 based on the information outside the vehicle acquired by the vehicle outside information detection unit 230.
- the microcomputer 251 controls the headlamps according to the position of the preceding vehicle or the oncoming vehicle detected by the outside information detection unit 230, and performs cooperative control for the purpose of anti-glare such as switching the high beam to the low beam. It can be carried out.
- the audio image output unit 252 transmits an output signal of at least one of audio and image to an output device capable of visually or audibly notifying the passenger of the vehicle or the outside of the vehicle.
- an audio speaker 261, a display unit 262, and an instrument panel 263 are shown as such an output device.
- the display unit 262 may include, for example, an onboard display or a head-up display.
- FIG. 18 is a plan view showing a specific example of the set position of the image pickup unit 231 of FIG.
- the vehicle 300 shown in FIG. 18 includes image pickup units 301, 302, 303, 304, and 305 as the image pickup unit 231.
- the image pickup units 301, 302, 303, 304, and 305 are provided, for example, at positions such as the front nose, side mirrors, rear bumpers, back doors, and the upper part of the windshield in the vehicle interior of the vehicle 300.
- the image pickup unit 301 provided in the front nose mainly acquires an image in front of the vehicle 300.
- the image pickup unit 302 provided in the left side mirror and the image pickup section 303 provided in the right side mirror mainly acquire an image of the side of the vehicle 300.
- the image pickup unit 304 provided on the rear bumper or the back door mainly acquires an image of the rear of the vehicle 300.
- the image pickup unit 305 provided on the upper part of the windshield in the vehicle interior mainly acquires an image in front of the vehicle 300.
- the image pickup unit 305 is used, for example, to detect a preceding vehicle, a pedestrian, an obstacle, a traffic light, a traffic sign, a lane, or the like.
- FIG. 18 shows an example of the imaging range of the imaging units 301, 302, 303, 304 (hereinafter referred to as “imaging unit 301 to 304”).
- the imaging range 311 indicates the imaging range of the imaging unit 301 provided on the front nose.
- the image pickup range 312 indicates the image pickup range of the image pickup unit 302 provided on the left side mirror.
- the image pickup range 313 indicates the image pickup range of the image pickup unit 303 provided on the right side mirror.
- the image pickup range 314 indicates the image pickup range of the image pickup unit 304 provided on the rear bumper or the back door. For example, by superimposing the image data captured by the image pickup units 301 to 304, a bird's-eye view image of the vehicle 300 viewed from above can be obtained.
- the imaging range 311, 312, 313, 314 will be referred to as "imaging range 311 to 314".
- At least one of the image pickup units 301 to 304 may have a function of acquiring distance information.
- at least one of the image pickup units 301 to 304 may be a stereo camera including a plurality of image pickup devices, or may be an image pickup device having pixels for phase difference detection.
- the microcomputer 251 (FIG. 17) has a distance to each three-dimensional object within the imaging range 311 to 314 based on the distance information obtained from the imaging units 301 to 304, and a temporal change of this distance (vehicle 300). Relative velocity to) is calculated. Based on these calculation results, the microcomputer 251 is the closest three-dimensional object on the traveling path of the vehicle 300, and is a three-dimensional object traveling at a predetermined speed (for example, 0 km / h or more) in almost the same direction as the vehicle 300. , Can be extracted as a preceding vehicle.
- a predetermined speed for example, 0 km / h or more
- the microcomputer 251 can set an inter-vehicle distance to be secured in front of the preceding vehicle in advance, and can perform automatic brake control (including follow-up stop control), automatic acceleration control (including follow-up start control), and the like. As described above, according to this example, it is possible to perform coordinated control for the purpose of automatic driving or the like that autonomously travels without the operation of the driver.
- the microcomputer 251 classifies three-dimensional object data related to a three-dimensional object into two-wheeled vehicles, ordinary vehicles, large vehicles, pedestrians, electric poles, and other three-dimensional objects based on the distance information obtained from the imaging units 301 to 304. It can be extracted and used for automatic avoidance of obstacles. For example, the microcomputer 251 distinguishes obstacles around the vehicle 300 into obstacles that can be seen by the driver of the vehicle 300 and obstacles that are difficult to see. Then, the microcomputer 251 determines the collision risk indicating the risk of collision with each obstacle, and when the collision risk is equal to or higher than the set value and there is a possibility of collision, the microcomputer 251 is used via the audio speaker 261 or the display unit 262. By outputting an alarm to the driver and performing forced deceleration and avoidance steering via the drive system control unit 210, driving support for collision avoidance can be provided.
- At least one of the image pickup units 301 to 304 may be an infrared camera that detects infrared rays.
- the microcomputer 251 can recognize a pedestrian by determining whether or not a pedestrian is present in the captured images of the imaging units 301 to 304. Such recognition of a pedestrian is, for example, whether or not the pedestrian is a pedestrian by performing a procedure for extracting feature points in the captured images of the image pickup units 301 to 304 as an infrared camera and a pattern matching process on a series of feature points showing the outline of the object. It is performed by the procedure for determining.
- the audio image output unit 252 When the microcomputer 251 determines that a pedestrian is present in the captured images of the imaging units 301 to 304 and recognizes the pedestrian, the audio image output unit 252 has a square contour line for emphasizing the recognized pedestrian.
- the display unit 262 is controlled so as to superimpose and display. Further, the audio image output unit 252 may control the display unit 262 so as to display an icon or the like indicating a pedestrian at a desired position.
- FIG. 19 is a diagram showing an example of a schematic configuration of an endoscopic surgery system to which the technique according to the present disclosure (the present technique) can be applied.
- FIG. 19 illustrates how the surgeon (doctor) 531 is performing surgery on patient 532 on patient bed 533 using the endoscopic surgery system 400.
- the endoscopic surgery system 400 includes an endoscope 500, other surgical tools 510 such as an abdominal tube 511 and an energy treatment tool 512, and a support arm device 520 that supports the endoscope 500.
- a cart 600 equipped with various devices for endoscopic surgery.
- the endoscope 500 is composed of a lens barrel 501 in which a region having a predetermined length from the tip is inserted into the body cavity of the patient 532, and a camera head 502 connected to the base end of the lens barrel 501.
- the endoscope 500 configured as a so-called rigid mirror having a rigid lens barrel 501 is shown, but the endoscope 500 may be configured as a so-called flexible mirror having a flexible lens barrel. good.
- An opening in which an objective lens is fitted is provided at the tip of the lens barrel 501.
- a light source device 603 is connected to the endoscope 500, and the light generated by the light source device 603 is guided to the tip of the lens barrel by a light guide extending inside the lens barrel 501, and is an objective. It is irradiated toward the observation target in the body cavity of the patient 532 through the lens.
- the endoscope 500 may be a direct endoscope, a perspective mirror, or a side endoscope.
- An optical system and an image pickup element are provided inside the camera head 502, and the reflected light (observation light) from the observation target is focused on the image pickup element by the optical system.
- the observation light is photoelectrically converted by the image pickup device, and an electric signal corresponding to the observation light, that is, an image signal corresponding to the observation image is generated.
- the image signal is transmitted as RAW data to the camera control unit (CCU: Camera Control Unit) 601.
- the CCU 601 is composed of a CPU (Central Processing Unit), a GPU (Graphics Processing Unit), and the like, and comprehensively controls the operations of the endoscope 500 and the display device 602. Further, the CCU 601 receives an image signal from the camera head 502, and performs various image processing on the image signal for displaying an image based on the image signal, such as a development process (demosaic process).
- a development process demosaic process
- the display device 602 displays an image based on the image signal processed by the CCU 601 under the control of the CCU 601.
- the light source device 603 is composed of, for example, a light source such as an LED (Light Emitting Diode), and supplies irradiation light for photographing an operating part or the like to the endoscope 500.
- a light source such as an LED (Light Emitting Diode)
- LED Light Emitting Diode
- the input device 604 is an input interface for the endoscopic surgery system 11000.
- the user can input various information and input instructions to the endoscopic surgery system 400 via the input device 604.
- the user inputs an instruction to change the imaging conditions (type of irradiation light, magnification, focal length, etc.) by the endoscope 500.
- the treatment tool control device 605 controls the drive of the energy treatment tool 512 for cauterizing tissue, incising, sealing a blood vessel, or the like.
- the pneumoperitoneum device 606 gas in the body cavity through the pneumoperitoneum tube 511 in order to inflate the body cavity of the patient 532 for the purpose of securing the field of view by the endoscope 500 and securing the work space of the operator. Is sent.
- the recorder 607 is a device capable of recording various information related to surgery.
- the printer 608 is a device capable of printing various information related to surgery in various formats such as text, images, and graphs.
- the light source device 603 that supplies the irradiation light to the endoscope 500 when photographing the surgical site can be composed of, for example, an LED, a laser light source, or a white light source composed of a combination thereof.
- a white light source is configured by a combination of RGB laser light sources, the output intensity and output timing of each color (each wavelength) can be controlled with high accuracy. Therefore, the light source device 603 adjusts the white balance of the captured image. It can be carried out.
- the laser light from each of the RGB laser light sources is irradiated to the observation target in a time-division manner, and the drive of the image sensor of the camera head 502 is controlled in synchronization with the irradiation timing to correspond to each of RGB. It is also possible to capture the image in a time-division manner. According to this method, a color image can be obtained without providing a color filter in the image pickup device.
- the drive of the light source device 603 may be controlled so as to change the intensity of the output light at predetermined time intervals.
- the drive of the image sensor of the camera head 502 in synchronization with the timing of the change of the light intensity to acquire an image in time division and synthesizing the image, so-called high dynamic without blackout and overexposure. Range images can be generated.
- the light source device 603 may be configured to be able to supply light in a predetermined wavelength band corresponding to special light observation.
- special light observation for example, by utilizing the wavelength dependence of light absorption in body tissue, the surface layer of the mucous membrane is irradiated with light in a narrower band than the irradiation light (that is, white light) during normal observation.
- a so-called narrow band imaging is performed in which a predetermined tissue such as a blood vessel is photographed with high contrast.
- fluorescence observation may be performed in which an image is obtained by fluorescence generated by irradiating with excitation light.
- the body tissue is irradiated with excitation light to observe the fluorescence from the body tissue (autofluorescence observation), or a reagent such as indocyanine green (ICG) is locally injected into the body tissue and the body tissue is injected. It is possible to obtain a fluorescence image by irradiating the excitation light corresponding to the fluorescence wavelength of the reagent.
- the light source device 603 may be configured to be capable of supplying narrowband light and / or excitation light corresponding to such special light observation.
- FIG. 20 is a block diagram showing an example of the functional configuration of the camera head 502 and CCU601 shown in FIG.
- the camera head 502 includes a lens unit 701, an image pickup unit 702, a drive unit 703, a communication unit 704, and a camera head control unit 705.
- the CCU 601 has a communication unit 711, an image processing unit 712, and a control unit 713.
- the camera head 502 and the CCU 601 are communicably connected to each other by a transmission cable 700.
- the lens unit 701 is an optical system provided at a connection portion with the lens barrel 501.
- the observation light taken in from the tip of the lens barrel 501 is guided to the camera head 502 and incident on the lens unit 701.
- the lens unit 701 is configured by combining a plurality of lenses including a zoom lens and a focus lens.
- the image pickup unit 702 is composed of an image pickup element.
- the image pickup element constituting the image pickup unit 702 may be one (so-called single plate type) or a plurality (so-called multi-plate type).
- each image pickup element may generate an image signal corresponding to each of RGB, and a color image may be obtained by synthesizing them.
- the image pickup unit 702 may be configured to have a pair of image pickup elements for acquiring image signals for the right eye and the left eye corresponding to 3D (Dimensional) display, respectively.
- the 3D display enables the operator 531 to more accurately grasp the depth of the living tissue in the surgical site.
- the image pickup unit 702 is composed of a multi-plate type, a plurality of lens units 701 may be provided corresponding to each image pickup element.
- the image pickup unit 702 is, for example, the solid-state image pickup device according to any one of the first to sixth embodiments.
- the image pickup unit 702 does not necessarily have to be provided on the camera head 502.
- the image pickup unit 702 may be provided inside the lens barrel 501 immediately after the objective lens.
- the drive unit 703 is composed of an actuator, and the zoom lens and the focus lens of the lens unit 701 are moved by a predetermined distance along the optical axis under the control of the camera head control unit 705. As a result, the magnification and focus of the image captured by the image pickup unit 702 can be adjusted as appropriate.
- the communication unit 704 is configured by a communication device for transmitting and receiving various information to and from the CCU 601.
- the communication unit 704 transmits the image signal obtained from the image pickup unit 702 as RAW data to the CCU 601 via the transmission cable 700.
- the communication unit 704 receives a control signal for controlling the drive of the camera head 502 from the CCU 601 and supplies the control signal to the camera head control unit 705.
- the control signal includes, for example, information to specify the frame rate of the captured image, information to specify the exposure value at the time of imaging, and / or information to specify the magnification and focus of the captured image. Contains information about the condition.
- the image pickup conditions such as the frame rate, exposure value, magnification, and focus may be appropriately specified by the user, or may be automatically set by the control unit 713 of the CCU 601 based on the acquired image signal. good.
- the endoscope 500 is equipped with a so-called AE (Auto Exposure) function, an AF (Auto Focus) function, and an AWB (Auto White Balance) function.
- the camera head control unit 705 controls the drive of the camera head 502 based on the control signal from the CCU 601 received via the communication unit 704.
- the communication unit 711 is composed of a communication device for transmitting and receiving various information to and from the camera head 502.
- the communication unit 711 receives an image signal transmitted from the camera head 502 via the transmission cable 700.
- the communication unit 711 transmits a control signal for controlling the drive of the camera head 502 to the camera head 502.
- Image signals and control signals can be transmitted by telecommunications, optical communication, or the like.
- the image processing unit 712 performs various image processing on the image signal which is the RAW data transmitted from the camera head 502.
- the control unit 713 performs various controls related to the imaging of the surgical site and the like by the endoscope 500 and the display of the captured image obtained by the imaging of the surgical site and the like. For example, the control unit 713 generates a control signal for controlling the drive of the camera head 502.
- control unit 713 causes the display device 602 to display an image captured by the surgical unit or the like based on the image signal processed by the image processing unit 712.
- the control unit 713 may recognize various objects in the captured image by using various image recognition techniques.
- the control unit 713 detects a surgical tool such as forceps, a specific biological part, bleeding, mist when using the energy treatment tool 512, etc. by detecting the shape, color, etc. of the edge of the object included in the captured image. Can be recognized.
- the control unit 713 displays the captured image on the display device 602
- the control unit 713 may superimpose and display various surgical support information on the image of the surgical unit by using the recognition result. By superimposing and displaying the surgery support information and presenting it to the surgeon 531 it is possible to reduce the burden on the surgeon 531 and to ensure that the surgeon 531 can proceed with the surgery.
- the transmission cable 700 connecting the camera head 502 and the CCU 601 is an electric signal cable compatible with electric signal communication, an optical fiber compatible with optical communication, or a composite cable thereof.
- the communication is performed by wire using the transmission cable 700, but the communication between the camera head 502 and the CCU 601 may be performed wirelessly.
- the first substrate including the first semiconductor substrate and A plurality of photoelectric conversion units provided in the first semiconductor substrate, and A pixel separation unit provided between the photoelectric conversion units in the first semiconductor substrate is provided.
- a solid-state image pickup device having a ⁇ 100 ⁇ surface at the interface between the side surface of the pixel separation portion and the first semiconductor substrate.
- the insulating film is provided at a first portion having a first film thickness in a plan view and a second portion provided at a corner portion of the pixel separation portion and having a second film thickness thicker than the first film thickness.
- the pixel separation portion includes a plurality of first portions extending in a first direction parallel to the surface of the first semiconductor substrate in a plan view, and a plurality of second portions extending in a second direction parallel to the surface of the first semiconductor substrate.
- the solid-state imaging device according to (1) which includes a portion.
- the pixel separation portion includes a plurality of first portions extending in a first direction parallel to the surface of the first semiconductor substrate in a plan view, and a plurality of second portions extending in a second direction parallel to the surface of the first semiconductor substrate.
- the first or second direction is parallel to the ⁇ 110> direction of the second semiconductor substrate.
- the first or second direction is parallel to the ⁇ 100> direction of the second semiconductor substrate.
- the first or second direction is parallel to the ⁇ 100> direction of the second semiconductor substrate.
- the first or second direction is parallel to the ⁇ 110> direction of the second semiconductor substrate.
- the first substrate including the first semiconductor substrate and A plurality of photoelectric conversion units provided in the first semiconductor substrate, and A pixel separation unit provided between the photoelectric conversion units in the first semiconductor substrate is provided.
- the pixel separation portion includes an insulating film and contains an insulating film.
- the insulating film is provided at a first portion having a first film thickness in a plan view and a second portion provided at a corner portion of the pixel separation portion and having a second film thickness thicker than the first film thickness.
- a solid-state image sensor including.
- a plurality of photoelectric conversion units are formed in the first semiconductor substrate of the first substrate, and a plurality of photoelectric conversion units are formed.
- a pixel separation section is formed between the photoelectric conversion sections in the first semiconductor substrate.
- the insulating film is provided at a first portion having a first film thickness in a plan view and a second portion provided at a corner portion of the pixel separation portion and having a second film thickness thicker than the first film thickness.
- Pixel 2 Pixel array area 3: Control circuit, 4: Vertical drive circuit, 5: Column signal processing circuit, 6: Horizontal drive circuit, 7: Output circuit, 8: Vertical signal line, 9: Horizontal signal line, 11: Semiconductor substrate, 11': Substrate, 11a: Chip region, 11b: Dicing region, 11c: Source diffusion layer, 11d: Drain diffusion layer, 12: photoelectric conversion unit, 13: n-type semiconductor region, 14: p-type semiconductor region, 21: Pixel separation groove, 21a: 1st linear portion, 21b: 2nd linear portion, 22: Pixel separation part, 22a: 1st linear part, 22b: 2nd linear part, 23: Insulating film, 23a: 1st part, 23b: 2nd part, 24: Light-shielding film, 25: light-shielding film, 26: flattening film, 27: color filter, 28: on-chip lens, 31: substrate, 32: insulating layer, 32a: insulating film, 32b: interlayer insulating film, 31
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Power Engineering (AREA)
- Electromagnetism (AREA)
- Condensed Matter Physics & Semiconductors (AREA)
- General Physics & Mathematics (AREA)
- Computer Hardware Design (AREA)
- Microelectronics & Electronic Packaging (AREA)
- Solid State Image Pick-Up Elements (AREA)
Abstract
Description
図1は、第1実施形態の固体撮像装置の構成を示すブロック図である。
光電変換部12は、半導体基板11内に各画素1ごとに設けられている。図2のAは、1つの画素1に含まれる1つの光電変換部12を示している。光電変換部12は、半導体基板11内に設けられたn型半導体領域13と、半導体基板11内でn型半導体領域13の周囲に設けられたp型半導体領域14とを含んでいる。光電変換部12では、n型半導体領域13とp型半導体領域14との間のpn接合によりフォトダイオードが実現されており、フォトダイオードが光を電荷に変換する。光電変換部12は、半導体基板11の裏面側から光を受光し、受光した光の光量に応じた信号電荷を生成し、生成した信号電荷をn型半導体領域13に蓄積する。
遮光膜25は、半導体基板11外にて画素分離部22上に設けられている。遮光膜25は例えば、W(タングステン)、Al(アルミニウム)、またはCu(銅)といった金属元素を含む膜であり、光を遮光する作用を有している。遮光膜25は、遮光膜24と同時に形成されてもよい。
図9は、第2実施形態の固体撮像装置の構造を示す断面図と平面図である。
図10は、第3実施形態の固体撮像装置の構造を示す断面図と平面図である。
図11は、第4実施形態の固体撮像装置の構造を示す断面図と平面図である。
図13は、第5実施形態の固体撮像装置の構造を示す断面図と平面図である。
図15は、第6実施形態の固体撮像装置の構造を示す断面図と平面図である。
図16は、電子機器の構成例を示すブロック図である。図16に示す電気機器は、カメラ100である。
第1半導体基板を含む第1基板と、
前記第1半導体基板内に設けられた複数の光電変換部と、
前記第1半導体基板内で前記光電変換部間に設けられた画素分離部とを備え、
前記画素分離部の側面と前記第1半導体基板との界面は、{100}面を有する、固体撮像装置。
前記画素分離部は、絶縁膜を含む、(1)に記載の固体撮像装置。
前記画素分離部はさらに、遮光膜を含む、(2)に記載の固体撮像装置。
前記絶縁膜は、前記第1半導体基板に含まれる元素と、酸素とを含む、(2)に記載の固体撮像装置。
前記絶縁膜は、平面視で第1の膜厚を有する第1部分と、前記画素分離部の角部に設けられ、前記第1の膜厚よりも厚い第2の膜厚を有する第2部分とを含む、(2)に記載の固体撮像装置。
前記画素分離部は、平面視で前記第1半導体基板の表面に平行な第1方向に延びる複数の第1部分と、前記第1半導体基板の表面に平行な第2方向に延びる複数の第2部分とを含む、(1)に記載の固体撮像装置。
前記平面視は、前記第1半導体基板の光入射面を視た状態に当たる、(5)に記載の固体撮像装置。
前記第1または第2方向は、前記第1半導体基板の<100>方向と平行である、(6)に記載の固体撮像装置。
前記画素分離部は、前記第1半導体基板を貫通する画素分離溝内に設けられている、(1)に記載の固体撮像装置。
前記画素分離部は、前記第1半導体基板を貫通しない画素分離溝内に設けられている、(1)に記載の固体撮像装置。
前記第1基板の光入射面とは反対側に設けられた第1絶縁層と、
前記第1絶縁層と対向するように設けられた第2半導体基板を含む第2基板とをさらに備え、
前記第2基板は、トランジスタを含む、(1)に記載の固体撮像装置。
前記画素分離部は、平面視で前記第1半導体基板の表面に平行な第1方向に延びる複数の第1部分と、前記第1半導体基板の表面に平行な第2方向に延びる複数の第2部分とを含む、(11)に記載の固体撮像装置。
前記第1または第2方向は、前記第2半導体基板の<110>方向に平行であり、
前記トランジスタは、<110>方向に平行なチャネル方向を有するn型平面トランジスタである、(12)に記載の固体撮像装置。
前記第1または第2方向は、前記第2半導体基板の<100>方向に平行であり、
前記トランジスタは、前記第2半導体基板の{100}面であるフィン側壁を有し、前記第1または第2方向に平行なチャネル方向を有するフィン型トランジスタである、(12)に記載の固体撮像装置。
前記第1または第2方向は、前記第2半導体基板の<100>方向に平行であり、
前記トランジスタは、<100>方向に平行なチャネル方向を有するp型平面トランジスタである、(12)に記載の固体撮像装置。
前記第1または第2方向は、前記第2半導体基板の<110>方向に平行であり、
前記トランジスタは、前記第2半導体基板の{100}面であるフィン側壁を有し、前記第1および第2方向に非平行なチャネル方向を有するフィン型トランジスタである、(12)に記載の固体撮像装置。
第1半導体基板を含む第1基板と、
前記第1半導体基板内に設けられた複数の光電変換部と、
前記第1半導体基板内で前記光電変換部間に設けられた画素分離部とを備え、
前記画素分離部は、絶縁膜を含み、
前記絶縁膜は、平面視で第1の膜厚を有する第1部分と、前記画素分離部の角部に設けられ、前記第1の膜厚よりも厚い第2の膜厚を有する第2部分とを含む、固体撮像装置。
第1基板の第1半導体基板内に複数の光電変換部を形成し、
前記第1半導体基板内で前記光電変換部間に画素分離部を形成する、
ことを含み、
前記画素分離部は、前記画素分離部の側面と前記第1半導体基板との界面が、{100}面を有するように形成される、固体撮像装置の製造方法。
前記画素分離部は、絶縁膜を含むように形成される、(18)に記載の固体撮像装置の製造方法。
前記絶縁膜は、平面視で第1の膜厚を有する第1部分と、前記画素分離部の角部に設けられ、前記第1の膜厚よりも厚い第2の膜厚を有する第2部分とを含むように形成される、(19)に記載の固体撮像装置の製造方法。
4:垂直駆動回路、5:カラム信号処理回路、6:水平駆動回路、
7:出力回路、8:垂直信号線、9:水平信号線、
11:半導体基板、11’:基板、11a:チップ領域、11b:ダイシング領域、
11c:ソース拡散層、11d:ドレイン拡散層、
12:光電変換部、13:n型半導体領域、14:p型半導体領域、
21:画素分離溝、21a:第1線状部分、21b:第2線状部分、
22:画素分離部、22a:第1線状部分、22b:第2線状部分、
23:絶縁膜、23a:第1部分、23b:第2部分、24:遮光膜、
25:遮光膜、26:平坦化膜、27:カラーフィルタ、28:オンチップレンズ、
31:基板、32:絶縁層、32a:絶縁膜、32b:層間絶縁膜、
33:半導体基板、33’:基板、33a:ソース拡散層、33b:ドレイン拡散層、
34:絶縁層、34a:絶縁膜、34b:層間絶縁膜、
35:ゲート電極、36:ゲート電極、36a:平面部分、36b:フィン部分、
41:プラグ、42:絶縁膜、43:プラグ、44:配線層
Claims (20)
- 第1半導体基板を含む第1基板と、
前記第1半導体基板内に設けられた複数の光電変換部と、
前記第1半導体基板内で前記光電変換部間に設けられた画素分離部とを備え、
前記画素分離部の側面と前記第1半導体基板との界面は、{100}面を有する、固体撮像装置。 - 前記画素分離部は、絶縁膜を含む、請求項1に記載の固体撮像装置。
- 前記画素分離部はさらに、遮光膜を含む、請求項2に記載の固体撮像装置。
- 前記絶縁膜は、前記第1半導体基板に含まれる元素と、酸素とを含む、請求項2に記載の固体撮像装置。
- 前記絶縁膜は、平面視で第1の膜厚を有する第1部分と、前記画素分離部の角部に設けられ、前記第1の膜厚よりも厚い第2の膜厚を有する第2部分とを含む、請求項2に記載の固体撮像装置。
- 前記画素分離部は、平面視で前記第1半導体基板の表面に平行な第1方向に延びる複数の第1部分と、前記第1半導体基板の表面に平行な第2方向に延びる複数の第2部分とを含む、請求項1に記載の固体撮像装置。
- 前記平面視は、前記第1半導体基板の光入射面を視た状態に当たる、請求項5に記載の固体撮像装置。
- 前記第1または第2方向は、前記第1半導体基板の<100>方向と平行である、請求項6に記載の固体撮像装置。
- 前記画素分離部は、前記第1半導体基板を貫通する画素分離溝内に設けられている、請求項1に記載の固体撮像装置。
- 前記画素分離部は、前記第1半導体基板を貫通しない画素分離溝内に設けられている、請求項1に記載の固体撮像装置。
- 前記第1基板の光入射面とは反対側に設けられた第1絶縁層と、
前記第1絶縁層と対向するように設けられた第2半導体基板を含む第2基板とをさらに備え、
前記第2基板は、トランジスタを含む、請求項1に記載の固体撮像装置。 - 前記画素分離部は、平面視で前記第1半導体基板の表面に平行な第1方向に延びる複数の第1部分と、前記第1半導体基板の表面に平行な第2方向に延びる複数の第2部分とを含む、請求項11に記載の固体撮像装置。
- 前記第1または第2方向は、前記第2半導体基板の<110>方向に平行であり、
前記トランジスタは、<110>方向に平行なチャネル方向を有するn型平面トランジスタである、請求項12に記載の固体撮像装置。 - 前記第1または第2方向は、前記第2半導体基板の<100>方向に平行であり、
前記トランジスタは、前記第2半導体基板の{100}面であるフィン側壁を有し、前記第1または第2方向に平行なチャネル方向を有するフィン型トランジスタである、請求項12に記載の固体撮像装置。 - 前記第1または第2方向は、前記第2半導体基板の<100>方向に平行であり、
前記トランジスタは、<100>方向に平行なチャネル方向を有するp型平面トランジスタである、請求項12に記載の固体撮像装置。 - 前記第1または第2方向は、前記第2半導体基板の<110>方向に平行であり、
前記トランジスタは、前記第2半導体基板の{100}面であるフィン側壁を有し、前記第1および第2方向に非平行なチャネル方向を有するフィン型トランジスタである、請求項12に記載の固体撮像装置。 - 第1半導体基板を含む第1基板と、
前記第1半導体基板内に設けられた複数の光電変換部と、
前記第1半導体基板内で前記光電変換部間に設けられた画素分離部とを備え、
前記画素分離部は、絶縁膜を含み、
前記絶縁膜は、平面視で第1の膜厚を有する第1部分と、前記画素分離部の角部に設けられ、前記第1の膜厚よりも厚い第2の膜厚を有する第2部分とを含む、固体撮像装置。 - 第1基板の第1半導体基板内に複数の光電変換部を形成し、
前記第1半導体基板内で前記光電変換部間に画素分離部を形成する、
ことを含み、
前記画素分離部は、前記画素分離部の側面と前記第1半導体基板との界面が、{100}面を有するように形成される、固体撮像装置の製造方法。 - 前記画素分離部は、絶縁膜を含むように形成される、請求項18に記載の固体撮像装置の製造方法。
- 前記絶縁膜は、平面視で第1の膜厚を有する第1部分と、前記画素分離部の角部に設けられ、前記第1の膜厚よりも厚い第2の膜厚を有する第2部分とを含むように形成される、請求項19に記載の固体撮像装置の製造方法。
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP21910247.2A EP4270451A1 (en) | 2020-12-25 | 2021-12-06 | Solid-state imaging device and method for manufacturing same |
CN202180085150.5A CN116670804A (zh) | 2020-12-25 | 2021-12-06 | 固态摄像装置及其制造方法 |
JP2022572071A JPWO2022138097A1 (ja) | 2020-12-25 | 2021-12-06 | |
US18/257,921 US20240006446A1 (en) | 2020-12-25 | 2021-12-06 | Solid-state imaging device and method for manufacturing the same |
KR1020237019258A KR20230123941A (ko) | 2020-12-25 | 2021-12-06 | 고체 촬상 장치 및 그 제조 방법 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2020217541 | 2020-12-25 | ||
JP2020-217541 | 2020-12-25 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022138097A1 true WO2022138097A1 (ja) | 2022-06-30 |
Family
ID=82159537
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2021/044625 WO2022138097A1 (ja) | 2020-12-25 | 2021-12-06 | 固体撮像装置およびその製造方法 |
Country Status (7)
Country | Link |
---|---|
US (1) | US20240006446A1 (ja) |
EP (1) | EP4270451A1 (ja) |
JP (1) | JPWO2022138097A1 (ja) |
KR (1) | KR20230123941A (ja) |
CN (1) | CN116670804A (ja) |
TW (1) | TW202226564A (ja) |
WO (1) | WO2022138097A1 (ja) |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2013175494A (ja) | 2011-03-02 | 2013-09-05 | Sony Corp | 固体撮像装置、固体撮像装置の製造方法及び電子機器 |
US20160268321A1 (en) * | 2015-03-11 | 2016-09-15 | Samsung Electronics Co., Ltd. | Image sensor devices using offset pixel patterns |
JP2018148116A (ja) | 2017-03-08 | 2018-09-20 | ソニーセミコンダクタソリューションズ株式会社 | 固体撮像装置、および電子機器 |
WO2019093151A1 (ja) * | 2017-11-09 | 2019-05-16 | ソニーセミコンダクタソリューションズ株式会社 | 固体撮像装置、および電子機器 |
WO2020170936A1 (ja) * | 2019-02-20 | 2020-08-27 | ソニーセミコンダクタソリューションズ株式会社 | 撮像装置 |
-
2021
- 2021-11-18 TW TW110142861A patent/TW202226564A/zh unknown
- 2021-12-06 JP JP2022572071A patent/JPWO2022138097A1/ja active Pending
- 2021-12-06 CN CN202180085150.5A patent/CN116670804A/zh active Pending
- 2021-12-06 KR KR1020237019258A patent/KR20230123941A/ko unknown
- 2021-12-06 EP EP21910247.2A patent/EP4270451A1/en active Pending
- 2021-12-06 US US18/257,921 patent/US20240006446A1/en active Pending
- 2021-12-06 WO PCT/JP2021/044625 patent/WO2022138097A1/ja active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2013175494A (ja) | 2011-03-02 | 2013-09-05 | Sony Corp | 固体撮像装置、固体撮像装置の製造方法及び電子機器 |
US20160268321A1 (en) * | 2015-03-11 | 2016-09-15 | Samsung Electronics Co., Ltd. | Image sensor devices using offset pixel patterns |
JP2018148116A (ja) | 2017-03-08 | 2018-09-20 | ソニーセミコンダクタソリューションズ株式会社 | 固体撮像装置、および電子機器 |
WO2019093151A1 (ja) * | 2017-11-09 | 2019-05-16 | ソニーセミコンダクタソリューションズ株式会社 | 固体撮像装置、および電子機器 |
WO2020170936A1 (ja) * | 2019-02-20 | 2020-08-27 | ソニーセミコンダクタソリューションズ株式会社 | 撮像装置 |
Also Published As
Publication number | Publication date |
---|---|
JPWO2022138097A1 (ja) | 2022-06-30 |
CN116670804A (zh) | 2023-08-29 |
US20240006446A1 (en) | 2024-01-04 |
KR20230123941A (ko) | 2023-08-24 |
TW202226564A (zh) | 2022-07-01 |
EP4270451A1 (en) | 2023-11-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11869914B2 (en) | Image sensor and method for manufacturing image sensor | |
WO2020012824A1 (ja) | 固体撮像素子および電子機器 | |
WO2020137203A1 (ja) | 撮像素子および撮像装置 | |
JP2022093359A (ja) | 測距素子 | |
WO2019155782A1 (ja) | 半導体装置および半導体装置の製造方法 | |
US20230103730A1 (en) | Solid-state imaging device | |
WO2021100332A1 (ja) | 半導体装置、固体撮像装置及び電子機器 | |
WO2019188131A1 (ja) | 半導体装置および半導体装置の製造方法 | |
WO2022009627A1 (ja) | 固体撮像装置および電子機器 | |
KR20240037943A (ko) | 촬상 장치 | |
WO2021186907A1 (ja) | 固体撮像装置及びその製造方法、並びに電子機器 | |
US20220311943A1 (en) | Imaging element and imaging apparatus | |
US20220254823A1 (en) | Imaging device | |
WO2020195180A1 (ja) | 撮像素子および撮像装置 | |
WO2022138097A1 (ja) | 固体撮像装置およびその製造方法 | |
WO2022130987A1 (ja) | 固体撮像装置およびその製造方法 | |
US20240038807A1 (en) | Solid-state imaging device | |
WO2022085695A1 (ja) | 固体撮像装置 | |
WO2024024269A1 (ja) | 固体撮像装置およびその製造方法 | |
WO2022249678A1 (ja) | 固体撮像装置およびその製造方法 | |
WO2023021787A1 (ja) | 光検出装置およびその製造方法 | |
WO2023080011A1 (ja) | 撮像装置及び電子機器 | |
WO2022201745A1 (ja) | 固体撮像装置及び固体撮像装置の製造方法 | |
US20230170360A1 (en) | Imaging apparatus and electronic device | |
WO2020090384A1 (ja) | 撮像装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21910247 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2022572071 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 202180085150.5 Country of ref document: CN Ref document number: 18257921 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 2021910247 Country of ref document: EP Effective date: 20230725 |