WO2018180575A1 - 固体撮像素子、電子機器、並びに製造方法 - Google Patents
固体撮像素子、電子機器、並びに製造方法 Download PDFInfo
- Publication number
- WO2018180575A1 WO2018180575A1 PCT/JP2018/010393 JP2018010393W WO2018180575A1 WO 2018180575 A1 WO2018180575 A1 WO 2018180575A1 JP 2018010393 W JP2018010393 W JP 2018010393W WO 2018180575 A1 WO2018180575 A1 WO 2018180575A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- solid
- state imaging
- photoelectric conversion
- electrode
- semiconductor substrate
- Prior art date
Links
- 238000003384 imaging method Methods 0.000 title claims abstract description 167
- 238000004519 manufacturing process Methods 0.000 title claims abstract description 33
- 238000006243 chemical reaction Methods 0.000 claims abstract description 106
- 239000004065 semiconductor Substances 0.000 claims abstract description 82
- 239000000758 substrate Substances 0.000 claims abstract description 79
- 238000009792 diffusion process Methods 0.000 claims abstract description 19
- 238000007667 floating Methods 0.000 claims abstract description 19
- 238000012545 processing Methods 0.000 claims description 72
- 238000000926 separation method Methods 0.000 claims description 21
- 239000000463 material Substances 0.000 claims description 17
- 230000003287 optical effect Effects 0.000 claims description 16
- 239000007769 metal material Substances 0.000 claims description 7
- 229910021417 amorphous silicon Inorganic materials 0.000 claims description 4
- 229910021420 polycrystalline silicon Inorganic materials 0.000 claims description 4
- 229920005591 polysilicon Polymers 0.000 claims description 4
- 230000009467 reduction Effects 0.000 abstract description 2
- 239000010410 layer Substances 0.000 description 82
- 238000005516 engineering process Methods 0.000 description 45
- 238000000034 method Methods 0.000 description 38
- 210000003128 head Anatomy 0.000 description 28
- 238000004891 communication Methods 0.000 description 27
- 238000001514 detection method Methods 0.000 description 23
- 230000000875 corresponding effect Effects 0.000 description 21
- 230000006870 function Effects 0.000 description 18
- 230000008569 process Effects 0.000 description 18
- 238000012546 transfer Methods 0.000 description 17
- 238000001727 in vivo Methods 0.000 description 16
- 239000002775 capsule Substances 0.000 description 13
- 238000002674 endoscopic surgery Methods 0.000 description 12
- 238000010586 diagram Methods 0.000 description 11
- XUIMIQQOPSSXEZ-UHFFFAOYSA-N Silicon Chemical compound [Si] XUIMIQQOPSSXEZ-UHFFFAOYSA-N 0.000 description 9
- 230000003321 amplification Effects 0.000 description 9
- 230000005540 biological transmission Effects 0.000 description 9
- 238000003199 nucleic acid amplification method Methods 0.000 description 9
- 229910052710 silicon Inorganic materials 0.000 description 9
- 239000010703 silicon Substances 0.000 description 9
- 210000001519 tissue Anatomy 0.000 description 9
- 239000011229 interlayer Substances 0.000 description 8
- 230000001681 protective effect Effects 0.000 description 8
- VYPSYNLAJGMNEJ-UHFFFAOYSA-N Silicium dioxide Chemical compound O=[Si]=O VYPSYNLAJGMNEJ-UHFFFAOYSA-N 0.000 description 7
- XLOMVQKBTHCTTD-UHFFFAOYSA-N Zinc monoxide Chemical compound [Zn]=O XLOMVQKBTHCTTD-UHFFFAOYSA-N 0.000 description 7
- 229910052814 silicon oxide Inorganic materials 0.000 description 7
- 230000035945 sensitivity Effects 0.000 description 6
- 238000001356 surgical procedure Methods 0.000 description 6
- 229910052581 Si3N4 Inorganic materials 0.000 description 5
- 229910052751 metal Inorganic materials 0.000 description 5
- 239000002184 metal Substances 0.000 description 5
- HQVNEWCFYHHQES-UHFFFAOYSA-N silicon nitride Chemical compound N12[Si]34N5[Si]62N3[Si]51N64 HQVNEWCFYHHQES-UHFFFAOYSA-N 0.000 description 5
- 239000010936 titanium Substances 0.000 description 5
- RTAQQCXQSZGOHL-UHFFFAOYSA-N Titanium Chemical compound [Ti] RTAQQCXQSZGOHL-UHFFFAOYSA-N 0.000 description 4
- 229910052782 aluminium Inorganic materials 0.000 description 4
- XAGFODPZIPBFFR-UHFFFAOYSA-N aluminium Chemical compound [Al] XAGFODPZIPBFFR-UHFFFAOYSA-N 0.000 description 4
- 230000001276 controlling effect Effects 0.000 description 4
- 230000005284 excitation Effects 0.000 description 4
- 230000007246 mechanism Effects 0.000 description 4
- 229910052719 titanium Inorganic materials 0.000 description 4
- WFKWXMTUELFFGS-UHFFFAOYSA-N tungsten Chemical compound [W] WFKWXMTUELFFGS-UHFFFAOYSA-N 0.000 description 4
- 229910052721 tungsten Inorganic materials 0.000 description 4
- 239000010937 tungsten Substances 0.000 description 4
- 239000011787 zinc oxide Substances 0.000 description 4
- 238000009825 accumulation Methods 0.000 description 3
- 230000015572 biosynthetic process Effects 0.000 description 3
- 210000004204 blood vessel Anatomy 0.000 description 3
- 230000008859 change Effects 0.000 description 3
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- JAONJTDQXUSBGG-UHFFFAOYSA-N dialuminum;dizinc;oxygen(2-) Chemical compound [O-2].[O-2].[O-2].[O-2].[O-2].[Al+3].[Al+3].[Zn+2].[Zn+2] JAONJTDQXUSBGG-UHFFFAOYSA-N 0.000 description 3
- 239000002019 doping agent Substances 0.000 description 3
- 238000001312 dry etching Methods 0.000 description 3
- 230000000694 effects Effects 0.000 description 3
- 238000010336 energy treatment Methods 0.000 description 3
- 239000012535 impurity Substances 0.000 description 3
- 230000001678 irradiating effect Effects 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 239000002356 single layer Substances 0.000 description 3
- 238000003860 storage Methods 0.000 description 3
- XOLBLPGZBRYERU-UHFFFAOYSA-N tin dioxide Chemical compound O=[Sn]=O XOLBLPGZBRYERU-UHFFFAOYSA-N 0.000 description 3
- GYHNNYVSQQEPJS-UHFFFAOYSA-N Gallium Chemical compound [Ga] GYHNNYVSQQEPJS-UHFFFAOYSA-N 0.000 description 2
- 208000005646 Pneumoperitoneum Diseases 0.000 description 2
- NRTOMJZYCJJWKI-UHFFFAOYSA-N Titanium nitride Chemical compound [Ti]#N NRTOMJZYCJJWKI-UHFFFAOYSA-N 0.000 description 2
- 230000000903 blocking effect Effects 0.000 description 2
- 239000003153 chemical reaction reagent Substances 0.000 description 2
- 238000005229 chemical vapour deposition Methods 0.000 description 2
- 229910017052 cobalt Inorganic materials 0.000 description 2
- 239000010941 cobalt Substances 0.000 description 2
- GUTLYIVDDKVIGB-UHFFFAOYSA-N cobalt atom Chemical compound [Co] GUTLYIVDDKVIGB-UHFFFAOYSA-N 0.000 description 2
- 238000012937 correction Methods 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 230000018109 developmental process Effects 0.000 description 2
- 238000005530 etching Methods 0.000 description 2
- 229910052733 gallium Inorganic materials 0.000 description 2
- 229910052735 hafnium Inorganic materials 0.000 description 2
- VBJZVLUMGGDVMO-UHFFFAOYSA-N hafnium atom Chemical compound [Hf] VBJZVLUMGGDVMO-UHFFFAOYSA-N 0.000 description 2
- MOFVSTNWEDAEEK-UHFFFAOYSA-M indocyanine green Chemical compound [Na+].[O-]S(=O)(=O)CCCCN1C2=CC=C3C=CC=CC3=C2C(C)(C)C1=CC=CC=CC=CC1=[N+](CCCCS([O-])(=O)=O)C2=CC=C(C=CC=C3)C3=C2C1(C)C MOFVSTNWEDAEEK-UHFFFAOYSA-M 0.000 description 2
- 229960004657 indocyanine green Drugs 0.000 description 2
- 238000009434 installation Methods 0.000 description 2
- 238000009413 insulation Methods 0.000 description 2
- 238000005468 ion implantation Methods 0.000 description 2
- 230000002093 peripheral effect Effects 0.000 description 2
- 229910052715 tantalum Inorganic materials 0.000 description 2
- GUVRBAGPIYLISA-UHFFFAOYSA-N tantalum atom Chemical compound [Ta] GUVRBAGPIYLISA-UHFFFAOYSA-N 0.000 description 2
- OAICVXFJPJFONN-UHFFFAOYSA-N Phosphorus Chemical compound [P] OAICVXFJPJFONN-UHFFFAOYSA-N 0.000 description 1
- 241000519995 Stachys sylvatica Species 0.000 description 1
- BOTDANWDWHJENH-UHFFFAOYSA-N Tetraethyl orthosilicate Chemical compound CCO[Si](OCC)(OCC)OCC BOTDANWDWHJENH-UHFFFAOYSA-N 0.000 description 1
- 229910003363 ZnMgO Inorganic materials 0.000 description 1
- 229910007694 ZnSnO3 Inorganic materials 0.000 description 1
- 238000002679 ablation Methods 0.000 description 1
- 238000010521 absorption reaction Methods 0.000 description 1
- 230000001133 acceleration Effects 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 230000002730 additional effect Effects 0.000 description 1
- 238000000137 annealing Methods 0.000 description 1
- 230000003796 beauty Effects 0.000 description 1
- 230000000740 bleeding effect Effects 0.000 description 1
- 230000003139 buffering effect Effects 0.000 description 1
- 238000002485 combustion reaction Methods 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 239000004020 conductor Substances 0.000 description 1
- 239000000470 constituent Substances 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 238000002073 fluorescence micrograph Methods 0.000 description 1
- CJNBYAVZURUTKZ-UHFFFAOYSA-N hafnium(iv) oxide Chemical compound O=[Hf]=O CJNBYAVZURUTKZ-UHFFFAOYSA-N 0.000 description 1
- 230000036541 health Effects 0.000 description 1
- 230000005525 hole transport Effects 0.000 description 1
- 229910052738 indium Inorganic materials 0.000 description 1
- APFVFJFRJDLVQX-UHFFFAOYSA-N indium atom Chemical compound [In] APFVFJFRJDLVQX-UHFFFAOYSA-N 0.000 description 1
- AMGQUBHHOARCQH-UHFFFAOYSA-N indium;oxotin Chemical compound [In].[Sn]=O AMGQUBHHOARCQH-UHFFFAOYSA-N 0.000 description 1
- 239000012212 insulator Substances 0.000 description 1
- 210000000936 intestine Anatomy 0.000 description 1
- 238000002955 isolation Methods 0.000 description 1
- 230000031700 light absorption Effects 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 239000003595 mist Substances 0.000 description 1
- 230000000116 mitigating effect Effects 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 210000004400 mucous membrane Anatomy 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 210000000056 organ Anatomy 0.000 description 1
- 230000035515 penetration Effects 0.000 description 1
- 230000002572 peristaltic effect Effects 0.000 description 1
- 229910052698 phosphorus Inorganic materials 0.000 description 1
- 239000011574 phosphorus Substances 0.000 description 1
- 238000005498 polishing Methods 0.000 description 1
- 238000007639 printing Methods 0.000 description 1
- 230000008929 regeneration Effects 0.000 description 1
- 238000011069 regeneration method Methods 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 210000004761 scalp Anatomy 0.000 description 1
- 238000007789 sealing Methods 0.000 description 1
- 239000002210 silicon-based material Substances 0.000 description 1
- 238000004611 spectroscopical analysis Methods 0.000 description 1
- 210000002784 stomach Anatomy 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 229910001887 tin oxide Inorganic materials 0.000 description 1
- 238000002834 transmittance Methods 0.000 description 1
- YVTHLONGBIQYBO-UHFFFAOYSA-N zinc indium(3+) oxygen(2-) Chemical compound [O--].[Zn++].[In+3] YVTHLONGBIQYBO-UHFFFAOYSA-N 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L21/00—Processes or apparatus adapted for the manufacture or treatment of semiconductor or solid state devices or of parts thereof
- H01L21/70—Manufacture or treatment of devices consisting of a plurality of solid state components formed in or on a common substrate or of parts thereof; Manufacture of integrated circuit devices or of parts thereof
- H01L21/71—Manufacture of specific parts of devices defined in group H01L21/70
- H01L21/768—Applying interconnections to be used for carrying current between separate components within a device comprising conductors and dielectrics
- H01L21/76898—Applying interconnections to be used for carrying current between separate components within a device comprising conductors and dielectrics formed through a semiconductor substrate
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/14636—Interconnect structures
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/14603—Special geometry or disposition of pixel-elements, address-lines or gate-electrodes
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/14609—Pixel-elements with integrated switching, control, storage or amplification elements
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/14609—Pixel-elements with integrated switching, control, storage or amplification elements
- H01L27/14612—Pixel-elements with integrated switching, control, storage or amplification elements involving a transistor
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/1462—Coatings
- H01L27/14623—Optical shielding
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14643—Photodiode arrays; MOS imagers
- H01L27/14645—Colour imagers
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14683—Processes or apparatus peculiar to the manufacture or treatment of these devices or parts thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/70—SSIS architectures; Circuits associated therewith
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/70—SSIS architectures; Circuits associated therewith
- H04N25/76—Addressed sensors, e.g. MOS or CMOS sensors
Definitions
- the present disclosure relates to a solid-state imaging device, an electronic device, and a manufacturing method, and more particularly, to a solid-state imaging device, an electronic device, and a manufacturing method that can further reduce an element layout area.
- an image sensor in which a photoelectric conversion film is stacked has a through electrode structure connecting a photoelectric conversion film and a floating diffusion as disclosed in Patent Document 1, for example.
- a through electrode structure connecting a photoelectric conversion film and a floating diffusion as disclosed in Patent Document 1, for example.
- the through electrodes are arranged close to each other.
- a shield electrode is arranged to prevent electrical color mixture due to capacitive coupling.
- Patent Document 1 since there is a concern that the wiring connected to the shield electrode and the floating diffusion may be short-circuited due to process variations such as overlay and line width, a countermeasure is taken to separate the through electrode from the shield electrode. . As a result, the pixel size is increased, the photodiode area is reduced, and the saturation signal amount and sensitivity characteristics are reduced.
- the present disclosure has been made in view of such a situation, and can further reduce the element layout area.
- a solid-state imaging device includes a photoelectric conversion element provided on a first surface side of a semiconductor substrate, connected to the photoelectric conversion element, the first surface of the semiconductor substrate, the first surface, A through electrode provided between the second surface, which are different surfaces, a transistor and a floating diffusion provided on the second surface and connected to the photoelectric conversion element via the through electrode, and the second surface And a dielectric layer thicker than the gate insulating film of the transistor, which is formed between the through electrodes.
- a shield electrode may be further provided inside the dielectric layer as viewed from the second surface side.
- the material of the shield electrode is polysilicon or amorphous silicon.
- the shield electrode is controlled to a predetermined voltage from a contact electrically connected to the shield electrode and a wiring electrically connected to the contact.
- the contact electrically connected to the shield electrode is provided between the through electrodes.
- the contacts electrically connected to the shield electrode are provided in a line shape between the through electrodes.
- the dielectric layer is provided around the through electrode.
- the through electrode penetrates the semiconductor substrate and is separated from the semiconductor substrate by a separation groove, and the dielectric layer and the separation groove are in contact with each other.
- One or more photodiodes provided in the semiconductor substrate can be further provided.
- An electronic apparatus includes: a photoelectric conversion element provided on a first surface side of a semiconductor substrate; and the first surface of the semiconductor substrate and the first surface connected to the photoelectric conversion element.
- a through electrode provided between a second surface and a different surface; a transistor and a floating diffusion provided on the second surface and connected to the photoelectric conversion element via the through electrode;
- a solid-state imaging device including a dielectric layer thicker than a gate insulating film of the transistor, formed between the through electrodes, a signal processing circuit for processing an output signal output from the solid-state imaging device, and incident light And an optical system incident on the solid-state imaging device.
- a manufacturing apparatus forms a through hole between a first surface of a semiconductor substrate and a second surface that is a surface different from the first surface, and the through hole in the second surface is formed.
- a dielectric layer thicker than the gate insulating film of the transistor formed on the second surface is formed, a gate wiring including the transistor is formed on the second surface, and a metal material film is embedded in the through hole.
- a through electrode is formed, and the photoelectric conversion element connected to the transistor and the floating diffusion through the through electrode is formed on the first surface.
- a through electrode provided with a photoelectric conversion element on a first surface side of a semiconductor substrate is connected to the photoelectric conversion element, and the first surface of the semiconductor substrate and the first surface
- the transistor and the floating diffusion are provided on the second surface and connected to the photoelectric conversion element through the through electrode.
- a dielectric layer thicker than the gate insulating film of the transistor is formed between the through electrodes on the second surface.
- the element layout area can be further reduced.
- FIG. 2 Use example of image sensor Example 4 of electronic device 4.
- FIG. 1 shows a schematic configuration example of an example of a CMOS (Complementary Metal Oxide Semiconductor) solid-state imaging device applied to each embodiment of the present technology.
- CMOS Complementary Metal Oxide Semiconductor
- a solid-state imaging device (element chip) 1 includes a pixel region (a pixel region in which pixels 2 including a plurality of photoelectric conversion elements are regularly arranged two-dimensionally on a semiconductor substrate 11 (for example, a silicon substrate). A so-called imaging region) 3 and a peripheral circuit region.
- the pixel 2 includes a photoelectric conversion element (for example, PD (Photo Diode)) and a plurality of pixel transistors (so-called MOS transistors).
- the plurality of pixel transistors can be constituted by three transistors, for example, a transfer transistor, a reset transistor, and an amplifying transistor, and can further be constituted by four transistors by adding a selection transistor.
- the pixel 2 can have a pixel sharing structure.
- the pixel sharing structure includes a plurality of photodiodes, a plurality of transfer transistors, one shared floating diffusion, and one other pixel transistor that is shared.
- the photodiode is a photoelectric conversion element.
- the peripheral circuit area includes a vertical drive circuit 4, a column signal processing circuit 5, a horizontal drive circuit 6, an output circuit 7, and a control circuit 8.
- the control circuit 8 receives data for instructing an input clock, an operation mode, and the like, and outputs data such as internal information of the solid-state imaging device 1. Specifically, the control circuit 8 is based on the vertical synchronization signal, the horizontal synchronization signal, and the master clock, and the clock signal or the reference signal for the operations of the vertical drive circuit 4, the column signal processing circuit 5, and the horizontal drive circuit 6 Generate a control signal. The control circuit 8 inputs these signals to the vertical drive circuit 4, the column signal processing circuit 5, and the horizontal drive circuit 6.
- the vertical drive circuit 4 is composed of, for example, a shift register, selects a pixel drive wiring, supplies a pulse for driving the pixel 2 to the selected pixel drive wiring, and drives the pixels 2 in units of rows. Specifically, the vertical drive circuit 4 selectively scans each pixel 2 in the pixel region 3 sequentially in the vertical direction in units of rows, and generates the signal according to the amount of light received by the photoelectric conversion element of each pixel 2 through the vertical signal line 9. A pixel signal based on the signal charge is supplied to the column signal processing circuit 5.
- the column signal processing circuit 5 is disposed, for example, for each column of the pixels 2 and performs signal processing such as noise removal on the signal output from the pixels 2 for one row for each pixel column. Specifically, the column signal processing circuit 5 performs signal processing such as CDS (Correlated Double Sampling) for removing fixed pattern noise specific to the pixel 2, signal amplification, A / D (Analog / Digital) conversion, and the like. .
- a horizontal selection switch (not shown) is provided connected to the horizontal signal line 10. A part of the signal processing may be subjected to signal processing for each pixel.
- the horizontal drive circuit 6 is constituted by, for example, a shift register, and sequentially outputs horizontal scanning pulses to select each of the column signal processing circuits 5 in order, and the pixel signal is output from each of the column signal processing circuits 5 to the horizontal signal line. 10 to output.
- the output circuit 7 performs signal processing on the signals sequentially supplied from each of the column signal processing circuits 5 through the horizontal signal line 10 and outputs the signals.
- the output circuit 7 may perform only buffering, or may perform black level adjustment, column variation correction, various digital signal processing, and the like.
- the input / output terminal 12 is provided for exchanging signals with the outside.
- FIG. 2 is a cross-sectional view illustrating a pixel structure example of the first embodiment of the solid-state imaging device to which the present technology is applied.
- the photoelectric conversion element 109 provided on the first surface 11-1 of the semiconductor substrate 11, and the first surface and the second surface of the semiconductor substrate 11 connected to the photoelectric conversion element 109 are connected.
- a through electrode 104 provided therebetween, an amplifier transistor 115 and a floating diffusion 116 provided on the second surface 11-2 of the semiconductor substrate 11 are provided.
- the photoelectric conversion element 109 is connected to the gate of the amplifier transistor 115 and the floating diffusion 116 via the through electrode 104.
- the dielectric layer 103 is provided between the through electrodes 104 of the second surface 11-2, and the shield electrode is formed inside the dielectric layer 103 when viewed from the second surface 11-2 side.
- 114 is provided.
- the dielectric layer 103 is formed thicker than the gate insulating film 120 of the transistor (for example, the amplifier transistor 115) disposed on the second surface 11-2 side.
- the shield electrode 114 is controlled to a predetermined voltage from a contact 128 electrically connected to the shield electrode 114 and a wiring electrically connected to the contact 128.
- the dielectric layer 103 is processed due to the feature that the etching rate of the dielectric layer 103 is lower than that of the semiconductor substrate 11 when the through electrode is formed from the first surface 11-1 of the semiconductor substrate 11.
- the diameter of the through electrode 104 is defined by the dielectric layer 103 in a self-aligning manner.
- a separation groove 105 that separates the through electrode 104 and the semiconductor substrate 11 is formed, and the separation groove 105 is filled with the dielectric 101. It is preferable to take. Impurity regions (P + in FIG. 2) are preferably provided in the semiconductor substrate on the outer surface of the separation groove 105 and the outer surface of the dielectric layer 103.
- the hole accumulation layer can be formed on the outer surface of the dielectric layer 103 of the semiconductor substrate 11 even under a condition where the shield electrode 114 is applied with a positive bias.
- the thickness of the dielectric layer 103 is preferably 20 nm or more.
- a shield electrode 114 is provided inside the dielectric layer 103 when viewed from the second surface 11-2 side.
- the shield electrode 114 is positioned at a distance of 20 nm or more from the end of the dielectric layer 103 to the inside. Is preferably arranged.
- the dielectric layer 103 may be disposed between the through electrodes 104 such that the distance between the separation groove 105 and the dielectric layer 103 is closer than 30% of the combined diameter of the through electrode 104 and the separation groove 105.
- the distance between the separation groove 105 and the dielectric layer 103 is preferably 100 nm or less.
- the material of the dielectric layer 103 is not particularly limited, and includes, for example, a silicon oxide film, a silicon nitride film, and a silicon oxynitride film.
- the shield electrode 114 is made of, for example, a doped silicon material such as polysilicon or PDAS (Phosphorus Doped Amorphous Silicon).
- the material forming the through electrode 104 is not particularly limited, but is preferably a metal such as aluminum, tungsten, titanium, cobalt, hafnium, or tantalum, or a conductive material.
- an on-chip lens or a color filter may be formed on the light incident surface 112 as necessary.
- This manufacturing process is a process executed by the manufacturing apparatus, and the process diagrams of FIGS. 5 to 10 are referred to as appropriate.
- step S111 the manufacturing apparatus forms, for example, a P well as the first conductivity type well in the semiconductor substrate 11, and the second conductivity type (for example, N-type photodiodes 123 and 124 are formed. At this time, a P + region is formed in the vicinity of the first surface 11-1 of the semiconductor substrate 11.
- step S112 the manufacturing apparatus penetrates from the first surface 11-1 to the second surface 11-2 of the semiconductor substrate 11 in the planned formation area 151 of the through electrode 104 and the separation groove 105, as shown in FIG.
- An impurity region (P + region) to be formed is formed.
- the manufacturing apparatus forms the dielectric layer 103 on the second surface 11-2 side between the formation scheduled areas 151.
- the dielectric layer 103 may be formed at the same time as element isolation disposed on the second surface 11-2 side, for example, STI.
- step S113 the manufacturing apparatus formed, on the second surface 11-2 of the semiconductor substrate 11, N + regions serving as floating diffusions (hereinafter referred to as FD) 116 to 118 as shown in FIG. After that, the gate insulating film 120 and the gate wiring including the gates of the vertical transfer gate 122, the transfer gate 121, the amplifier transistor 115, and the reset transistor 129 are formed.
- FD floating diffusions
- step S114 the manufacturing apparatus forms a multilayer wiring 127 including the contact 128, the wiring layer 125, and the insulating film 126 on the second surface 11-2 of the semiconductor substrate 11.
- an SOI (Silicon On Insulator) substrate in which a semiconductor substrate 11, a buried oxide film (not shown), and a holding substrate (not shown) are stacked is used.
- the buried oxide film and the holding substrate are bonded to the first surface 11-1 of the semiconductor substrate 11. After ion implantation, an annealing process is performed.
- step S115 the manufacturing apparatus joins a support substrate (not shown) or another semiconductor substrate to the second surface 11-2 side (multilayer wiring) of the semiconductor substrate 11, as shown in FIG. Then, the solid-state imaging device 1 is turned upside down. Then, the semiconductor substrate 11 is separated from the buried oxide film of the SOI substrate and the holding substrate, and the first surface 11-1 of the semiconductor substrate 11 is exposed.
- CMOS complementary metal-oxide-semiconductor
- CVD Chemical Vapor Deposition
- step S116 the manufacturing apparatus processes the semiconductor substrate 11 from the first surface 11-1 side by dry etching, for example, and separates the ring or ring into the semiconductor substrate 11. A groove 105 is formed.
- step S117 the manufacturing apparatus places a film 102 having a fixed charge and a dielectric 101 on the outer and bottom surfaces of the separation groove 105 and the first surface 11-1 of the semiconductor substrate 11, as shown in FIG. Form a film.
- a material of the dielectric 101 it is possible to use a silicon oxide film or a silicon nitride film formed by TEOS or ALD.
- step S118 of FIG. 4 the manufacturing apparatus retracts the film 102 having a fixed charge and the dielectric 101 by dry etching or the like.
- step S119 the manufacturing apparatus embeds the metal material film in the separation groove 105, and then recedes or planarizes the metal material film by dry etching or CMP (Chemical-Mechanical-Polishing).
- CMP Chemical-Mechanical-Polishing
- step S 120 the manufacturing apparatus forms the interlayer insulating film 111 and the upper contact as shown in FIG. 2, and connects the upper contact to the upper end of the through electrode 104.
- step S121 the manufacturing apparatus forms the photoelectric conversion element 109 including the transparent electrode 106, the photoelectric conversion film 107, and the transparent electrode 108, and the protective film 110, and finally, an optical member such as a planarization film.
- An on-chip lens (not shown) is disposed.
- FIG. 11 is a top view illustrating a pixel structure example of the second embodiment of the solid-state imaging device to which the present technology is applied.
- the pixel structure of FIG. 11 is different from the pixel structure of FIG. 2 only in that the contact 128 is provided between the through-electrodes 104, and the others are basically the same as the pixel structure of FIG. Yes. That is, the pixel structure shown in FIG. 11 is characterized in that the contact 128 electrically connected to the shield electrode 114 is provided between the through electrodes 104.
- capacitive coupling between the through electrodes 104 can be further prevented as compared with the pixel structure of FIG.
- the material of the contact 128 is not particularly limited, but includes a metal material such as aluminum, tungsten, titanium, cobalt, hafnium, and tantalum.
- the corners of the dielectric layer 103 and the shield electrode 114 may be rounded.
- FIG. 12 is a top view illustrating a pixel structure example of the third embodiment of the solid-state imaging device to which the present technology is applied.
- the pixel structure of FIG. 12 is different from the pixel structure of FIG. 11 only in that the contact 128 is provided in a line shape between the through electrodes 104, and the rest is basically the same as the pixel structure of FIG. It is common.
- the pixel structure shown in FIG. 12 is characterized in that the contact 128 is provided in a line shape between the through electrodes 104.
- FIG. 13 is a top view illustrating a pixel structure example of the third embodiment of the solid-state imaging device to which the present technology is applied.
- FIG. 14 is a cross-sectional view of the pixel structure of FIG.
- the pixel structure of FIG. 13 is different from the pixel structure of FIG. 2 only in that the dielectric layer 103 is provided around the through electrode 104, and is otherwise basically the same as the pixel structure of FIG. is doing.
- the pixel structure shown in FIG. 13 is characterized in that the dielectric layer 103 is provided around the through electrode 104.
- FIG. 15 is a cross-sectional view illustrating a pixel structure example of a fifth embodiment of a solid-state imaging device to which the present technology is applied.
- the through electrode 104 penetrates the semiconductor substrate 11 and is separated from the semiconductor substrate 11 by the separation groove 105, and the dielectric layer 103 and the separation groove 105 are in contact with each other.
- 2 is basically the same as the pixel structure of FIG. That is, in the pixel structure shown in FIG. 15, the through electrode 104 penetrates the semiconductor substrate 11 and is separated from the semiconductor substrate 11 by the separation groove 105, and the dielectric layer 103 and the separation groove 105 are in contact with each other. Is characterized.
- the pixel size can be further reduced as compared with the pixel structure of FIG.
- the dielectric layer is difficult to be processed because the etching rate of the dielectric layer is lower than that of the semiconductor substrate.
- the diameter of the through electrode is defined in a self-aligned manner by the dielectric layer. Therefore, in the present technology, since the shield electrode is disposed on the dielectric layer, it is difficult for the through electrode and the shield electrode to be short-circuited.
- a P-type impurity region is provided in the semiconductor substrate on the outer surface of the dielectric layer. Furthermore, by providing a dielectric layer under the shield electrode, a hole accumulation layer can be formed in the semiconductor substrate on the outer surface of the dielectric layer under the condition that the shield electrode is applied with a positive bias.
- FIG. 16 is a diagram illustrating a usage example in which the above-described solid-state imaging device is used.
- the solid-state imaging device (image sensor) described above can be used in various cases for sensing light such as visible light, infrared light, ultraviolet light, and X-ray as follows.
- Devices for taking images for viewing such as digital cameras and mobile devices with camera functions
- Devices used for traffic such as in-vehicle sensors that capture the back, surroundings, and interiors of vehicles, surveillance cameras that monitor traveling vehicles and roads, and ranging sensors that measure distances between vehicles, etc.
- Equipment used for home appliances such as TVs, refrigerators, air conditioners, etc. to take pictures and operate the equipment according to the gestures ⁇ Endoscopes, equipment that performs blood vessel photography by receiving infrared light, etc.
- Equipment used for medical and health care ⁇ Security equipment such as security surveillance cameras and personal authentication cameras ⁇ Skin measuring instrument for photographing skin and scalp photography Such as a microscope to do beauty Equipment used for sports such as action cameras and wearable cameras for sports applications etc.
- Equipment used for agriculture such as cameras for monitoring the condition of fields and crops
- the present technology is not limited to application to a solid-state imaging device, but can also be applied to an imaging apparatus.
- the imaging apparatus refers to a camera system such as a digital still camera or a digital video camera, or an electronic apparatus having an imaging function such as a mobile phone.
- a module-like form mounted on an electronic device that is, a camera module is used as an imaging device.
- the 17 includes a solid-state image sensor (element chip) 301, an optical lens 302, a shutter device 303, a drive circuit 304, and a signal processing circuit 305.
- the solid-state imaging device 301 the above-described solid-state imaging device 1 of the present technology is provided.
- the optical lens 302 forms image light (incident light) from the subject on the imaging surface of the solid-state imaging device 301. Thereby, signal charges are accumulated in the solid-state imaging device 301 for a certain period.
- the shutter device 303 controls a light irradiation period and a light shielding period with respect to the solid-state imaging element 301.
- the drive circuit 304 supplies drive signals for controlling the signal transfer operation of the solid-state imaging device 301, the shutter operation of the shutter device 303, and the light emission operation of a light emitting unit (not shown).
- the drive circuit 304 controls each operation using parameters set by a CPU (not shown).
- the solid-state imaging device 301 performs signal transfer by a drive signal (timing signal) supplied from the drive circuit 304.
- the signal processing circuit 305 performs various types of signal processing on the signal output from the solid-state image sensor 301.
- the video signal subjected to the signal processing is stored in a storage medium such as a memory or output to a monitor.
- the technology according to the present disclosure can be applied to various products.
- the technology according to the present disclosure may be applied to an endoscopic surgery system.
- FIG. 18 is a block diagram illustrating an example of a schematic configuration of a patient in-vivo information acquisition system using a capsule endoscope to which the technique (present technique) according to the present disclosure can be applied.
- the in-vivo information acquisition system 10001 includes a capsule endoscope 10100 and an external control device 10200.
- the capsule endoscope 10100 is swallowed by the patient at the time of examination.
- the capsule endoscope 10100 has an imaging function and a wireless communication function, and moves inside the organ such as the stomach and the intestine by peristaltic motion or the like until it is spontaneously discharged from the patient.
- Images (hereinafter also referred to as in-vivo images) are sequentially captured at predetermined intervals, and information about the in-vivo images is sequentially wirelessly transmitted to the external control device 10200 outside the body.
- the external control device 10200 comprehensively controls the operation of the in-vivo information acquisition system 10001. Further, the external control device 10200 receives information about the in-vivo image transmitted from the capsule endoscope 10100 and, based on the received information about the in-vivo image, displays the in-vivo image on the display device (not shown). The image data for displaying is generated.
- an in-vivo image obtained by imaging the inside of the patient's body can be obtained at any time in this manner until the capsule endoscope 10100 is swallowed and discharged.
- the capsule endoscope 10100 includes a capsule-type casing 10101.
- a light source unit 10111 In the casing 10101, a light source unit 10111, an imaging unit 10112, an image processing unit 10113, a wireless communication unit 10114, a power supply unit 10115, and a power supply unit 10116 and the control unit 10117 are stored.
- the light source unit 10111 is composed of a light source such as an LED (Light Emitting Diode), for example, and irradiates the imaging field of the imaging unit 10112 with light.
- a light source such as an LED (Light Emitting Diode), for example, and irradiates the imaging field of the imaging unit 10112 with light.
- the image capturing unit 10112 includes an image sensor and an optical system including a plurality of lenses provided in front of the image sensor. Reflected light (hereinafter referred to as observation light) of light irradiated on the body tissue to be observed is collected by the optical system and enters the image sensor. In the imaging unit 10112, in the imaging element, the observation light incident thereon is photoelectrically converted, and an image signal corresponding to the observation light is generated. The image signal generated by the imaging unit 10112 is provided to the image processing unit 10113.
- the image processing unit 10113 is configured by a processor such as a CPU (Central Processing Unit) or a GPU (Graphics Processing Unit), and performs various signal processing on the image signal generated by the imaging unit 10112.
- the image processing unit 10113 provides the radio communication unit 10114 with the image signal subjected to signal processing as RAW data.
- the wireless communication unit 10114 performs predetermined processing such as modulation processing on the image signal that has been subjected to signal processing by the image processing unit 10113, and transmits the image signal to the external control apparatus 10200 via the antenna 10114A.
- the wireless communication unit 10114 receives a control signal related to drive control of the capsule endoscope 10100 from the external control device 10200 via the antenna 10114A.
- the wireless communication unit 10114 provides a control signal received from the external control device 10200 to the control unit 10117.
- the power feeding unit 10115 includes a power receiving antenna coil, a power regeneration circuit that regenerates power from a current generated in the antenna coil, a booster circuit, and the like. In the power feeding unit 10115, electric power is generated using a so-called non-contact charging principle.
- the power supply unit 10116 is composed of a secondary battery, and stores the electric power generated by the power supply unit 10115.
- FIG. 19 in order to avoid complication of the drawing, illustration of an arrow or the like indicating a power supply destination from the power supply unit 10116 is omitted.
- the imaging unit 10112, the image processing unit 10113, the wireless communication unit 10114, and the control unit 10117 can be used for driving them.
- the control unit 10117 includes a processor such as a CPU, and a control signal transmitted from the external control device 10200 to drive the light source unit 10111, the imaging unit 10112, the image processing unit 10113, the wireless communication unit 10114, and the power feeding unit 10115. Control accordingly.
- a processor such as a CPU
- the external control device 10200 is configured by a processor such as a CPU or GPU, or a microcomputer or a control board in which a processor and a storage element such as a memory are mounted.
- the external control device 10200 controls the operation of the capsule endoscope 10100 by transmitting a control signal to the control unit 10117 of the capsule endoscope 10100 via the antenna 10200A.
- the capsule endoscope 10100 for example, the light irradiation condition for the observation target in the light source unit 10111 can be changed by a control signal from the external control device 10200.
- an imaging condition for example, a frame rate or an exposure value in the imaging unit 10112
- a control signal from the external control device 10200 can be changed by a control signal from the external control device 10200.
- the contents of processing in the image processing unit 10113 and the conditions (for example, the transmission interval, the number of transmission images, etc.) by which the wireless communication unit 10114 transmits an image signal may be changed by a control signal from the external control device 10200. .
- the external control device 10200 performs various image processing on the image signal transmitted from the capsule endoscope 10100, and generates image data for displaying the captured in-vivo image on the display device.
- image processing for example, development processing (demosaic processing), high image quality processing (band enhancement processing, super-resolution processing, NR (Noise reduction) processing and / or camera shake correction processing, etc.), and / or enlargement processing ( Various signal processing such as electronic zoom processing can be performed.
- the external control device 10200 controls driving of the display device to display an in-vivo image captured based on the generated image data.
- the external control device 10200 may cause the generated image data to be recorded on a recording device (not shown) or may be printed out on a printing device (not shown).
- the technology according to the present disclosure can be applied to the imaging unit 10112 among the configurations described above.
- the solid-state imaging device 1 in FIG. 1 can be applied to the imaging unit 10112.
- the element layout area can be further reduced, and the pixel size can be further reduced, and the saturation signal amount and sensitivity characteristics can be improved. Be expected.
- the technology according to the present disclosure can be applied to various products.
- the technology according to the present disclosure may be applied to an endoscopic surgery system.
- FIG. 19 is a diagram illustrating an example of a schematic configuration of an endoscopic surgery system to which the technology (present technology) according to the present disclosure can be applied.
- FIG. 19 shows a state where an operator (doctor) 11131 is performing an operation on a patient 11132 on a patient bed 11133 using an endoscopic operation system 11000.
- an endoscopic surgery system 11000 includes an endoscope 11100, other surgical instruments 11110 such as an insufflation tube 11111 and an energy treatment instrument 11112, and a support arm device 11120 that supports the endoscope 11100. And a cart 11200 on which various devices for endoscopic surgery are mounted.
- the endoscope 11100 includes a lens barrel 11101 in which a region having a predetermined length from the distal end is inserted into the body cavity of the patient 11132, and a camera head 11102 connected to the proximal end of the lens barrel 11101.
- a lens barrel 11101 in which a region having a predetermined length from the distal end is inserted into the body cavity of the patient 11132, and a camera head 11102 connected to the proximal end of the lens barrel 11101.
- an endoscope 11100 configured as a so-called rigid mirror having a rigid lens barrel 11101 is illustrated, but the endoscope 11100 may be configured as a so-called flexible mirror having a flexible lens barrel. Good.
- An opening into which the objective lens is fitted is provided at the tip of the lens barrel 11101.
- a light source device 11203 is connected to the endoscope 11100, and light generated by the light source device 11203 is guided to the tip of the lens barrel by a light guide extending inside the lens barrel 11101. Irradiation is performed toward the observation target in the body cavity of the patient 11132 through the lens.
- the endoscope 11100 may be a direct endoscope, a perspective mirror, or a side endoscope.
- An optical system and an image sensor are provided inside the camera head 11102, and reflected light (observation light) from the observation target is condensed on the image sensor by the optical system. Observation light is photoelectrically converted by the imaging element, and an electrical signal corresponding to the observation light, that is, an image signal corresponding to the observation image is generated.
- the image signal is transmitted to a camera control unit (CCU: Camera Control Unit) 11201 as RAW data.
- CCU Camera Control Unit
- the CCU 11201 is configured by a CPU (Central Processing Unit), a GPU (Graphics Processing Unit), and the like, and comprehensively controls operations of the endoscope 11100 and the display device 11202. Further, the CCU 11201 receives an image signal from the camera head 11102 and performs various kinds of image processing for displaying an image based on the image signal, such as development processing (demosaic processing), for example.
- a CPU Central Processing Unit
- GPU Graphics Processing Unit
- the display device 11202 displays an image based on an image signal subjected to image processing by the CCU 11201 under the control of the CCU 11201.
- the light source device 11203 is composed of a light source such as an LED (Light Emitting Diode), for example, and supplies irradiation light to the endoscope 11100 when photographing a surgical site or the like.
- a light source such as an LED (Light Emitting Diode), for example, and supplies irradiation light to the endoscope 11100 when photographing a surgical site or the like.
- the input device 11204 is an input interface for the endoscopic surgery system 11000.
- a user can input various information and instructions to the endoscopic surgery system 11000 via the input device 11204.
- the user inputs an instruction to change the imaging conditions (type of irradiation light, magnification, focal length, etc.) by the endoscope 11100.
- the treatment instrument control device 11205 controls the drive of the energy treatment instrument 11112 for tissue ablation, incision, blood vessel sealing, or the like.
- the pneumoperitoneum device 11206 passes gas into the body cavity via the pneumoperitoneum tube 11111.
- the recorder 11207 is an apparatus capable of recording various types of information related to surgery.
- the printer 11208 is a device that can print various types of information related to surgery in various formats such as text, images, or graphs.
- the light source device 11203 that supplies the irradiation light when the surgical site is imaged to the endoscope 11100 can be configured by, for example, a white light source configured by an LED, a laser light source, or a combination thereof.
- a white light source is configured by a combination of RGB laser light sources, the output intensity and output timing of each color (each wavelength) can be controlled with high accuracy. Therefore, the light source device 11203 adjusts the white balance of the captured image. It can be carried out.
- the driving of the light source device 11203 may be controlled so as to change the intensity of the output light every predetermined time. Synchronously with the timing of changing the intensity of the light, the drive of the image sensor of the camera head 11102 is controlled to acquire an image in a time-sharing manner, and the image is synthesized, so that high dynamic without so-called blackout and overexposure A range image can be generated.
- the light source device 11203 may be configured to be able to supply light of a predetermined wavelength band corresponding to special light observation.
- special light observation for example, by utilizing the wavelength dependence of light absorption in body tissue, the surface of the mucous membrane is irradiated by irradiating light in a narrow band compared to irradiation light (ie, white light) during normal observation.
- a so-called narrow band imaging is performed in which a predetermined tissue such as a blood vessel is imaged with high contrast.
- fluorescence observation may be performed in which an image is obtained by fluorescence generated by irradiating excitation light.
- the body tissue is irradiated with excitation light to observe fluorescence from the body tissue (autofluorescence observation), or a reagent such as indocyanine green (ICG) is locally administered to the body tissue and applied to the body tissue. It is possible to obtain a fluorescence image by irradiating excitation light corresponding to the fluorescence wavelength of the reagent.
- the light source device 11203 can be configured to be able to supply narrowband light and / or excitation light corresponding to such special light observation.
- FIG. 20 is a block diagram illustrating an example of functional configurations of the camera head 11102 and the CCU 11201 illustrated in FIG.
- the camera head 11102 includes a lens unit 11401, an imaging unit 11402, a drive unit 11403, a communication unit 11404, and a camera head control unit 11405.
- the CCU 11201 includes a communication unit 11411, an image processing unit 11412, and a control unit 11413.
- the camera head 11102 and the CCU 11201 are connected to each other by a transmission cable 11400 so that they can communicate with each other.
- the lens unit 11401 is an optical system provided at a connection portion with the lens barrel 11101. Observation light taken from the tip of the lens barrel 11101 is guided to the camera head 11102 and enters the lens unit 11401.
- the lens unit 11401 is configured by combining a plurality of lenses including a zoom lens and a focus lens.
- the imaging unit 11402 includes an imaging element.
- One (so-called single plate type) image sensor may be included in the imaging unit 11402, or a plurality (so-called multi-plate type) may be used.
- image signals corresponding to RGB may be generated by each imaging element, and a color image may be obtained by combining them.
- the imaging unit 11402 may be configured to include a pair of imaging elements for acquiring right-eye and left-eye image signals corresponding to 3D (Dimensional) display. By performing the 3D display, the operator 11131 can more accurately grasp the depth of the living tissue in the surgical site.
- 3D 3D
- the imaging unit 11402 is not necessarily provided in the camera head 11102.
- the imaging unit 11402 may be provided inside the lens barrel 11101 immediately after the objective lens.
- the driving unit 11403 is configured by an actuator, and moves the zoom lens and the focus lens of the lens unit 11401 by a predetermined distance along the optical axis under the control of the camera head control unit 11405. Thereby, the magnification and the focus of the image captured by the imaging unit 11402 can be adjusted as appropriate.
- the communication unit 11404 is configured by a communication device for transmitting and receiving various types of information to and from the CCU 11201.
- the communication unit 11404 transmits the image signal obtained from the imaging unit 11402 as RAW data to the CCU 11201 via the transmission cable 11400.
- the communication unit 11404 receives a control signal for controlling driving of the camera head 11102 from the CCU 11201 and supplies the control signal to the camera head control unit 11405.
- the control signal includes, for example, information for designating the frame rate of the captured image, information for designating the exposure value at the time of imaging, and / or information for designating the magnification and focus of the captured image. Contains information about the condition.
- the imaging conditions such as the frame rate, exposure value, magnification, and focus may be appropriately specified by the user, or may be automatically set by the control unit 11413 of the CCU 11201 based on the acquired image signal. Good. In the latter case, a so-called AE (Auto Exposure) function, AF (Auto Focus) function, and AWB (Auto White Balance) function are mounted on the endoscope 11100.
- AE Auto Exposure
- AF Automatic Focus
- AWB Auto White Balance
- the camera head control unit 11405 controls driving of the camera head 11102 based on a control signal from the CCU 11201 received via the communication unit 11404.
- the communication unit 11411 is configured by a communication device for transmitting and receiving various types of information to and from the camera head 11102.
- the communication unit 11411 receives an image signal transmitted from the camera head 11102 via the transmission cable 11400.
- the communication unit 11411 transmits a control signal for controlling driving of the camera head 11102 to the camera head 11102.
- the image signal and the control signal can be transmitted by electrical communication, optical communication, or the like.
- the image processing unit 11412 performs various types of image processing on the image signal that is RAW data transmitted from the camera head 11102.
- the control unit 11413 performs various types of control related to imaging of the surgical site by the endoscope 11100 and display of a captured image obtained by imaging of the surgical site. For example, the control unit 11413 generates a control signal for controlling driving of the camera head 11102.
- control unit 11413 causes the display device 11202 to display a picked-up image showing the surgical part or the like based on the image signal subjected to the image processing by the image processing unit 11412.
- the control unit 11413 may recognize various objects in the captured image using various image recognition techniques.
- the control unit 11413 detects surgical tools such as forceps, specific biological parts, bleeding, mist when using the energy treatment tool 11112, and the like by detecting the shape and color of the edge of the object included in the captured image. Can be recognized.
- the control unit 11413 may display various types of surgery support information superimposed on the image of the surgical unit using the recognition result. Surgery support information is displayed in a superimposed manner and presented to the operator 11131, thereby reducing the burden on the operator 11131 and allowing the operator 11131 to proceed with surgery reliably.
- the transmission cable 11400 for connecting the camera head 11102 and the CCU 11201 is an electric signal cable corresponding to electric signal communication, an optical fiber corresponding to optical communication, or a composite cable thereof.
- communication is performed by wire using the transmission cable 11400.
- communication between the camera head 11102 and the CCU 11201 may be performed wirelessly.
- the technology according to the present disclosure can be applied to the endoscope 11100 and the camera head 11102 (the imaging unit 11402 thereof) among the configurations described above.
- the solid-state imaging device 1 in FIG. 1 includes an endoscope 11100 and a camera head 11102 (the imaging unit 11402).
- the technique according to the present disclosure to the endoscope 11100 and the camera head 11102 (the imaging unit 11402), the element layout area can be further reduced, the pixel size can be further reduced, and saturation can be achieved. Expected to improve signal volume and sensitivity characteristics.
- the technology according to the present disclosure can be applied to various products.
- the technology according to the present disclosure is realized as a device that is mounted on any type of mobile body such as an automobile, an electric vehicle, a hybrid electric vehicle, a motorcycle, a bicycle, personal mobility, an airplane, a drone, a ship, and a robot. May be.
- FIG. 21 is a block diagram illustrating a schematic configuration example of a vehicle control system that is an example of a mobile control system to which the technology according to the present disclosure can be applied.
- the vehicle control system 12000 includes a plurality of electronic control units connected via a communication network 12001.
- the vehicle control system 12000 includes a drive system control unit 12010, a body system control unit 12020, a vehicle exterior information detection unit 12030, a vehicle interior information detection unit 12040, and an integrated control unit 12050.
- a microcomputer 12051, an audio image output unit 12052, and an in-vehicle network I / F (interface) 12053 are illustrated.
- the drive system control unit 12010 controls the operation of the device related to the drive system of the vehicle according to various programs.
- the drive system control unit 12010 includes a driving force generator for generating a driving force of a vehicle such as an internal combustion engine or a driving motor, a driving force transmission mechanism for transmitting the driving force to wheels, and a steering angle of the vehicle. It functions as a control device such as a steering mechanism that adjusts and a braking device that generates a braking force of the vehicle.
- the body system control unit 12020 controls the operation of various devices mounted on the vehicle body according to various programs.
- the body system control unit 12020 functions as a keyless entry system, a smart key system, a power window device, or a control device for various lamps such as a headlamp, a back lamp, a brake lamp, a blinker, or a fog lamp.
- the body control unit 12020 can be input with radio waves transmitted from a portable device that substitutes for a key or signals from various switches.
- the body system control unit 12020 receives input of these radio waves or signals, and controls a door lock device, a power window device, a lamp, and the like of the vehicle.
- the vehicle outside information detection unit 12030 detects information outside the vehicle on which the vehicle control system 12000 is mounted.
- the imaging unit 12031 is connected to the vehicle exterior information detection unit 12030.
- the vehicle exterior information detection unit 12030 causes the imaging unit 12031 to capture an image outside the vehicle and receives the captured image.
- the vehicle outside information detection unit 12030 may perform an object detection process or a distance detection process such as a person, a car, an obstacle, a sign, or a character on a road surface based on the received image.
- the imaging unit 12031 is an optical sensor that receives light and outputs an electrical signal corresponding to the amount of received light.
- the imaging unit 12031 can output an electrical signal as an image, or can output it as distance measurement information. Further, the light received by the imaging unit 12031 may be visible light or invisible light such as infrared rays.
- the vehicle interior information detection unit 12040 detects vehicle interior information.
- a driver state detection unit 12041 that detects a driver's state is connected to the in-vehicle information detection unit 12040.
- the driver state detection unit 12041 includes, for example, a camera that images the driver, and the vehicle interior information detection unit 12040 determines the degree of fatigue or concentration of the driver based on the detection information input from the driver state detection unit 12041. It may be calculated or it may be determined whether the driver is asleep.
- the microcomputer 12051 calculates a control target value of the driving force generator, the steering mechanism, or the braking device based on the information inside / outside the vehicle acquired by the vehicle outside information detection unit 12030 or the vehicle interior information detection unit 12040, and the drive system control unit A control command can be output to 12010.
- the microcomputer 12051 realizes an ADAS (Advanced Driver Assistance System) function including vehicle collision avoidance or impact mitigation, following traveling based on inter-vehicle distance, vehicle speed maintaining traveling, vehicle collision warning, or vehicle lane departure warning, etc. It is possible to perform cooperative control for the purpose.
- ADAS Advanced Driver Assistance System
- the microcomputer 12051 controls the driving force generator, the steering mechanism, the braking device, and the like based on the information around the vehicle acquired by the vehicle exterior information detection unit 12030 or the vehicle interior information detection unit 12040. It is possible to perform cooperative control for the purpose of automatic driving that autonomously travels without depending on the operation.
- the microcomputer 12051 can output a control command to the body system control unit 12020 based on information outside the vehicle acquired by the vehicle outside information detection unit 12030.
- the microcomputer 12051 controls the headlamp according to the position of the preceding vehicle or the oncoming vehicle detected by the outside information detection unit 12030, and performs cooperative control for the purpose of anti-glare, such as switching from a high beam to a low beam. It can be carried out.
- the sound image output unit 12052 transmits an output signal of at least one of sound and image to an output device capable of visually or audibly notifying information to a vehicle occupant or the outside of the vehicle.
- an audio speaker 12061, a display unit 12062, and an instrument panel 12063 are illustrated as output devices.
- the display unit 12062 may include at least one of an on-board display and a head-up display, for example.
- FIG. 22 is a diagram illustrating an example of an installation position of the imaging unit 12031.
- the vehicle 12100 includes imaging units 12101, 12102, 12103, 12104, and 12105 as the imaging unit 12031.
- the imaging units 12101, 12102, 12103, 12104, and 12105 are provided, for example, at positions such as a front nose, a side mirror, a rear bumper, a back door, and an upper part of a windshield in the vehicle interior of the vehicle 12100.
- the imaging unit 12101 provided in the front nose and the imaging unit 12105 provided in the upper part of the windshield in the vehicle interior mainly acquire an image in front of the vehicle 12100.
- the imaging units 12102 and 12103 provided in the side mirror mainly acquire an image of the side of the vehicle 12100.
- the imaging unit 12104 provided in the rear bumper or the back door mainly acquires an image behind the vehicle 12100.
- the forward images acquired by the imaging units 12101 and 12105 are mainly used for detecting a preceding vehicle or a pedestrian, an obstacle, a traffic light, a traffic sign, a lane, or the like.
- FIG. 22 shows an example of the shooting range of the imaging units 12101 to 12104.
- the imaging range 12111 indicates the imaging range of the imaging unit 12101 provided in the front nose
- the imaging ranges 12112 and 12113 indicate the imaging ranges of the imaging units 12102 and 12103 provided in the side mirrors, respectively
- the imaging range 12114 The imaging range of the imaging part 12104 provided in the rear bumper or the back door is shown. For example, by superimposing the image data captured by the imaging units 12101 to 12104, an overhead image when the vehicle 12100 is viewed from above is obtained.
- At least one of the imaging units 12101 to 12104 may have a function of acquiring distance information.
- at least one of the imaging units 12101 to 12104 may be a stereo camera including a plurality of imaging elements, or may be an imaging element having pixels for phase difference detection.
- the microcomputer 12051 based on the distance information obtained from the imaging units 12101 to 12104, the distance to each three-dimensional object in the imaging range 12111 to 12114 and the temporal change in this distance (relative speed with respect to the vehicle 12100).
- a predetermined speed for example, 0 km / h or more
- the microcomputer 12051 can set an inter-vehicle distance to be secured in advance before the preceding vehicle, and can perform automatic brake control (including follow-up stop control), automatic acceleration control (including follow-up start control), and the like.
- automatic brake control including follow-up stop control
- automatic acceleration control including follow-up start control
- cooperative control for the purpose of autonomous driving or the like autonomously traveling without depending on the operation of the driver can be performed.
- the microcomputer 12051 converts the three-dimensional object data related to the three-dimensional object to other three-dimensional objects such as a two-wheeled vehicle, a normal vehicle, a large vehicle, a pedestrian, and a utility pole based on the distance information obtained from the imaging units 12101 to 12104. It can be classified and extracted and used for automatic avoidance of obstacles.
- the microcomputer 12051 identifies obstacles around the vehicle 12100 as obstacles that are visible to the driver of the vehicle 12100 and obstacles that are difficult to see.
- the microcomputer 12051 determines the collision risk indicating the risk of collision with each obstacle, and when the collision risk is equal to or higher than the set value and there is a possibility of collision, the microcomputer 12051 is connected via the audio speaker 12061 or the display unit 12062. By outputting an alarm to the driver and performing forced deceleration or avoidance steering via the drive system control unit 12010, driving assistance for collision avoidance can be performed.
- At least one of the imaging units 12101 to 12104 may be an infrared camera that detects infrared rays.
- the microcomputer 12051 can recognize a pedestrian by determining whether a pedestrian is present in the captured images of the imaging units 12101 to 12104. Such pedestrian recognition is, for example, whether or not the user is a pedestrian by performing a pattern matching process on a sequence of feature points indicating the outline of an object and a procedure for extracting feature points in the captured images of the imaging units 12101 to 12104 as infrared cameras. It is carried out by the procedure for determining.
- the audio image output unit 12052 When the microcomputer 12051 determines that there is a pedestrian in the captured images of the imaging units 12101 to 12104 and recognizes the pedestrian, the audio image output unit 12052 has a rectangular contour line for emphasizing the recognized pedestrian.
- the display unit 12062 is controlled so as to be superimposed and displayed.
- voice image output part 12052 may control the display part 12062 so that the icon etc. which show a pedestrian may be displayed on a desired position.
- the technology according to the present disclosure can be applied to the imaging unit 12031 (including the imaging units 12101 to 12104) among the configurations described above.
- the solid-state imaging device 1 in FIG. 1 can be applied to the imaging unit 12031 (including the imaging units 12101 to 12104).
- FIG. 23 is a cross-sectional view illustrating a configuration example of a solid-state imaging device having pixels on which photoelectric conversion units are stacked, to which the technology according to the present disclosure can be applied.
- FIG. 23 shows a configuration example of one pixel of a solid-state imaging device having pixels on which photoelectric conversion units are stacked.
- a multilayer wiring layer 22030 in which a transfer Tr (MOS FET) or the like is formed is provided on the surface 22011 side opposite to the light receiving surface, which is the surface of the semiconductor substrate 22010.
- the solid-state imaging device includes one organic photoelectric conversion unit 22040 that selectively detects light in different wavelength ranges and performs photoelectric conversion, and two inorganic photoelectric conversion units 22013 and 22014 stacked in the vertical direction.
- the organic photoelectric conversion unit 22040 includes, for example, three types of organic semiconductor materials.
- the two inorganic photoelectric conversion units 22013 and 22014 and the one organic photoelectric conversion unit 22040 are stacked, so that one element (pixel) has red (Red), green (Green), and blue. Each color signal of (Blue) can be acquired.
- the organic photoelectric conversion unit 22040 is formed on the surface 22012 which is the back surface of the semiconductor substrate 22010, and the inorganic photoelectric conversion units 22013 and 22014 are formed so as to be embedded in the semiconductor substrate 22010.
- the organic photoelectric conversion unit 22040 is composed of an organic photoelectric conversion element that generates an electron-hole pair by absorbing light in a selective wavelength region, that is, green light here, using an organic semiconductor.
- the organic photoelectric conversion unit 22040 has a configuration in which an organic photoelectric conversion layer (organic semiconductor layer) 22043 is sandwiched between a lower electrode 22041 and an upper electrode 22042 which are a pair of electrodes for extracting signal charges.
- the lower electrode 22041 and the upper electrode 22042 are electrically connected to conductive plugs 22015 and 22016 embedded in the semiconductor substrate 22010 via a wiring layer and a contact metal layer.
- interlayer insulating films 22045 and 22046 are formed on the surface 22012 of the semiconductor substrate 22010, and through holes are provided in the interlayer insulating film 22045 in regions facing the conductive plugs 22015 and 22016, respectively.
- conductive plugs 22047 and 22048 are embedded in each through hole.
- wiring layers 22049 and 22050 are embedded in regions facing the conductive plugs 22047 and 22048, respectively.
- a lower electrode 22041 is provided over the interlayer insulating film 22046, and a wiring layer 22052 electrically separated by the lower electrode 22041 and the insulating film 22051 is provided.
- the organic photoelectric conversion layer 22043 is formed on the lower electrode 22041, and the upper electrode 22042 is formed so as to cover the organic photoelectric conversion layer 22043.
- a protective film 22053 is formed on the upper electrode 22042 so as to cover the surface thereof.
- a contact hole 22054 is provided in a predetermined region of the protective film 22053, and a contact metal layer 22055 that fills the contact hole 22054 and extends to the upper surface of the wiring layer 22052 is formed on the protective film 22053.
- the conductive plug 22047 functions as a connector together with the conductive plug 22015, and forms a charge (electron) transmission path from the lower electrode 22041 to the green power storage layer 22017 together with the conductive plug 22015 and the wiring layer 22049.
- the conductive plug 22048 functions as a connector together with the conductive plug 22016 and, together with the conductive plug 22016, the wiring layer 22050, the wiring layer 22052, and the contact metal layer 22055, provides a discharge path for charges (holes) from the upper electrode 22042.
- the conductive plugs 22047 and 22048 can be composed of a laminated film of a metal material such as titanium (Ti), titanium nitride (TiN), and tungsten, for example.
- a metal material such as titanium (Ti), titanium nitride (TiN), and tungsten, for example.
- contact with silicon can be ensured even when the conductive plugs 22015 and 22016 are formed as n-type or p-type semiconductor layers.
- the interlayer insulating film 22045 includes an insulating film having a small interface state in order to reduce the interface state between the semiconductor substrate 22010 and the silicon layer 22018 and to suppress generation of dark current from the interface with the silicon layer 22018. can do.
- an insulating film for example, a stacked film of a hafnium oxide (HfO2) film and a silicon oxide (SiO2) film can be used.
- the interlayer insulating film 22046 may be formed of, for example, a single layer film made of one of silicon oxide, silicon nitride, silicon oxynitride (SiON), or the like, or a laminated film made of two or more of these. it can.
- the insulating film 22051 is formed of, for example, a single layer film made of one of silicon oxide, silicon nitride, silicon oxynitride (SiON), or the like, or a laminated film made of two or more of these.
- the insulating film 22051 has a planarized surface, and has a shape and a pattern substantially free of steps from the lower electrode 22041.
- the insulating film 22051 has a function of electrically separating the lower electrodes 22041 of each pixel of the solid-state imaging device.
- the lower electrode 22041 is provided in a region covering the light receiving surfaces facing the light receiving surfaces of the inorganic photoelectric conversion units 22013 and 22014 formed in the semiconductor substrate 22010.
- the lower electrode 22041 is made of a light-transmitting conductive film, for example, ITO (indium tin oxide).
- ITO indium tin oxide
- a tin oxide (SnO2) material added with a dopant or a zinc oxide material obtained by adding a dopant to aluminum zinc oxide (ZnO) is used. It may be used.
- zinc oxide-based material examples include aluminum zinc oxide (AZO) added with aluminum (Al) as a dopant, gallium zinc oxide (GZO) added with gallium (Ga), and indium zinc oxide added with indium (In). (IZO).
- AZO aluminum zinc oxide
- Ga gallium zinc oxide
- IZO indium zinc oxide added with indium (In).
- CuI, InSbO4, ZnMgO, CuInO2, MgIN2O4, CdO, ZnSnO3, etc. may be used.
- the lower electrode 22041 is formed separately for each pixel.
- the organic photoelectric conversion layer 22043 includes, for example, three types of a first organic semiconductor material, a second organic semiconductor material, and / or a third organic semiconductor material, and any one of these three types of organic semiconductor materials is In addition to being one or both of an organic p-type semiconductor and an organic n-type semiconductor, light in a selective wavelength range is photoelectrically converted while light in another wavelength range is transmitted. Specifically, the organic photoelectric conversion layer 22043 has a maximum absorption wavelength in a range of 450 nm to 650 nm as a green wavelength, for example.
- an undercoat film, a hole transport layer, an electron blocking film, an organic photoelectric conversion layer 22043, a hole blocking film, a buffer film, an electron transport layer, and a work function adjusting film may be stacked in this order from the lower electrode 22041 side. Good.
- the upper electrode 22042 is made of a conductive film having the same light transmittance as the lower electrode 22041.
- the upper electrode 22042 may be separated for each pixel or may be formed as a common electrode for each pixel.
- the thickness of the upper electrode 22042 is, for example, 10 nm to 200 nm.
- the protective film 22053 is made of a light-transmitting material.
- the protective film 22053 is a single-layer film made of any of silicon oxide, silicon nitride, silicon oxynitride, or the like, or a laminated film made of two or more of them. It is.
- the thickness of the protective film 22053 is, for example, 100 nm to 30000 nm.
- the contact metal layer 22055 is made of, for example, any one of titanium, tungsten, titanium nitride, aluminum and the like, or a laminated film made of two or more of them.
- the inorganic photoelectric conversion units 22013 and 22014 are PDs (photodiodes) each having a pn junction, and are formed in the order of the inorganic photoelectric conversion units 22013 and 22014 from the surface 22012 side on the optical path in the semiconductor substrate 22010.
- the inorganic photoelectric conversion unit 22013 selectively detects blue light and accumulates signal charges corresponding to blue.
- the inorganic photoelectric conversion unit 22013 is formed to extend from a selective region along the surface 22012 of the semiconductor substrate 22010 to a region near the interface with the multilayer wiring layer 22030.
- the inorganic photoelectric conversion unit 22014 selectively detects red light and accumulates signal charges corresponding to red.
- the inorganic photoelectric conversion unit 22014 is formed, for example, over a region below the inorganic photoelectric conversion unit 22013 (on the side of the surface 22011).
- blue is a color corresponding to a wavelength range of 450 nm to 495 nm
- red is a color corresponding to a wavelength range of 620 nm to 750 nm, for example, and the inorganic photoelectric conversion units 22013 and 22014 are respectively It is only necessary that light in a part or all of the wavelength range can be detected.
- the pixel in FIG. 23 has a stacked structure in which an organic photoelectric conversion unit 22040 and two inorganic photoelectric conversion units 22013 and 22014 are stacked in the vertical direction, and the organic photoelectric conversion unit 22040 emits green light and an inorganic photoelectric conversion unit.
- 22013 absorbs (detects) blue light
- inorganic photoelectric conversion unit 22014 absorbs (detects) and photoelectrically converts the light, so vertical spectroscopy in the vertical (layer) direction is performed by one pixel, and each color signal of red, green, and blue Can be obtained.
- the technology according to the present disclosure can be applied to the solid-state imaging device as described above.
- FIG. 24 is a plan view illustrating a configuration example of a solid-state imaging device having pixels in which photoelectric conversion units are stacked, to which the technology according to the present disclosure can be applied.
- FIG. 24 shows a configuration example of one pixel of a solid-state imaging device having pixels on which photoelectric conversion units are stacked.
- the pixel 25010 includes a red photoelectric conversion unit, a green photoelectric conversion unit, and a blue photoelectric conversion unit (all illustrated in the drawing) that photoelectrically convert light of each wavelength of R (Red), G (Green), and B (Blue). (Not shown) has, for example, a photoelectric conversion region 25021 stacked in three layers in the order of a green photoelectric conversion unit, a blue photoelectric conversion unit, and a red photoelectric conversion unit. Further, the pixel 25010 includes Tr groups 25110 and 25120 serving as charge read-out units that read out charges corresponding to light of RGB wavelengths from the red photoelectric conversion unit, the green photoelectric conversion unit, and the blue photoelectric conversion unit, and 25130.
- each of RGB In the solid-state imaging device, in one pixel 25010, in the vertical direction, that is, in each layer as a red photoelectric conversion unit, a green photoelectric conversion unit, and a blue photoelectric conversion unit stacked in the photoelectric conversion region 25021, each of RGB The light is split.
- the Tr groups 25110, 25120, and 25130 are formed around the photoelectric conversion region 25021.
- the Tr group 25110 outputs signal charges corresponding to R light generated and accumulated by the red photoelectric conversion unit as pixel signals.
- the Tr group 25110 includes a transfer Tr (MOS-FET) 25111, a reset Tr 25112, an amplification Tr 25113, and a selection Tr 25114.
- the Tr group 25120 outputs a signal charge corresponding to the G light generated and accumulated by the green photoelectric conversion unit as a pixel signal.
- the Tr group 25120 includes a transfer Tr 25121, a reset Tr 25122, an amplification Tr 25123, and a selection Tr 25124.
- the Tr group 25130 outputs a signal charge corresponding to the B light generated and accumulated by the blue photoelectric conversion unit as a pixel signal.
- the Tr group 25130 includes a transfer Tr 25131, a reset Tr 25132, an amplification Tr 25133, and a selection Tr 25134.
- the transfer Tr 25111 includes a gate G, a source / drain region S / D, and an FD (floating diffusion) 25115 (source / drain region).
- the transfer Tr 25121 includes a gate G, a green photoelectric conversion unit (a source / drain region connected to the photoelectric conversion region 25021), and an FD 25125.
- the transfer Tr 25131 includes a gate G, a source / drain region S / D, and an FD 25135.
- the source / drain region of the transfer Tr 25111 is connected to the red photoelectric conversion unit in the photoelectric conversion region 25021, and the source / drain region of the transfer Tr 25131 is connected to the blue photoelectric conversion unit in the photoelectric conversion region 25021. Yes.
- Each of the reset Trs 25112, 25122, and 25132, the amplification Trs 25113, 25123, and 25133, and the selection Trs 25114, 25124, and 25134 is a pair arranged in such a manner as to sandwich the gate G. Source / drain regions S / D.
- the FDs 25115, 25125, and 25135 are connected to the source / drain regions S / D that are the sources of the reset Trs 25112, 25122, and 25132, respectively, and to the gates G of the amplification Trs 25113, 25123, and 25133, respectively. It is connected.
- a source Vdd is connected to the common source / drain region S / D in each of the reset Tr 25112 and the amplification Tr 25113, the reset Tr 25122 and the amplification Tr 25123, and the reset 25132 and the amplification Tr 25133.
- a VSL (vertical signal line) is connected to the source / drain region S / D which is the source of the selection Trs 25114, 25124 and 25134.
- the technology according to the present disclosure can be applied to the solid-state imaging device as described above.
- steps describing the series of processes described above are not limited to the processes performed in time series according to the described order, but are not necessarily performed in time series, either in parallel or individually.
- the process to be executed is also included.
- the configuration described as one device (or processing unit) may be divided and configured as a plurality of devices (or processing units).
- the configurations described above as a plurality of devices (or processing units) may be combined into a single device (or processing unit).
- a configuration other than that described above may be added to the configuration of each device (or each processing unit).
- a part of the configuration of a certain device (or processing unit) may be included in the configuration of another device (or other processing unit). . That is, the present technology is not limited to the above-described embodiment, and various modifications can be made without departing from the gist of the present technology.
- this technique can also take the following structures.
- a photoelectric conversion element provided on the first surface side of the semiconductor substrate;
- a through electrode connected to the photoelectric conversion element and provided between the first surface of the semiconductor substrate and a second surface which is a surface different from the first surface;
- a transistor and a floating diffusion provided on the second surface and connected to the photoelectric conversion element via the through electrode;
- a solid-state imaging device comprising: a dielectric layer formed between the through electrodes on the second surface and thicker than a gate insulating film of the transistor.
- the solid-state imaging device according to (1) further including a shield electrode inside the dielectric layer when viewed from the second surface side.
- the material of the shield electrode is polysilicon or amorphous silicon.
- the shield electrode is controlled to a predetermined voltage from a contact electrically connected to the shield electrode and a wiring electrically connected to the contact.
- (6) The solid-state imaging device according to (4) or (5), wherein the contact electrically connected to the shield electrode is provided in a line shape between the through electrodes.
- the penetration electrode penetrates the semiconductor substrate and is separated from the semiconductor substrate by a separation groove, and the dielectric layer and the separation groove are in contact with each other.
- a solid-state imaging device according to claim 1.
- the solid-state imaging device according to any one of (1) to (7), further including one or a plurality of photodiodes provided in the semiconductor substrate.
- a photoelectric conversion element provided on the first surface side of the semiconductor substrate;
- a through electrode connected to the photoelectric conversion element and provided between the first surface of the semiconductor substrate and a second surface which is a surface different from the first surface;
- a transistor and a floating diffusion provided on the second surface and connected to the photoelectric conversion element via the through electrode;
- a solid-state imaging device comprising a dielectric layer formed between the through electrodes on the second surface and thicker than a gate insulating film of the transistor;
- a signal processing circuit for processing an output signal output from the solid-state imaging device;
- an optical system that makes incident light incident on the solid-state imaging device.
- the manufacturing equipment is Forming a through hole between the first surface of the semiconductor substrate and a second surface which is a surface different from the first surface; Forming a dielectric layer thicker than the gate insulating film of the transistor formed on the second surface between the through holes in the second surface; Forming a gate wiring including the transistor on the second surface; A through electrode is formed by embedding a metal material film in the through hole, The manufacturing method of forming the photoelectric conversion element connected to the transistor and the floating diffusion through the through electrode on the first surface.
- 1 solid-state imaging device 2 pixels, 11 semiconductor substrate, 11-1 first surface, 11-2 second surface, 101 dielectric, 102 film, 103 dielectric layer, 104 through electrode, 105 separation groove, 106 transparent electrode, 107 photoelectric conversion film, 108 transparent electrode, 109 photoelectric conversion element, 110 protective film, 111 interlayer insulation film, 112 light incident surface, 114 shield electrode, 115 amplifier transistor, 116 to 118 floating diffusion, 120 gate insulation film, 121 transfer gate , 122 vertical transfer gate, 123, 124 photodiode, 125 wiring layer, 126 insulating film, 127 multilayer substrate, 128 contacts, 129 reset transistor, 151 formation Constant region
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Power Engineering (AREA)
- General Physics & Mathematics (AREA)
- Condensed Matter Physics & Semiconductors (AREA)
- Computer Hardware Design (AREA)
- Microelectronics & Electronic Packaging (AREA)
- Electromagnetism (AREA)
- Signal Processing (AREA)
- Multimedia (AREA)
- Manufacturing & Machinery (AREA)
- Solid State Image Pick-Up Elements (AREA)
- Transforming Light Signals Into Electric Signals (AREA)
- Internal Circuitry In Semiconductor Integrated Circuit Devices (AREA)
Abstract
Description
0.装置の説明
1.実施の形態
2.イメージセンサの使用例
3.電子機器の例
4.体内情報取得システムへの応用例
5.内視鏡手術システムへの応用例
6.移動体への応用例
7.積層型固体撮像素子の構成例
<固体撮像素子の概略構成例>
図1は、本技術の各実施の形態に適用されるCMOS(Complementary Metal Oxide Semiconductor)固体撮像素子の一例の概略構成例を示している。
<固体撮像素子の第1の実施の形態の画素構造例>
図2は、本技術を適用した固体撮像素子の第1の実施の形態の画素構造例を示す断面図である。
図11は、本技術を適用した固体撮像素子の第2の実施の形態の画素構造例を示す上面図である。
図12は、本技術を適用した固体撮像素子の第3の実施の形態の画素構造例を示す上面図である。
図13は、本技術を適用した固体撮像素子の第3の実施の形態の画素構造例を示す上面図である。また、図14は、図13の画素構造の断面図である。
図15は、本技術を適用した固体撮像素子の第5の実施の形態の画素構造例を示す断面図である。
図16は、上述の固体撮像素子を使用する使用例を示す図である。
・自動停止等の安全運転や、運転者の状態の認識等のために、自動車の前方や後方、周囲、車内等を撮影する車載用センサ、走行車両や道路を監視する監視カメラ、車両間等の測距を行う測距センサ等の、交通の用に供される装置
・ユーザのジェスチャを撮影して、そのジェスチャに従った機器操作を行うために、TVや、冷蔵庫、エアーコンディショナ等の家電に供される装置
・内視鏡や、赤外光の受光による血管撮影を行う装置等の、医療やヘルスケアの用に供される装置
・防犯用途の監視カメラや、人物認証用途のカメラ等の、セキュリティの用に供される装置
・肌を撮影する肌測定器や、頭皮を撮影するマイクロスコープ等の、美容の用に供される装置
・スポーツ用途等向けのアクションカメラやウェアラブルカメラ等の、スポーツの用に供される装置
・畑や作物の状態を監視するためのカメラ等の、農業の用に供される装置
<電子機器の構成例>
本開示に係る技術(本技術)は、様々な製品へ応用することができる。例えば、本開示に係る技術は、内視鏡手術システムに適用されてもよい。
本開示に係る技術(本技術)は、様々な製品へ応用することができる。例えば、本開示に係る技術は、内視鏡手術システムに適用されてもよい。
本開示に係る技術(本技術)は、様々な製品へ応用することができる。例えば、本開示に係る技術は、自動車、電気自動車、ハイブリッド電気自動車、自動二輪車、自転車、パーソナルモビリティ、飛行機、ドローン、船舶、ロボット等のいずれかの種類の移動体に搭載される装置として実現されてもよい。
<本開示に係る技術を適用し得る、光電変換部が積層された画素を有する固体撮像装置の断面構成例>
(1) 半導体基板の第1面側に設けられた光電変換素子と、
前記光電変換素子に接続され、前記半導体基板の前記第1面と、前記第1面とは異なる面である第2面との間に設けられた貫通電極と、
前記第2面に設けられ、前記貫通電極を介して光電変換素子と接続されるトランジスタおよびフローティングディフュージョンと、
前記第2面における前記貫通電極の間に形成される、前記トランジスタのゲート絶縁膜よりも厚い誘電体層と
を備える固体撮像素子。
(2) 前記第2面側からみて、前記誘電体層の内側にシールド電極を
さらに備える前記(1)に記載の固体撮像素子。
(3) 前記シールド電極の材料は、ポリシリコンまたはアモルファスシリコンである
前記(2)に記載の固体撮像素子。
(4) 前記シールド電極は、前記シールド電極と電気的に接続されたコンタクトと、前記コンタクトに電気的に接続された配線から所定の電圧に制御されている
前記(2)または(3)に記載の固体撮像素子。
(5) 前記シールド電極に電気的に接続されたコンタクトは、前記貫通電極の間に設けられている
前記(4)に記載の固体撮像素子。
(6) 前記シールド電極に電気的に接続されたコンタクトは、前記貫通電極の間にライン状に設けられている
前記(4)または(5)に記載の固体撮像素子。
(7) 前記誘電体層は、前記貫通電極の周囲に設けられている
前記(1)乃至(6)のいずれかに記載の固体撮像素子。
(8) 前記貫通電極は、前記半導体基板を貫通すると共に、分離溝により前記半導体基板と分離されていて、前記誘電体層と前記分離溝が接している
前記(1)乃至(7)のいずれかに記載の固体撮像素子。
(9) 前記半導体基板内に設けられた1または複数のフォトダイオードを
さらに備える前記(1)乃至(7)のいずれかに記載の固体撮像素子。
(10) 半導体基板の第1面側に設けられた光電変換素子と、
前記光電変換素子に接続され、前記半導体基板の前記第1面と、前記第1面とは異なる面である第2面との間に設けられた貫通電極と、
前記第2面に設けられ、前記貫通電極を介して光電変換素子と接続されるトランジスタおよびフローティングディフュージョンと、
前記第2面における前記貫通電極の間に形成される、前記トランジスタのゲート絶縁膜よりも厚い誘電体層と
を備える固体撮像素子と、
前記固体撮像素子から出力される出力信号を処理する信号処理回路と、
入射光を前記固体撮像素子に入射する光学系と
を有する電子機器。
(11) 製造装置が、
半導体基板の第1面と、前記第1面とは異なる面である第2面との間に貫通孔を形成し、
前記第2面における貫通孔の間に、前記第2面に形成されるトランジスタのゲート絶縁膜より厚い誘電体層を形成し、
前記第2面に前記トランジスタを含むゲート配線を形成し、
前記貫通孔に金属材料膜を埋め込んで貫通電極を形成し、
前記第1面に、前記貫通電極を介して前記トランジスタおよびフローティングディフュージョンと接続する前記光電変換素子を形成する
製造方法。
Claims (11)
- 半導体基板の第1面側に設けられた光電変換素子と、
前記光電変換素子に接続され、前記半導体基板の前記第1面と、前記第1面とは異なる面である第2面との間に設けられた貫通電極と、
前記第2面に設けられ、前記貫通電極を介して光電変換素子と接続されるトランジスタおよびフローティングディフュージョンと、
前記第2面における前記貫通電極の間に形成される、前記トランジスタのゲート絶縁膜よりも厚い誘電体層と
を備える固体撮像素子。 - 前記第2面側からみて、前記誘電体層の内側にシールド電極を
さらに備える請求項1に記載の固体撮像素子。 - 前記シールド電極の材料は、ポリシリコンまたはアモルファスシリコンである
請求項2に記載の固体撮像素子。 - 前記シールド電極は、前記シールド電極と電気的に接続されたコンタクトと、前記コンタクトに電気的に接続された配線から所定の電圧に制御されている
請求項2に記載の固体撮像素子。 - 前記シールド電極に電気的に接続されたコンタクトは、前記貫通電極の間に設けられている
請求項4に記載の固体撮像素子。 - 前記シールド電極に電気的に接続されたコンタクトは、前記貫通電極の間にライン状に設けられている
請求項5に記載の固体撮像素子。 - 前記誘電体層は、前記貫通電極の周囲に設けられている
請求項1に記載の固体撮像素子。 - 前記貫通電極は、前記半導体基板を貫通すると共に、分離溝により前記半導体基板と分離されていて、前記誘電体層と前記分離溝が接している
請求項1に記載の固体撮像素子。 - 前記半導体基板内に設けられた1または複数のフォトダイオードを
さらに備える請求項1に記載の固体撮像素子。 - 半導体基板の第1面側に設けられた光電変換素子と、
前記光電変換素子に接続され、前記半導体基板の前記第1面と、前記第1面とは異なる面である第2面との間に設けられた貫通電極と、
前記第2面に設けられ、前記貫通電極を介して光電変換素子と接続されるトランジスタおよびフローティングディフュージョンと、
前記第2面における前記貫通電極の間に形成される、前記トランジスタのゲート絶縁膜よりも厚い誘電体層と
を備える固体撮像素子と、
前記固体撮像素子から出力される出力信号を処理する信号処理回路と、
入射光を前記固体撮像素子に入射する光学系と
を有する電子機器。 - 製造装置が、
半導体基板の第1面と、前記第1面とは異なる面である第2面との間に貫通孔を形成し、
前記第2面における貫通孔の間に、前記第2面に形成されるトランジスタのゲート絶縁膜より厚い誘電体層を形成し、
前記第2面に前記トランジスタを含むゲート配線を形成し、
前記貫通孔に金属材料膜を埋め込んで貫通電極を形成し、
前記第1面に、前記貫通電極を介して前記トランジスタおよびフローティングディフュージョンと接続する前記光電変換素子を形成する
製造方法。
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201880018750.8A CN110431667B (zh) | 2017-03-31 | 2018-03-16 | 固态成像元件、电子设备和制造方法 |
KR1020197025859A KR102550831B1 (ko) | 2017-03-31 | 2018-03-16 | 고체 촬상 소자, 전자 기기, 및 제조 방법 |
US16/494,845 US11049895B2 (en) | 2017-03-31 | 2018-03-16 | Solid-state imaging element, electronic device, and fabrication method |
JP2019509272A JP7187440B2 (ja) | 2017-03-31 | 2018-03-16 | 固体撮像素子、電子機器、並びに製造方法 |
DE112018001649.8T DE112018001649T5 (de) | 2017-03-31 | 2018-03-16 | Halbleiter-bilderzeugungselement, elektronische vorrichtung und herstellungsverfahren |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2017069803 | 2017-03-31 | ||
JP2017-069803 | 2017-03-31 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2018180575A1 true WO2018180575A1 (ja) | 2018-10-04 |
Family
ID=63675508
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2018/010393 WO2018180575A1 (ja) | 2017-03-31 | 2018-03-16 | 固体撮像素子、電子機器、並びに製造方法 |
Country Status (6)
Country | Link |
---|---|
US (1) | US11049895B2 (ja) |
JP (1) | JP7187440B2 (ja) |
KR (1) | KR102550831B1 (ja) |
CN (1) | CN110431667B (ja) |
DE (1) | DE112018001649T5 (ja) |
WO (1) | WO2018180575A1 (ja) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11393854B2 (en) * | 2019-01-07 | 2022-07-19 | Samsung Electronics Co., Ltd. | Image sensor with photoelectric part and transfer gate on opposite sides of the substrate |
WO2023153300A1 (ja) * | 2022-02-14 | 2023-08-17 | ソニーセミコンダクタソリューションズ株式会社 | 固体撮像素子および製造方法、並びに電子機器 |
JP7530835B2 (ja) | 2019-01-28 | 2024-08-08 | ソニーグループ株式会社 | 固体撮像素子、電子機器、および固体撮像素子の製造方法 |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11121169B2 (en) * | 2019-06-25 | 2021-09-14 | Omnivision Technologies, Inc. | Metal vertical transfer gate with high-k dielectric passivation lining |
WO2022198487A1 (zh) * | 2021-03-24 | 2022-09-29 | 京东方科技集团股份有限公司 | 显示基板及其制备方法、显示装置 |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013001809A1 (ja) * | 2011-06-30 | 2013-01-03 | パナソニック株式会社 | 固体撮像装置 |
WO2016143531A1 (ja) * | 2015-03-09 | 2016-09-15 | ソニー株式会社 | 撮像素子及びその製造方法、並びに電子機器 |
JP2016225330A (ja) * | 2015-05-27 | 2016-12-28 | ソニーセミコンダクタソリューションズ株式会社 | 固体撮像素子および撮像装置 |
Family Cites Families (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5180538B2 (ja) * | 2007-08-27 | 2013-04-10 | キヤノン株式会社 | 撮像素子及び撮像装置 |
JP5509846B2 (ja) * | 2009-12-28 | 2014-06-04 | ソニー株式会社 | 固体撮像装置とその製造方法、及び電子機器 |
JP5412316B2 (ja) * | 2010-02-23 | 2014-02-12 | パナソニック株式会社 | 半導体装置、積層型半導体装置及び半導体装置の製造方法 |
JP6232589B2 (ja) * | 2013-06-24 | 2017-11-22 | パナソニックIpマネジメント株式会社 | 固体撮像装置およびその製造方法 |
JP2015012241A (ja) * | 2013-07-01 | 2015-01-19 | ソニー株式会社 | 撮像素子およびその製造方法、ならびに電子機器 |
JP6079502B2 (ja) | 2013-08-19 | 2017-02-15 | ソニー株式会社 | 固体撮像素子および電子機器 |
JP2014146820A (ja) * | 2014-03-14 | 2014-08-14 | Panasonic Corp | 固体撮像装置 |
KR102328769B1 (ko) * | 2014-06-20 | 2021-11-18 | 삼성전자주식회사 | 이미지 센서와 이를 포함하는 이미지 처리 시스템 |
JP2016009777A (ja) * | 2014-06-25 | 2016-01-18 | ソニー株式会社 | 固体撮像素子およびその製造方法、並びに電子機器 |
TWI788994B (zh) * | 2014-11-05 | 2023-01-01 | 日商索尼半導體解決方案公司 | 固體攝像元件及其製造方法以及電子機器 |
JP6725231B2 (ja) * | 2015-10-06 | 2020-07-15 | ソニーセミコンダクタソリューションズ株式会社 | 固体撮像素子、および電子装置 |
-
2018
- 2018-03-16 WO PCT/JP2018/010393 patent/WO2018180575A1/ja active Application Filing
- 2018-03-16 DE DE112018001649.8T patent/DE112018001649T5/de active Pending
- 2018-03-16 US US16/494,845 patent/US11049895B2/en active Active
- 2018-03-16 KR KR1020197025859A patent/KR102550831B1/ko active IP Right Grant
- 2018-03-16 CN CN201880018750.8A patent/CN110431667B/zh active Active
- 2018-03-16 JP JP2019509272A patent/JP7187440B2/ja active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2013001809A1 (ja) * | 2011-06-30 | 2013-01-03 | パナソニック株式会社 | 固体撮像装置 |
WO2016143531A1 (ja) * | 2015-03-09 | 2016-09-15 | ソニー株式会社 | 撮像素子及びその製造方法、並びに電子機器 |
JP2016225330A (ja) * | 2015-05-27 | 2016-12-28 | ソニーセミコンダクタソリューションズ株式会社 | 固体撮像素子および撮像装置 |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11393854B2 (en) * | 2019-01-07 | 2022-07-19 | Samsung Electronics Co., Ltd. | Image sensor with photoelectric part and transfer gate on opposite sides of the substrate |
JP7530835B2 (ja) | 2019-01-28 | 2024-08-08 | ソニーグループ株式会社 | 固体撮像素子、電子機器、および固体撮像素子の製造方法 |
US12120897B2 (en) | 2019-01-28 | 2024-10-15 | Sony Group Corporation | Solid-state imaging element, electronic device, and manufacturing method of solid-state imaging element |
WO2023153300A1 (ja) * | 2022-02-14 | 2023-08-17 | ソニーセミコンダクタソリューションズ株式会社 | 固体撮像素子および製造方法、並びに電子機器 |
Also Published As
Publication number | Publication date |
---|---|
CN110431667B (zh) | 2024-03-22 |
US11049895B2 (en) | 2021-06-29 |
JP7187440B2 (ja) | 2022-12-12 |
KR102550831B1 (ko) | 2023-07-04 |
KR20190131482A (ko) | 2019-11-26 |
CN110431667A (zh) | 2019-11-08 |
DE112018001649T5 (de) | 2019-12-24 |
US20200286941A1 (en) | 2020-09-10 |
JPWO2018180575A1 (ja) | 2020-05-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11923385B2 (en) | Solid-state imaging device and solid-state imaging apparatus | |
US11398514B2 (en) | Solid-state image pickup device, manufacturing method therefor, and electronic apparatus | |
WO2017150167A1 (ja) | 固体撮像素子 | |
WO2020012824A1 (ja) | 固体撮像素子および電子機器 | |
US11942502B2 (en) | Solid-state imaging device, method for manufacturing the same, and electronic apparatus | |
JP7187440B2 (ja) | 固体撮像素子、電子機器、並びに製造方法 | |
JP2018206837A (ja) | 固体撮像装置および固体撮像装置の製造方法、並びに電子機器 | |
WO2019239754A1 (ja) | 固体撮像素子および固体撮像素子の製造方法ならびに電子機器 | |
WO2020162196A1 (ja) | 撮像装置および撮像システム | |
WO2022270039A1 (ja) | 固体撮像装置 | |
WO2019181466A1 (ja) | 撮像素子、電子機器 | |
WO2017122537A1 (ja) | 受光素子、受光素子の製造方法、撮像素子および電子機器 | |
WO2024038703A1 (ja) | 光検出装置 | |
WO2023188899A1 (ja) | 光検出装置及び電子機器 | |
WO2023149187A1 (ja) | 縦型トランジスタ、光検出装置、及び、電子機器 | |
WO2024024515A1 (ja) | 光検出素子および測距システム | |
US20220344390A1 (en) | Organic cis image sensor | |
WO2023248925A1 (ja) | 撮像素子及び電子機器 | |
WO2020059495A1 (ja) | 撮像素子、半導体素子および電子機器 | |
WO2018155183A1 (ja) | 撮像素子および電子機器 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 18777553 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 20197025859 Country of ref document: KR Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2019509272 Country of ref document: JP Kind code of ref document: A |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 18777553 Country of ref document: EP Kind code of ref document: A1 |