WO2022172711A1 - 光電変換素子および電子機器 - Google Patents
光電変換素子および電子機器 Download PDFInfo
- Publication number
- WO2022172711A1 WO2022172711A1 PCT/JP2022/001854 JP2022001854W WO2022172711A1 WO 2022172711 A1 WO2022172711 A1 WO 2022172711A1 JP 2022001854 W JP2022001854 W JP 2022001854W WO 2022172711 A1 WO2022172711 A1 WO 2022172711A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- pixel
- transistor
- gate
- photoelectric conversion
- wiring
- Prior art date
Links
- 238000006243 chemical reaction Methods 0.000 title claims abstract description 80
- 239000004065 semiconductor Substances 0.000 claims abstract description 107
- 238000012546 transfer Methods 0.000 claims abstract description 60
- 230000003321 amplification Effects 0.000 claims description 43
- 238000003199 nucleic acid amplification method Methods 0.000 claims description 43
- 238000007667 floating Methods 0.000 claims description 23
- 238000002955 isolation Methods 0.000 claims description 19
- 230000008859 change Effects 0.000 claims description 5
- 238000005192 partition Methods 0.000 claims description 4
- 230000035945 sensitivity Effects 0.000 claims description 2
- 238000003384 imaging method Methods 0.000 description 144
- 239000000758 substrate Substances 0.000 description 141
- 239000010410 layer Substances 0.000 description 113
- 238000010586 diagram Methods 0.000 description 31
- 238000012986 modification Methods 0.000 description 31
- 230000004048 modification Effects 0.000 description 31
- 238000001514 detection method Methods 0.000 description 23
- 238000012545 processing Methods 0.000 description 23
- 230000000875 corresponding effect Effects 0.000 description 22
- 238000009792 diffusion process Methods 0.000 description 20
- 238000004891 communication Methods 0.000 description 18
- 238000000034 method Methods 0.000 description 12
- 230000003287 optical effect Effects 0.000 description 12
- 238000005516 engineering process Methods 0.000 description 11
- 230000006866 deterioration Effects 0.000 description 10
- 239000010949 copper Substances 0.000 description 8
- 230000000694 effects Effects 0.000 description 8
- 238000002674 endoscopic surgery Methods 0.000 description 8
- 230000006870 function Effects 0.000 description 8
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 6
- 230000005540 biological transmission Effects 0.000 description 6
- 230000001276 controlling effect Effects 0.000 description 6
- 229910052802 copper Inorganic materials 0.000 description 6
- 238000009825 accumulation Methods 0.000 description 5
- 239000004020 conductor Substances 0.000 description 5
- 238000001356 surgical procedure Methods 0.000 description 5
- 239000000470 constituent Substances 0.000 description 4
- 230000005284 excitation Effects 0.000 description 4
- 230000007246 mechanism Effects 0.000 description 4
- 229910021420 polycrystalline silicon Inorganic materials 0.000 description 4
- 229920005591 polysilicon Polymers 0.000 description 4
- WFKWXMTUELFFGS-UHFFFAOYSA-N tungsten Chemical compound [W] WFKWXMTUELFFGS-UHFFFAOYSA-N 0.000 description 4
- 229910052721 tungsten Inorganic materials 0.000 description 4
- 239000010937 tungsten Substances 0.000 description 4
- 208000005646 Pneumoperitoneum Diseases 0.000 description 3
- XUIMIQQOPSSXEZ-UHFFFAOYSA-N Silicon Chemical compound [Si] XUIMIQQOPSSXEZ-UHFFFAOYSA-N 0.000 description 3
- 238000010336 energy treatment Methods 0.000 description 3
- 238000004519 manufacturing process Methods 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 229910052710 silicon Inorganic materials 0.000 description 3
- 239000010703 silicon Substances 0.000 description 3
- 229910004298 SiO 2 Inorganic materials 0.000 description 2
- 229910052782 aluminium Inorganic materials 0.000 description 2
- XAGFODPZIPBFFR-UHFFFAOYSA-N aluminium Chemical compound [Al] XAGFODPZIPBFFR-UHFFFAOYSA-N 0.000 description 2
- 210000004204 blood vessel Anatomy 0.000 description 2
- 239000003153 chemical reaction reagent Substances 0.000 description 2
- 230000000295 complement effect Effects 0.000 description 2
- 238000005401 electroluminescence Methods 0.000 description 2
- MOFVSTNWEDAEEK-UHFFFAOYSA-M indocyanine green Chemical compound [Na+].[O-]S(=O)(=O)CCCCN1C2=CC=C3C=CC=CC3=C2C(C)(C)C1=CC=CC=CC=CC1=[N+](CCCCS([O-])(=O)=O)C2=CC=C(C=CC=C3)C3=C2C1(C)C MOFVSTNWEDAEEK-UHFFFAOYSA-M 0.000 description 2
- 229960004657 indocyanine green Drugs 0.000 description 2
- 238000009434 installation Methods 0.000 description 2
- 239000011810 insulating material Substances 0.000 description 2
- 239000011229 interlayer Substances 0.000 description 2
- 238000010030 laminating Methods 0.000 description 2
- 229910052751 metal Inorganic materials 0.000 description 2
- 239000002184 metal Substances 0.000 description 2
- 229910044991 metal oxide Inorganic materials 0.000 description 2
- 150000004706 metal oxides Chemical class 0.000 description 2
- 238000004806 packaging method and process Methods 0.000 description 2
- 230000000149 penetrating effect Effects 0.000 description 2
- 230000002194 synthesizing effect Effects 0.000 description 2
- 240000004050 Pentaglottis sempervirens Species 0.000 description 1
- 235000004522 Pentaglottis sempervirens Nutrition 0.000 description 1
- VYPSYNLAJGMNEJ-UHFFFAOYSA-N Silicium dioxide Chemical compound O=[Si]=O VYPSYNLAJGMNEJ-UHFFFAOYSA-N 0.000 description 1
- GWEVSGVZZGPLCZ-UHFFFAOYSA-N Titan oxide Chemical compound O=[Ti]=O GWEVSGVZZGPLCZ-UHFFFAOYSA-N 0.000 description 1
- 230000001133 acceleration Effects 0.000 description 1
- 230000000740 bleeding effect Effects 0.000 description 1
- 238000002485 combustion reaction Methods 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000005684 electric field Effects 0.000 description 1
- 238000005530 etching Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 238000002073 fluorescence micrograph Methods 0.000 description 1
- 229910000449 hafnium oxide Inorganic materials 0.000 description 1
- WIHZLLGSGQNAGK-UHFFFAOYSA-N hafnium(4+);oxygen(2-) Chemical compound [O-2].[O-2].[Hf+4] WIHZLLGSGQNAGK-UHFFFAOYSA-N 0.000 description 1
- 230000001678 irradiating effect Effects 0.000 description 1
- 230000031700 light absorption Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000012423 maintenance Methods 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 239000003595 mist Substances 0.000 description 1
- 230000000116 mitigating effect Effects 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- TWNQGVIAIRXVLR-UHFFFAOYSA-N oxo(oxoalumanyloxy)alumane Chemical compound O=[Al]O[Al]=O TWNQGVIAIRXVLR-UHFFFAOYSA-N 0.000 description 1
- BPUBBGLMJRNUCC-UHFFFAOYSA-N oxygen(2-);tantalum(5+) Chemical compound [O-2].[O-2].[O-2].[O-2].[O-2].[Ta+5].[Ta+5] BPUBBGLMJRNUCC-UHFFFAOYSA-N 0.000 description 1
- RVTZCBVAJQQJTK-UHFFFAOYSA-N oxygen(2-);zirconium(4+) Chemical compound [O-2].[O-2].[Zr+4] RVTZCBVAJQQJTK-UHFFFAOYSA-N 0.000 description 1
- 238000007639 printing Methods 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 238000007789 sealing Methods 0.000 description 1
- 230000035939 shock Effects 0.000 description 1
- 229910052814 silicon oxide Inorganic materials 0.000 description 1
- 239000002344 surface layer Substances 0.000 description 1
- 229910001936 tantalum oxide Inorganic materials 0.000 description 1
- OGIDPMRJRNCKJF-UHFFFAOYSA-N titanium oxide Inorganic materials [Ti]=O OGIDPMRJRNCKJF-UHFFFAOYSA-N 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
- 229910001928 zirconium oxide Inorganic materials 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/14638—Structures specially adapted for transferring the charges across the imager perpendicular to the imaging plane
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/14609—Pixel-elements with integrated switching, control, storage or amplification elements
- H01L27/14612—Pixel-elements with integrated switching, control, storage or amplification elements involving a transistor
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/14603—Special geometry or disposition of pixel-elements, address-lines or gate-electrodes
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/1463—Pixel isolation structures
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/14634—Assemblies, i.e. Hybrid structures
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/14636—Interconnect structures
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L27/00—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
- H01L27/14—Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
- H01L27/144—Devices controlled by radiation
- H01L27/146—Imager structures
- H01L27/14601—Structural or functional details thereof
- H01L27/14641—Electronic components shared by two or more pixel-elements, e.g. one amplifier shared by two pixel elements
Definitions
- the present disclosure relates to photoelectric conversion elements and electronic devices.
- a solid-state imaging device having a three-dimensional structure includes, for example, a semiconductor substrate having a plurality of photoelectric conversion units and a semiconductor substrate having an amplification transistor for generating a voltage signal corresponding to the level of charge obtained by each photoelectric conversion unit. They are stacked on each other (see Patent Document 1, for example).
- a photoelectric conversion element includes a first semiconductor layer, a second semiconductor layer laminated on the first semiconductor layer, and provided between the first semiconductor layer and the second semiconductor layer. and a wiring layer.
- a photoelectric conversion portion for each pixel, a photoelectric conversion portion, a charge accumulation portion in which signal charges generated in the photoelectric conversion portion are accumulated, and a transfer transistor for transferring the signal charges from the photoelectric conversion portion to the charge accumulation portion.
- the second semiconductor layer is provided with pixel transistors for reading out signal charges in the charge storage section for each of one or more pixels.
- An interlayer insulating film and a gate wiring are provided in the wiring layer.
- the gate wiring is provided in the interlayer insulating film and connected to the gate of the transfer transistor for each pixel.
- a pixel transistor is arranged in a region between the first gate line and the second gate line in plan view.
- the first gate wiring is connected to the gate of the transfer transistor included in the first pixel in the first pixel and the second pixel, which are two pixels adjacent to each other.
- the second gate wiring is connected to the gate of the transfer transistor included in the second pixel.
- An electronic device includes the photoelectric conversion element.
- the pixel transistor is located between the first gate wiring and the second gate wiring in plan view. located in the area.
- FIG. 1 is a diagram illustrating an example of a schematic configuration of a solid-state imaging device according to an embodiment of the present disclosure
- FIG. 2 is a diagram showing an example of circuit configurations of sensor pixels and a readout circuit in FIG. 1;
- FIG. It is a figure showing an example of the cross-sectional structure of the solid-state image sensor of FIG. It is a figure showing an example of the cross-sectional structure of the solid-state image sensor of FIG. It is a figure showing an example of a cross-sectional structure in Sec1 of FIG.3 and FIG.4. It is a figure showing an example of a cross-sectional structure in Sec2 of FIG.3 and FIG.4. 1.
- FIG. 7B is a diagram showing a cross-sectional configuration example of a step following FIG. 7A;
- FIG. 7B is a diagram showing a cross-sectional configuration example of a step following FIG. 7B;
- FIG. 7D is a diagram showing a cross-sectional configuration example of a step following FIG. 7C;
- FIG. 7D is a diagram showing a cross-sectional configuration example of a process subsequent to FIG. 7D;
- FIG. 7C is a diagram illustrating a cross-sectional configuration example of a step following FIG. 7E;
- FIG. 7F is a diagram showing a cross-sectional configuration example of a step following FIG.
- FIG. 7F is a diagram showing a cross-sectional configuration example of a step following FIG. 7G
- FIG. 7H is a diagram showing a cross-sectional configuration example of a step following FIG. 7H
- FIG. 7I is a diagram showing a cross-sectional configuration example of a process subsequent to FIG. 7I
- FIG. 7J is a diagram showing a cross-sectional configuration example of a step following FIG. 7J
- FIG. 9 is a diagram showing a cross-sectional configuration example in Sec1 of FIG. 8
- FIG. 9 is a diagram showing a cross-sectional configuration example in Sec2 of FIG.
- FIG. 8 6 is a diagram showing a modified example of the cross-sectional configuration of FIG. 5;
- FIG. FIG. 7 is a diagram showing a modified example of the cross-sectional configuration of FIG. 6;
- FIG. 2 is a diagram showing a modified example of wiring connected to the sensor pixels in FIG. 1;
- FIG. 2 is a diagram showing a modified example of wiring connected to the sensor pixels in FIG. 1;
- FIG. 10 is a diagram showing a modified example of the cross-sectional configuration of FIG. 9;
- FIG. 11 is a diagram showing a modified example of the cross-sectional configuration of FIG. 10;
- 2 is a diagram showing a modified example of the circuit configuration of the sensor pixels and the readout circuit of FIG. 1;
- FIG. 5 is a diagram showing a modified example of the cross-sectional configuration of the amplification transistor in FIG. 4 ;
- FIG. 2 is a diagram showing an example of a schematic configuration of an imaging system including a solid-state imaging device according to the embodiment and its modification;
- 1 is a block diagram showing an example of a schematic configuration of a vehicle control system;
- FIG. 4 is an explanatory diagram showing an example of installation positions of an outside information detection unit and an imaging unit;
- 1 is a diagram showing an example of a schematic configuration of an endoscopic surgery system;
- FIG. 3 is a block diagram showing an example of functional configurations of a camera head and a CCU;
- Embodiment Solid-State Imaging Device
- Modified example solid-state imaging device
- Application example imaging system
- Application example Application example 1 (moving object): Fig. 20, Fig. 21
- a solid-state imaging device 1 according to an embodiment of the present disclosure will be described.
- the solid-state imaging device 1 is, for example, a back-illuminated image sensor such as a CMOS (Complementary Metal Oxide Semiconductor) image sensor.
- the solid-state imaging device 1 captures an image by receiving light from a subject and photoelectrically converting the light to generate an image signal.
- the solid-state imaging device 1 outputs pixel signals corresponding to incident light.
- a back-illuminated image sensor is an image sensor in which a photoelectric conversion section is provided between a light-receiving surface on which light from a subject is incident and a wiring layer in which wiring such as transistors for driving each pixel is provided.
- the photoelectric conversion unit is a photodiode or the like that receives light from an object and converts it into an electric signal. Note that the present disclosure is not limited to application to CMOS image sensors.
- FIG. 1 illustrates an example of a schematic configuration of a solid-state imaging device 1 according to an embodiment of the present disclosure.
- the solid-state imaging device 1 includes three substrates (first substrate 10, second substrate 20, and third substrate 30).
- the solid-state imaging device 1 is an imaging device with a three-dimensional structure configured by bonding together three substrates (first substrate 10, second substrate 20, and third substrate 30).
- the first substrate 10, the second substrate 20 and the third substrate 30 are laminated in this order.
- the first substrate 10 has a pixel area 13 in which a plurality of sensor pixels 12 that perform photoelectric conversion are arranged in a matrix. A pixel region 13 is formed on the semiconductor substrate 11 .
- the second substrate 20 has a plurality of readout circuits 22 that output pixel signals based on charges (signal charges) output from the sensor pixels 12 .
- a set of sensor pixels 12 and readout circuit 22 may be referred to as imaging pixels.
- the plurality of readout circuits 22 are formed on the semiconductor substrate 21, and are assigned to each of the plurality of sensor pixels 12, for example, as shown in FIG. In this case, one readout circuit 22 is shared by a plurality of imaging pixels.
- the second substrate 20 has a plurality of pixel drive lines 23 extending in the row direction and a plurality of vertical signal lines 24 extending in the column direction.
- the third substrate 30 has a logic circuit 32 for processing pixel signals.
- a logic circuit 32 is formed on a semiconductor substrate 31 .
- the logic circuit 32 has, for example, a vertical drive circuit 33, a column signal processing circuit 34, a horizontal drive circuit 35 and a system control circuit 36.
- the logic circuit 32 (specifically, the horizontal drive circuit 35) outputs the output voltage Vout for each sensor pixel 12 to the outside.
- the vertical drive circuit 33 selects a plurality of sensor pixels 12 in order in units of rows.
- the column signal processing circuit 34 performs, for example, correlated double sampling (CDS) processing on pixel signals output from each sensor pixel 12 in a row selected by the vertical driving circuit 33 .
- the column signal processing circuit 34 extracts the signal level of the pixel signal by performing CDS processing, for example, and holds pixel data corresponding to the amount of light received by each sensor pixel 12 .
- the horizontal driving circuit 35 for example, sequentially outputs the pixel data held in the column signal processing circuit 34 to the outside.
- the system control circuit 36 controls driving of each block (the vertical drive circuit 33, the column signal processing circuit 34 and the horizontal drive circuit 35) in the logic circuit 32, for example.
- FIG. 2 shows an example of the sensor pixel 12 and the readout circuit 22.
- “shared” means that the outputs of the four sensor pixels 12 are input to the common readout circuit 22 .
- Each sensor pixel 12 has components common to each other.
- identification numbers (1, 2, 3, 4) are added to the end of the reference numerals of the constituent elements of each sensor pixel 12 in order to distinguish the constituent elements of each sensor pixel 12 from each other.
- an identification number is added to the end of the reference numerals of the components of each sensor pixel 12 .
- the identification number at the end of the reference numerals of the constituent elements of each sensor pixel 12 is omitted.
- Each sensor pixel 12 has, for example, a photodiode PD, a transfer transistor TR, and a floating diffusion FD.
- the transfer transistor TR is electrically connected to the photodiode PD.
- the floating diffusion FD temporarily holds charges transferred from the photodiode PD via the transfer transistor TR.
- the photodiode PD corresponds to a specific example of the "photoelectric conversion section" of the present disclosure.
- the floating diffusion FD corresponds to a specific example of the "charge accumulation section" of the present disclosure.
- the photodiode PD performs photoelectric conversion to generate electric charge according to the amount of light received.
- a cathode of the photodiode PD is electrically connected to the source of the transfer transistor TR, and an anode of the photodiode PD is electrically connected to a reference potential line (eg ground).
- a drain of the transfer transistor TR is electrically connected to the floating diffusion FD, and a gate of the transfer transistor TR is electrically connected to the pixel driving line 23 through connection wirings 57 and 58 which will be described later.
- the transfer transistor TR is, for example, a CMOS (Complementary Metal Oxide Semiconductor) transistor.
- the floating diffusions FD of each sensor pixel 12 sharing one readout circuit 22 are electrically connected to each other and to the input terminal of the common readout circuit 22 .
- the readout circuit 22 has, for example, a reset transistor RST, a conversion transistor FDG, a selection transistor SEL, and an amplification transistor AMP. At least one of the selection transistor SEL and the conversion transistor FDG may be omitted if necessary.
- the source of the conversion transistor FDG (the input terminal of the readout circuit 22) is electrically connected to the floating diffusion FD through connection wirings 54 and 65.
- a drain of the conversion transistor FDG is electrically connected to a source of the reset transistor RST.
- a drain of the reset transistor RST is electrically connected to the power supply line VDD and a drain of the amplification transistor AMP.
- the source of the amplification transistor AMP is electrically connected to the drain of the selection transistor SEL, and the gate of the amplification transistor AMP is electrically connected to the source of the conversion transistor FDG through connection wirings 55 and 65 .
- the source of the selection transistor SEL (the output terminal of the readout circuit 22) is electrically connected to the vertical signal line 24.
- FIG. Gates of the conversion transistor FDG, the reset transistor RST, and the selection transistor SEL are electrically connected to the pixel drive line 23 (see FIG. 1).
- the transfer transistor TR transfers the charge of the photodiode PD to the floating diffusion FD when the transfer transistor TR is turned on.
- the transfer transistor TR is, for example, of a planar type having a gate (transfer gate TRG) formed on the surface of the semiconductor substrate 11, as shown in FIG. 3 which will be described later.
- the transfer transistor TR may be of a vertical type having a gate (vertical gate) extending from the surface of the semiconductor substrate 11 to a predetermined depth.
- the conversion transistor FDG is used when switching the conversion efficiency.
- pixel signals are small when shooting in a dark place.
- the FD capacitance C needs to be large so that V when converted into voltage by the amplification transistor AMP does not become too large (in other words, so that it becomes small).
- the conversion transistor FDG when the conversion transistor FDG is turned on, the gate capacitance corresponding to the conversion transistor FDG increases, so the overall FD capacitance C increases. On the other hand, when the conversion transistor FDG is turned off, the overall FD capacitance C becomes smaller. By switching the conversion transistor FDG on and off in this manner, the FD capacitance C can be made variable and the conversion efficiency can be switched.
- the reset transistor RST resets the potential of the floating diffusion FD to a predetermined potential.
- the selection transistor SEL controls the output timing of the pixel signal from the readout circuit 22 .
- the amplification transistor AMP generates a voltage signal corresponding to the level of the charge held in the floating diffusion FD as a pixel signal.
- the amplification transistor AMP constitutes a source follower type amplifier, and outputs a pixel signal having a voltage corresponding to the level of the charge generated in the photodiode PD.
- the amplification transistor AMP amplifies the potential of the floating diffusion FD when the selection transistor SEL is turned on, and outputs a voltage corresponding to the potential to the column signal processing circuit 34 via the vertical signal line 24 .
- the conversion transistor FDG, reset transistor RST, amplification transistor AMP, and selection transistor SEL are, for example, CMOS transistors.
- the conversion transistor FDG, the reset transistor RST, the amplification transistor AMP, and the selection transistor SEL are of planar type having gates formed on the surface of the semiconductor substrate 21, for example.
- the selection transistor SEL may be provided between the power supply line VDD and the amplification transistor AMP.
- the drain of the reset transistor RST is electrically connected to the power supply line VDD and the drain of the select transistor SEL.
- the source of the selection transistor SEL is electrically connected to the drain of the amplification transistor AMP, and the gate of the selection transistor SEL is electrically connected to the pixel drive line 23 (see FIG. 1).
- the source of the amplification transistor AMP (output end of the readout circuit 22) is electrically connected to the vertical signal line 24, and the gate of the amplification transistor AMP is electrically connected to the source of the reset transistor RST.
- FIG. 3 and 4 show an example of the vertical cross-sectional configuration of the solid-state imaging device 1.
- FIG. 3 and 4 illustrate cross-sectional configurations of portions of the solid-state imaging device 1 facing the sensor pixels 12.
- FIG. 3 illustrates a cross-sectional structure of a portion corresponding to line AA in FIG. 5 described later.
- FIG. 4 illustrates a cross-sectional structure of a portion corresponding to line AA in FIG. 6 which will be described later.
- 5 and 6 show an example of a horizontal cross-sectional configuration of the solid-state imaging device 1.
- FIG. FIG. 5 illustrates a cross-sectional configuration at Sec1 of FIGS. 3 and 4. As shown in FIG. In FIG. 5, the insulating layer 46 is omitted and the surface configuration of the semiconductor substrate 11 is superimposed.
- FIG. 6 illustrates a cross-sectional configuration at Sec2 of FIGS. 3 and 4. As shown in FIG. 6, the insulating layer 52 is omitted, and the surface structure of the semiconductor substrate 21, the connection wirings 57 and 58, the gate electrode TRG, and the element isolation portion 43 of FIG. 5 are shown superimposed.
- the solid-state imaging device 1 is configured by laminating a first substrate 10, a second substrate 20 and a third substrate 30 in this order. 70 and a receiving lens 80 are provided. For example, one color filter 70 and one light receiving lens 80 are provided for each sensor pixel 12 . That is, the solid-state imaging device 1 is a back-illuminated imaging device.
- the first substrate 10 is configured by stacking an insulating layer 46 on the semiconductor substrate 11 .
- the insulating layer 46 corresponds to a specific example of the "insulating layer" of the present disclosure.
- the insulating layer 46 is made of an inorganic insulating material such as SiO 2 or SiN, for example.
- the first substrate 10 has an insulating layer 46 as part of the wiring layer 51, which will be described later.
- the insulating layer 46 is provided in the gap between the semiconductor substrates 11 and 21 . That is, the semiconductor substrate 21 is stacked on the semiconductor substrate 11 with the insulating layer 46 interposed therebetween.
- the semiconductor substrate 11 is composed of a silicon substrate.
- the semiconductor substrate 11 has, for example, a p-well layer 42 on a part of the surface and its vicinity, and a conductive layer different from that of the p-well layer 42 in other regions (regions deeper than the p-well layer 42).
- PD41 of the type The p-well layer 42 is composed of a p-type semiconductor region.
- the PD 41 is composed of a semiconductor region of a conductivity type (specifically, n-type) different from that of the p-well layer 42 .
- the semiconductor substrate 11 has a floating diffusion FD in the p-well layer 42 as a semiconductor region of a conductivity type (specifically, n-type) different from that of the p-well layer 42 .
- the first substrate 10 (semiconductor substrate 11 ) has a photodiode PD, a transfer transistor TR and a floating diffusion FD for each sensor pixel 12 .
- the first substrate 10 has a configuration in which a transfer transistor TR and a floating diffusion FD are provided on the front surface side of the semiconductor substrate 11 (the side opposite to the light incident surface side, the second substrate 20 side).
- the first substrate 10 (semiconductor substrate 11 ) has an element isolation portion 43 that isolates each sensor pixel 12 .
- the element isolation portion 43 is formed extending in the normal direction of the semiconductor substrate 11 (the direction perpendicular to the surface of the semiconductor substrate 11).
- the element isolation portion 43 is provided between two sensor pixels 12 adjacent to each other.
- the element isolation section 43 electrically isolates the sensor pixels 12 adjacent to each other.
- the element isolation part 43 is made of, for example, silicon oxide.
- the element isolation part 43 penetrates the semiconductor substrate 11, for example.
- the first substrate 10 further has, for example, a p-well layer 44 which is a side surface of the element isolation portion 43 and is in contact with the surface on the side of the photodiode PD.
- the p-well layer 44 is composed of a semiconductor region of a conductivity type (specifically, p-type) different from that of the photodiode PD.
- the first substrate 10 further has a fixed charge film 45 in contact with the back surface of the semiconductor substrate 11, for example.
- the fixed charge film 45 is negatively charged in order to suppress the generation of dark current due to the interface level on the light receiving surface side of the semiconductor substrate 11 .
- the fixed charge film 45 is formed of, for example, an insulating film having negative fixed charges.
- a hole accumulation layer is formed at the interface on the light receiving surface side of the semiconductor substrate 11 by the electric field induced by the fixed charge film 45 . This hole accumulation layer suppresses the generation of electrons from the interface.
- Color filter 70 is provided on the back side of semiconductor substrate 11 .
- the color filter 70 is provided, for example, in contact with the fixed charge film 45 and provided at a position facing the sensor pixel 12 with the fixed charge film 45 interposed therebetween.
- the light-receiving lens 80 is provided, for example, in contact with the color filter 70 and is provided at a position facing the sensor pixel 12 with the color filter 70 and the fixed charge film 45 interposed therebetween.
- the second substrate 20 is configured by stacking an insulating layer 52 on the semiconductor substrate 21 .
- the insulating layer 52 is made of, for example, an inorganic insulating material such as SiO 2 or SiN.
- the second substrate 20 has an insulating layer 52 as part of the wiring layer 51 .
- the insulating layer 52 is provided in a gap between the semiconductor substrates 21 and 31 .
- the semiconductor substrate 21 is composed of a silicon substrate.
- the second substrate 20 semiconductor substrate 21
- the second substrate 20 (semiconductor substrate 21) has one readout circuit 22 for every four sensor pixels 12, for example.
- the second substrate 20 has a configuration in which a readout circuit 22 is provided on the surface side (third substrate 30 side) of the semiconductor substrate 21 .
- the second substrate 20 is bonded to the first substrate 10 with the back surface of the semiconductor substrate 21 facing the front surface of the semiconductor substrate 11 .
- the semiconductor substrate 21 has a plurality of openings penetrating through the semiconductor substrate 21 .
- An insulating layer 52 is embedded in each opening provided in the semiconductor substrate 21, and, for example, connection wirings 54 and 58, which will be described later, pass therethrough.
- a laminate consisting of the first substrate 10 and the second substrate 20 has a wiring layer 51 .
- the wiring layer 51 has a connection portion 53 and connection wirings 54 and 55 for each of the plurality of sensor pixels 12 sharing the readout circuit 22 .
- Each connection portion 3 and connection wirings 54 and 55 are made of a conductive material such as polysilicon, tungsten or copper.
- a part of the connection part 53 and the connection wiring 54 is provided in the insulating layer 46 of the wiring layer 51 .
- a part of the connection wiring 54 and the connection wiring 55 are provided in the insulating layer 52 of the wiring layer 51 .
- connection portion 53 is electrically connected to the floating diffusion FD of each of the multiple sensor pixels 12 sharing the readout circuit 22 .
- the four floating diffusions FD are arranged close to each other with the element isolation portion 43 interposed therebetween. Therefore, the four floating diffusions FD are electrically connected to each other by one connection portion 53 .
- connection wiring 54 is formed through the opening of the semiconductor substrate 21 and extends in the normal direction of the semiconductor substrate 21 .
- One end of the connection wiring 54 is connected to the connection portion 53 .
- the other end of the connection wiring 54 is connected to a connection wiring 65 in the wiring layer 61 which will be described later.
- the first substrate 10 and the second substrate 20 are electrically connected to each other by connecting portions 53 and connection wirings 54 and 55 .
- the connection wiring 65 is connected to the gate of the amplification transistor AMP and the source of the conversion transistor FDG.
- the connection wiring 55 is formed through the insulating layer 52 and extends in the normal direction of the insulating layer 52 .
- One end of the connection wiring 55 is connected to the gate of the amplification transistor AMP.
- the other end of the connection wiring 55 is connected to the connection wiring 65 .
- the wiring layer 51 further has a connection wiring 57 connected to the gate of the transfer transistor TR (transfer gate TRG) and a connection wiring 58 connected to the connection wiring 57 for each sensor pixel 12 .
- the connection wiring 57 corresponds to a specific example of the "gate wiring" of the present disclosure.
- the connection wiring 57 extends in a predetermined direction (first direction V) as shown in FIGS. 5 and 6, for example.
- Each connection wiring 57 is made of a conductive material such as polysilicon, tungsten or copper.
- the connection wiring 58 is formed through the opening of the semiconductor substrate 21 and extends in the normal direction of the semiconductor substrate 21 .
- One end of the connection wiring 58 is connected to the connection wiring 57 .
- the other end of the connection wiring 58 is electrically connected to the pixel drive line 23 via wiring in the insulating layer 52 .
- Each connection wiring 58 is made of a conductive material such as polysilicon, tungsten or copper.
- the connection wiring 58 is provided, for example, in a region facing the element isolation portion 43 (immediately above the element isolation portion 43).
- the connection wiring 58 is provided, for example, in a portion of the element isolation portion 43 that forms the outer edges of the plurality of sensor pixels 12 that share the readout circuit 22 .
- four sensor pixels 12 (four sensor pixels 12 included in the first imaging pixel) sharing the readout circuit 22, and a second sensor pixel adjacent in the second direction H to the first imaging pixel.
- the second direction H is a direction orthogonal to the first direction V.
- region ⁇ As shown in FIG. At this time, in the region (hereinafter referred to as “region ⁇ ” (see FIG.
- a region A connection wiring 58 is provided for each of the four sensor pixels 12 in contact with ⁇ . That is, in the region A, four connection wirings 58 are arranged side by side in the second direction H perpendicular to the first direction V. As shown in FIG.
- the region ⁇ 1 is between the connection wiring 57 (first gate wiring) of one sensor pixel 12 and the connection wiring 57 (second gate wiring) of the other sensor pixel 12 in two specific sensor pixels 12.
- the specific two sensor pixels 12 are two sensor pixels 12 arranged side by side in the second direction H among the four sensor pixels 12 sharing the readout circuit 22 .
- the amplification transistor AMP is arranged in the region ⁇ 1 in plan view.
- the region ⁇ 2 includes the connection wiring 57 (first gate wiring) of one sensor pixel 12 and the connection wiring 57 (first gate wiring) of the other sensor pixel 12 ( second gate wiring).
- the selection transistor SEL is arranged in the region ⁇ 2 in plan view.
- the area ⁇ 3 includes the connection wiring 57 (first gate wiring) of one sensor pixel 12 and the connection wiring 57 (first gate wiring) of the other sensor pixel 12 ( second gate wiring).
- reset transistor RST and conversion transistor FDG are arranged in region ⁇ 3 in plan view.
- the second substrate 20 further has a wiring layer 61 in contact with the wiring layer 51 (insulating layer 52).
- the wiring layer 61 is also in contact with the surface of the third substrate 30 on the second substrate 20 side.
- the wiring layer 61 has, for example, an insulating layer 64 and various wirings (for example, a plurality of pixel drive lines 23, a plurality of vertical signal lines 24, and a plurality of connection wirings 65) provided in the insulating layer 64. ing.
- Each pixel driving line 23, each vertical signal line 24 and each connection wiring 65 are made of a conductive material such as polysilicon, tungsten or copper.
- the wiring layer 61 further has, for example, a plurality of pad electrodes 66 within the insulating layer 64 .
- Each pad electrode 66 is made of metal such as Cu (copper) or Al (aluminum).
- Each pad electrode 66 is exposed on the surface of the wiring layer 61 .
- Each pad electrode 66 is used for electrical connection between the second substrate 20 and the third substrate 30 and for bonding the second substrate 20 and the third substrate 30 together.
- one pad electrode 66 is provided for each pixel drive line 23 and vertical signal line 24 .
- the third substrate 30 is configured by laminating a wiring layer 63 on the semiconductor substrate 31, for example.
- the semiconductor substrate 31 is composed of a silicon substrate.
- the third substrate 30 has a configuration in which a logic circuit 32 is provided on the surface side portion of the semiconductor substrate 31 .
- the third substrate 30 further has, for example, a wiring layer 62 on the wiring layer 63 .
- the wiring layer 62 has, for example, an insulating layer 68 and a plurality of pad electrodes 67 provided in the insulating layer 68 .
- a plurality of pad electrodes 67 are electrically connected to the logic circuit 32 .
- Each pad electrode 67 is made of metal such as Cu (copper) or Al (aluminum).
- Each pad electrode 67 is exposed on the surface of the wiring layer 62 .
- Each pad electrode 67 is used for electrical connection between the second substrate 20 and the third substrate 30 and bonding between the second substrate 20 and the third substrate 30 .
- the number of pad electrodes 67 does not necessarily have to be plural, and even one pad electrode 67 can be electrically connected to the logic circuit 32 .
- the second substrate 20 and the third substrate 30 are electrically connected to each other by bonding the pad electrodes 66 and 67 together.
- a gate (transfer gate TG) of the transfer transistor TR is electrically connected to the logic circuit 32 via the connection wiring 58 and the pad electrodes 66 and 67 .
- the third substrate 30 is bonded to the second substrate 20 with the surface of the semiconductor substrate 31 facing the front surface of the semiconductor substrate 21 .
- the first substrate 10 and the second substrate 20 are electrically connected to each other by connection wirings 54 and 58 .
- the second substrate 20 and the third substrate 30 are electrically connected to each other by bonding the pad electrodes 66 and 67 to each other.
- the readout circuit 22 is formed on the second substrate 20 and the logic circuit 32 is formed on the third substrate 30 .
- the structure for electrically connecting the second substrate 20 and the third substrate 30 to each other can be compared with the structure for electrically connecting the first substrate 10 and the second substrate 20 to each other. and the number of contacts for connection can be formed in a more flexible layout. Therefore, bonding between the pad electrodes 66 and 67 can be used as a structure for electrically connecting the second substrate 20 and the third substrate 30 to each other.
- a p-well layer 42 , an element isolation portion 43 and a p-well layer 44 are formed on the semiconductor substrate 11 .
- a photodiode PD, a transfer transistor TR (transfer gate TRG), and a floating diffusion FD are formed on the semiconductor substrate 11 (FIG. 7A).
- the sensor pixels 12 are formed on the semiconductor substrate 11 .
- an insulating layer 46a is formed on the semiconductor substrate 11 (FIG. 7B). At this time, an opening portion H1 is formed right above the insulating layer 46a in the insulating layer 46a so that the surface of the insulating layer 46a is exposed.
- a connection wiring 57 is formed on the surface of the insulating layer 46a including the opening H1 (FIG. 7C). Subsequently, an insulating layer 46b is formed so as to bury the connection wiring 57 (FIG. 7D). Thereby, an insulating layer 46 is formed on the semiconductor substrate 11 . Next, the semiconductor substrate 21 formed with the readout circuit 22 is placed on the surface of the insulating layer 46 (FIG. 7E). Next, openings H2 and H3 are formed at predetermined locations of the semiconductor substrate 21 (FIG. 7F).
- an opening H4 is formed through the opening H3 in the insulating layer 52a where the opening H3 is embedded (FIG. 7G).
- the connection wiring 57 is exposed at the bottom surface of the opening H4.
- a connection wiring 58 is formed so as to fill the opening H4 (FIG. 7H).
- an insulating layer 52b is formed on the surface including the connection wiring 58.
- an insulating layer 52 is formed on the semiconductor substrate 21 .
- an opening H5 penetrating through the opening H2 is formed in the insulating layer 52 where the opening H2 is embedded (FIG. 7I).
- the connection portion 53 is exposed on the bottom surface of the opening H5.
- a connection wiring 54 is formed so as to fill the opening H5 (FIG. 7J).
- a connection wiring 65 is formed on the surface of the insulating layer 52 in contact with the connection wiring 54 (FIG. 7K).
- the wiring layer 61 is formed, and the third substrate 30 is bonded onto the wiring layer 61 .
- the solid-state imaging device 1 is manufactured.
- a solid-state imaging device having a three-dimensional structure includes, for example, a semiconductor substrate having a plurality of photoelectric conversion units and a semiconductor substrate having an amplification transistor for generating a voltage signal corresponding to the level of charge obtained by each photoelectric conversion unit. They are stacked on each other (see Patent Document 1, for example).
- one transistor (pixel transistor) that configures the readout circuit 22 is connected to two connection wirings 57 (first gate wiring and second gate wiring) adjacent to each other in plan view. are arranged in intermediate regions (for example, regions ⁇ 1, ⁇ 2, ⁇ 3).
- the amplification transistor AMP is provided in a region of the element isolation portion 43 that faces a portion that partitions two adjacent sensor pixels 12 . Thereby, a sufficient space for forming the readout circuit 22 can be secured in the semiconductor substrate 21 .
- connection wirings 57 intersect the second direction H facing each other with the amplification transistor AMP in between. It extends in a first direction V. This reduces the possibility that the signal applied to the connection wiring 57 interferes with the amplification transistor AMP, for example, compared to the case where the connection wiring 57 is arranged directly under the amplification transistor AMP. As a result, deterioration of noise characteristics of the amplification transistor AMP can be suppressed.
- the conductive layer 59 as shown in FIGS. 8, 9 and 10 may be provided in the insulating layer 46 in the wiring layer 51, for example.
- FIG. 9 shows a horizontal cross-sectional configuration example of a portion corresponding to Sec1 in FIG.
- FIG. 10 shows a horizontal cross-sectional configuration example of a portion corresponding to Sec2 in FIG.
- the conductive layer 59 is provided in a region facing the amplification transistor AMP (especially the channel region of the amplification transistor AMP). This reduces the possibility that the signal from the semiconductor substrate 11 side interferes with the amplification transistor AMP. As a result, deterioration of noise characteristics of the amplification transistor AMP can be suppressed.
- the conductive layer 59 may be connected to the connection wiring 54 as shown in FIG. 8, for example.
- the potential of the conductive layer 59 can be controlled through the connection wiring 54 .
- the potential of the connection wiring 54 may be, for example, the potential of the power supply line VDD or the ground potential.
- connection wiring 57 Of the two connection wirings 57 adjacent to each other in the second direction H, one of the connection wirings 57 (first gate wiring) is focused.
- the first gate wiring is a plurality of sensors including sensor pixels 12 to which the first gate wiring is connected, as shown in FIGS. It may be connected to the gate (transfer gate TRG) of the transfer transistor TG of each pixel 12 . Also, attention is paid to the other connection wiring 57 (second gate wiring) of the two connection wirings 57 adjacent to each other in the second direction H.
- the second gate wiring is a plurality of sensors including the sensor pixels 12 to which the second gate wiring is connected, as shown in FIGS. It may be connected to the gate (transfer gate TRG) of the transfer transistor TG of each pixel 12 .
- the number of vertical wirings (connection wirings 58) electrically connecting the first substrate 10 and the second substrate 20 to each other can be reduced compared to the embodiment and its modification.
- a sufficient space for forming the readout circuit 22 can be secured on the semiconductor substrate 21 .
- the gate of the transfer transistor TR of the sensor pixel 12 included in one imaging pixel is referred to as a first gate
- the gate of the transfer transistor TR of the sensor pixel 12 included in the other imaging pixel is referred to as a first gate.
- the connection wiring 57 may be configured to connect the first gate and the second gate to each other, as shown in FIG. 13, for example. .
- the number of connection wirings 57 can be reduced compared to the case where one connection wiring 57 is provided for each sensor pixel 12 .
- a sufficient space for forming the readout circuit 22 can be secured on the semiconductor substrate 21 .
- the gate of the transfer transistor TR of the two sensor pixels 12 included in one of the two imaging pixels adjacent to each other is referred to as a third gate, and the transfer transistor TR of the two sensor pixels 12 included in the other imaging pixel is referred to as a third gate.
- the gate of transistor TR is called the fourth gate.
- the two connection wirings 57 are arranged so as to connect the two third gates and the two fourth gates to each other, as shown in FIG. may be configured to In this case, the number of connection wirings 57 can be reduced compared to the case where one connection wiring 57 is provided for each sensor pixel 12 . As a result, a sufficient space for forming the readout circuit 22 can be secured on the semiconductor substrate 21 .
- the conductive layer 59 may be provided in a region facing the entire amplification transistor AMP, as shown in FIG. 15, for example.
- the possibility that the signal from the semiconductor substrate 11 side interferes with the amplification transistor AMP is further reduced.
- deterioration of the noise characteristics of the amplification transistor AMP can be further suppressed.
- the conductive layer 59 may be insulated and separated from other conductors such as the connection wiring 54 as shown in FIG. 16, for example. At this time, the conductive layer 59 is floating. Even in this case, the possibility that the signal from the semiconductor substrate 11 side interferes with the amplification transistor AMP is reduced. As a result, deterioration of noise characteristics of the amplification transistor AMP can be suppressed.
- one readout circuit 22 may be connected to only one sensor pixel 12 as shown in FIG. 17, for example. Even in this case, the signal applied to the connection wiring 57 is less likely to interfere with the pixel transistor, as in the above embodiment and its modification. As a result, deterioration of noise characteristics of the pixel transistor can be suppressed.
- the amplification transistor AMP may be configured by a FinFET.
- the amplification transistor AMP has a channel region, a source region and a drain region within the inner side surface of an opening formed by selective etching of the semiconductor substrate 21 . That is, the amplification transistor AMP has a channel region, a source region, and a drain region within a plane intersecting the surface of the semiconductor substrate 21 .
- the amplification transistor AMP further has a gate insulating film 82 in contact with the channel region, and has a gate electrode 81 facing the channel region with the gate insulating film 82 interposed therebetween.
- the signal applied to the connection wiring 57 may interfere with the pixel transistor, as in the above embodiment and its modification. is reduced. As a result, deterioration of noise characteristics of the pixel transistor can be suppressed.
- FIG. 19 shows an example of a schematic configuration of an imaging system 2 including the solid-state imaging device 1 according to the above embodiments and modifications thereof.
- the imaging system 2 is, for example, an imaging device such as a digital still camera or a video camera, or an electronic device such as a mobile terminal device such as a smart phone or a tablet terminal.
- the imaging system 2 includes, for example, the solid-state imaging device 1 according to the above embodiment and its modification, an optical system 141, a shutter device 142, a control circuit 143, a DSP circuit 144, a frame memory 145, a display unit 146, a storage unit 147, An operation unit 148 and a power supply unit 149 are provided.
- the solid-state imaging device 1 the DSP circuit 144, the frame memory 145, the display unit 146, the storage unit 147, the operation unit 148, and the power supply unit 149 according to the above embodiment and its modification are connected via a bus line 150. connected to each other.
- the optical system 141 is configured with one or more lenses, guides light (incident light) from the subject to the solid-state imaging device 1, and forms an image on the light-receiving surface of the solid-state imaging device 1.
- the shutter device 142 is arranged between the optical system 141 and the solid-state imaging device 1 and controls the light irradiation period and the light shielding period for the solid-state imaging device 1 according to the control of the control circuit 143 .
- the solid-state imaging device 1 accumulates signal charges for a certain period of time according to the light imaged on the light receiving surface via the optical system 141 and the shutter device 142 .
- the signal charges accumulated in the solid-state imaging device 1 are transferred as pixel signals (image data) to the DSP circuit 144 according to the drive signal (timing signal) supplied from the control circuit 143 . That is, the solid-state imaging device 1 receives image light (incident light) that is incident via the optical system 141 and the shutter device 142, and outputs pixel signals corresponding to the received image light (incident light) to the DSP circuit 144. do.
- the control circuit 143 drives the solid-state image pickup device 1 and the shutter device 142 by outputting drive signals for controlling the transfer operation of the solid-state image pickup device 1 and the shutter operation of the shutter device 142 .
- the DSP circuit 144 is a signal processing circuit that processes pixel signals (image data) output from the solid-state imaging device 1 .
- the frame memory 145 temporarily holds the image data processed by the DSP circuit 144 on a frame-by-frame basis.
- the display unit 146 is, for example, a panel type display device such as a liquid crystal panel or an organic EL (Electro Luminescence) panel, and displays moving images or still images captured by the solid-state imaging device 1 .
- the storage unit 147 records image data of moving images or still images captured by the solid-state imaging device 1 in a recording medium such as a semiconductor memory or a hard disk.
- the operation unit 148 issues operation commands for various functions of the imaging system 2 in accordance with user's operations.
- the power supply unit 149 appropriately supplies various power supplies to the solid-state imaging device 1, the DSP circuit 144, the frame memory 145, the display unit 146, the storage unit 147, and the operation unit 148 as operating power supplies.
- the solid-state imaging device 1 according to the above embodiment and its modification is applied to the imaging system 2 .
- the solid-state imaging device 1 can be miniaturized or have high definition, so that a compact or high-definition imaging system 2 can be provided.
- the technology (the present technology) according to the present disclosure can be applied to various products.
- the technology according to the present disclosure is implemented as a device mounted on any type of moving object such as automobiles, electric vehicles, hybrid electric vehicles, motorcycles, bicycles, personal mobility, airplanes, drones, ships, and robots. may
- FIG. 20 is a block diagram showing a schematic configuration example of a vehicle control system, which is an example of a mobile control system to which the technology according to the present disclosure can be applied.
- a vehicle control system 12000 includes a plurality of electronic control units connected via a communication network 12001.
- the vehicle control system 12000 includes a drive system control unit 12010, a body system control unit 12020, an exterior information detection unit 12030, an interior information detection unit 12040, and an integrated control unit 12050.
- a microcomputer 12051, an audio/image output unit 12052, and an in-vehicle network I/F (interface) 12053 are illustrated.
- the drive system control unit 12010 controls the operation of devices related to the drive system of the vehicle according to various programs.
- the driving system control unit 12010 includes a driving force generator for generating driving force of the vehicle such as an internal combustion engine or a driving motor, a driving force transmission mechanism for transmitting the driving force to the wheels, and a steering angle of the vehicle. It functions as a control device such as a steering mechanism to adjust and a brake device to generate braking force of the vehicle.
- the body system control unit 12020 controls the operation of various devices equipped on the vehicle body according to various programs.
- the body system control unit 12020 functions as a keyless entry system, a smart key system, a power window device, or a control device for various lamps such as headlamps, back lamps, brake lamps, winkers or fog lamps.
- the body system control unit 12020 can receive radio waves transmitted from a portable device that substitutes for a key or signals from various switches.
- the body system control unit 12020 receives the input of these radio waves or signals and controls the door lock device, power window device, lamps, etc. of the vehicle.
- the vehicle exterior information detection unit 12030 detects information outside the vehicle in which the vehicle control system 12000 is installed.
- the vehicle exterior information detection unit 12030 is connected with an imaging section 12031 .
- the vehicle exterior information detection unit 12030 causes the imaging unit 12031 to capture an image of the exterior of the vehicle, and receives the captured image.
- the vehicle exterior information detection unit 12030 may perform object detection processing or distance detection processing such as people, vehicles, obstacles, signs, or characters on the road surface based on the received image.
- the imaging unit 12031 is an optical sensor that receives light and outputs an electrical signal according to the amount of received light.
- the imaging unit 12031 can output the electric signal as an image, and can also output it as distance measurement information.
- the light received by the imaging unit 12031 may be visible light or non-visible light such as infrared rays.
- the in-vehicle information detection unit 12040 detects in-vehicle information.
- the in-vehicle information detection unit 12040 is connected to, for example, a driver state detection section 12041 that detects the state of the driver.
- the driver state detection unit 12041 includes, for example, a camera that captures an image of the driver, and the in-vehicle information detection unit 12040 detects the degree of fatigue or concentration of the driver based on the detection information input from the driver state detection unit 12041. It may be calculated, or it may be determined whether the driver is dozing off.
- the microcomputer 12051 calculates control target values for the driving force generator, the steering mechanism, or the braking device based on the information inside and outside the vehicle acquired by the vehicle exterior information detection unit 12030 or the vehicle interior information detection unit 12040, and controls the drive system control unit.
- a control command can be output to 12010 .
- the microcomputer 12051 realizes the functions of ADAS (Advanced Driver Assistance System) including collision avoidance or shock mitigation, follow-up driving based on inter-vehicle distance, vehicle speed maintenance driving, vehicle collision warning, or vehicle lane deviation warning. Cooperative control can be performed for the purpose of ADAS (Advanced Driver Assistance System) including collision avoidance or shock mitigation, follow-up driving based on inter-vehicle distance, vehicle speed maintenance driving, vehicle collision warning, or vehicle lane deviation warning. Cooperative control can be performed for the purpose of ADAS (Advanced Driver Assistance System) including collision avoidance or shock mitigation, follow-up driving based on inter-vehicle distance, vehicle speed maintenance driving, vehicle collision warning, or vehicle
- the microcomputer 12051 controls the driving force generator, the steering mechanism, the braking device, etc. based on the information about the vehicle surroundings acquired by the vehicle exterior information detection unit 12030 or the vehicle interior information detection unit 12040, so that the driver's Cooperative control can be performed for the purpose of autonomous driving, etc., in which vehicles autonomously travel without depending on operation.
- the microcomputer 12051 can output a control command to the body system control unit 12020 based on the information outside the vehicle acquired by the information detection unit 12030 outside the vehicle.
- the microcomputer 12051 controls the headlamps according to the position of the preceding vehicle or the oncoming vehicle detected by the vehicle exterior information detection unit 12030, and performs cooperative control aimed at anti-glare such as switching from high beam to low beam. It can be carried out.
- the audio/image output unit 12052 transmits at least one of audio and/or image output signals to an output device capable of visually or audibly notifying the passengers of the vehicle or the outside of the vehicle.
- an audio speaker 12061, a display unit 12062, and an instrument panel 12063 are illustrated as output devices.
- the display unit 12062 may include at least one of an on-board display and a head-up display, for example.
- FIG. 21 is a diagram showing an example of the installation position of the imaging unit 12031.
- the vehicle 12100 has imaging units 12101, 12102, 12103, 12104, and 12105 as the imaging unit 12031.
- the imaging units 12101, 12102, 12103, 12104, and 12105 are provided at positions such as the front nose of the vehicle 12100, the side mirrors, the rear bumper, the back door, and the upper part of the windshield in the vehicle interior, for example.
- An image pickup unit 12101 provided in the front nose and an image pickup unit 12105 provided above the windshield in the passenger compartment mainly acquire images in front of the vehicle 12100 .
- Imaging units 12102 and 12103 provided in the side mirrors mainly acquire side images of the vehicle 12100 .
- An imaging unit 12104 provided in the rear bumper or back door mainly acquires an image behind the vehicle 12100 .
- Forward images acquired by the imaging units 12101 and 12105 are mainly used for detecting preceding vehicles, pedestrians, obstacles, traffic lights, traffic signs, lanes, and the like.
- FIG. 21 shows an example of the imaging range of the imaging units 12101 to 12104.
- the imaging range 12111 indicates the imaging range of the imaging unit 12101 provided in the front nose
- the imaging ranges 12112 and 12113 indicate the imaging ranges of the imaging units 12102 and 12103 provided in the side mirrors, respectively
- the imaging range 12114 The imaging range of an imaging unit 12104 provided on the rear bumper or back door is shown. For example, by superimposing the image data captured by the imaging units 12101 to 12104, a bird's-eye view image of the vehicle 12100 viewed from above can be obtained.
- At least one of the imaging units 12101 to 12104 may have a function of acquiring distance information.
- at least one of the imaging units 12101 to 12104 may be a stereo camera composed of a plurality of imaging elements, or may be an imaging element having pixels for phase difference detection.
- the microcomputer 12051 determines the distance to each three-dimensional object within the imaging ranges 12111 to 12114 and changes in this distance over time (relative velocity with respect to the vehicle 12100). , it is possible to extract, as the preceding vehicle, the closest three-dimensional object on the course of the vehicle 12100, which runs at a predetermined speed (for example, 0 km/h or more) in substantially the same direction as the vehicle 12100. can. Furthermore, the microcomputer 12051 can set the inter-vehicle distance to be secured in advance in front of the preceding vehicle, and perform automatic brake control (including following stop control) and automatic acceleration control (including following start control). In this way, cooperative control can be performed for the purpose of automatic driving in which the vehicle runs autonomously without relying on the operation of the driver.
- automatic brake control including following stop control
- automatic acceleration control including following start control
- the microcomputer 12051 converts three-dimensional object data related to three-dimensional objects to other three-dimensional objects such as motorcycles, ordinary vehicles, large vehicles, pedestrians, and utility poles. It can be classified and extracted and used for automatic avoidance of obstacles. For example, the microcomputer 12051 distinguishes obstacles around the vehicle 12100 into those that are visible to the driver of the vehicle 12100 and those that are difficult to see. Then, the microcomputer 12051 judges the collision risk indicating the degree of danger of collision with each obstacle, and when the collision risk is equal to or higher than the set value and there is a possibility of collision, an audio speaker 12061 and a display unit 12062 are displayed. By outputting an alarm to the driver via the drive system control unit 12010 and performing forced deceleration and avoidance steering via the drive system control unit 12010, driving support for collision avoidance can be performed.
- At least one of the imaging units 12101 to 12104 may be an infrared camera that detects infrared rays.
- the microcomputer 12051 can recognize a pedestrian by determining whether or not the pedestrian exists in the captured images of the imaging units 12101 to 12104 .
- recognition of a pedestrian is performed by, for example, a procedure for extracting feature points in images captured by the imaging units 12101 to 12104 as infrared cameras, and performing pattern matching processing on a series of feature points indicating the outline of an object to determine whether or not the pedestrian is a pedestrian.
- the audio image output unit 12052 outputs a rectangular outline for emphasis to the recognized pedestrian. is superimposed on the display unit 12062 . Also, the audio/image output unit 12052 may control the display unit 12062 to display an icon or the like indicating a pedestrian at a desired position.
- the technology according to the present disclosure can be applied to the imaging unit 12031 among the configurations described above.
- the solid-state imaging device 1 according to the above embodiment and its modification can be applied to the imaging section 12031 .
- FIG. 22 is a diagram showing an example of a schematic configuration of an endoscopic surgery system to which the technique (the present technique) according to the present disclosure can be applied.
- FIG. 22 illustrates a state in which an operator (doctor) 11131 is performing surgery on a patient 11132 on a patient bed 11133 using an endoscopic surgery system 11000 .
- an endoscopic surgery system 11000 includes an endoscope 11100, other surgical instruments 11110 such as a pneumoperitoneum tube 11111 and an energy treatment instrument 11112, and a support arm device 11120 for supporting the endoscope 11100. , and a cart 11200 loaded with various devices for endoscopic surgery.
- An endoscope 11100 is composed of a lens barrel 11101 whose distal end is inserted into the body cavity of a patient 11132 and a camera head 11102 connected to the proximal end of the lens barrel 11101 .
- an endoscope 11100 configured as a so-called rigid scope having a rigid lens barrel 11101 is illustrated, but the endoscope 11100 may be configured as a so-called flexible scope having a flexible lens barrel. good.
- the tip of the lens barrel 11101 is provided with an opening into which the objective lens is fitted.
- a light source device 11203 is connected to the endoscope 11100, and light generated by the light source device 11203 is guided to the tip of the lens barrel 11101 by a light guide extending inside the lens barrel 11101, where it reaches the objective. Through the lens, the light is irradiated toward the observation object inside the body cavity of the patient 11132 .
- the endoscope 11100 may be a straight scope, a perspective scope, or a side scope.
- An optical system and an imaging element are provided inside the camera head 11102, and the reflected light (observation light) from the observation target is focused on the imaging element by the optical system.
- the imaging device photoelectrically converts the observation light to generate an electrical signal corresponding to the observation light, that is, an image signal corresponding to the observation image.
- the image signal is transmitted to a camera control unit (CCU: Camera Control Unit) 11201 as RAW data.
- CCU Camera Control Unit
- the CCU 11201 is composed of a CPU (Central Processing Unit), a GPU (Graphics Processing Unit), etc., and controls the operations of the endoscope 11100 and the display device 11202 in an integrated manner. Further, the CCU 11201 receives an image signal from the camera head 11102 and performs various image processing such as development processing (demosaicing) for displaying an image based on the image signal.
- CPU Central Processing Unit
- GPU Graphics Processing Unit
- the display device 11202 displays an image based on an image signal subjected to image processing by the CCU 11201 under the control of the CCU 11201 .
- the light source device 11203 is composed of a light source such as an LED (Light Emitting Diode), for example, and supplies the endoscope 11100 with irradiation light for photographing a surgical site or the like.
- a light source such as an LED (Light Emitting Diode), for example, and supplies the endoscope 11100 with irradiation light for photographing a surgical site or the like.
- the input device 11204 is an input interface for the endoscopic surgery system 11000.
- the user can input various information and instructions to the endoscopic surgery system 11000 via the input device 11204 .
- the user inputs an instruction or the like to change imaging conditions (type of irradiation light, magnification, focal length, etc.) by the endoscope 11100 .
- the treatment instrument control device 11205 controls driving of the energy treatment instrument 11112 for tissue cauterization, incision, blood vessel sealing, or the like.
- the pneumoperitoneum device 11206 inflates the body cavity of the patient 11132 for the purpose of securing the visual field of the endoscope 11100 and securing the operator's working space, and injects gas into the body cavity through the pneumoperitoneum tube 11111. send in.
- the recorder 11207 is a device capable of recording various types of information regarding surgery.
- the printer 11208 is a device capable of printing various types of information regarding surgery in various formats such as text, images, and graphs.
- the light source device 11203 that supplies the endoscope 11100 with irradiation light for photographing the surgical site can be composed of, for example, a white light source composed of an LED, a laser light source, or a combination thereof.
- a white light source is configured by a combination of RGB laser light sources
- the output intensity and output timing of each color (each wavelength) can be controlled with high accuracy. It can be carried out.
- the observation target is irradiated with laser light from each of the RGB laser light sources in a time-division manner, and by controlling the drive of the imaging element of the camera head 11102 in synchronization with the irradiation timing, each of RGB can be handled. It is also possible to pick up images by time division. According to this method, a color image can be obtained without providing a color filter in the imaging device.
- the driving of the light source device 11203 may be controlled so as to change the intensity of the output light every predetermined time.
- the drive of the imaging device of the camera head 11102 in synchronism with the timing of the change in the intensity of the light to obtain an image in a time-division manner and synthesizing the images, a high dynamic A range of images can be generated.
- the light source device 11203 may be configured to be able to supply light in a predetermined wavelength band corresponding to special light observation.
- special light observation for example, the wavelength dependence of light absorption in body tissues is used to irradiate a narrower band of light than the irradiation light (i.e., white light) used during normal observation, thereby observing the mucosal surface layer.
- narrow band imaging is performed, in which a predetermined tissue such as a blood vessel is imaged with high contrast.
- fluorescence observation may be performed in which an image is obtained from fluorescence generated by irradiation with excitation light.
- the body tissue is irradiated with excitation light and the fluorescence from the body tissue is observed (autofluorescence observation), or a reagent such as indocyanine green (ICG) is locally injected into the body tissue and the body tissue is A fluorescence image can be obtained by irradiating excitation light corresponding to the fluorescence wavelength of the reagent.
- the light source device 11203 can be configured to be able to supply narrowband light and/or excitation light corresponding to such special light observation.
- FIG. 23 is a block diagram showing an example of functional configurations of the camera head 11102 and CCU 11201 shown in FIG.
- the camera head 11102 has a lens unit 11401, an imaging section 11402, a drive section 11403, a communication section 11404, and a camera head control section 11405.
- the CCU 11201 has a communication section 11411 , an image processing section 11412 and a control section 11413 .
- the camera head 11102 and the CCU 11201 are communicably connected to each other via a transmission cable 11400 .
- a lens unit 11401 is an optical system provided at a connection with the lens barrel 11101 . Observation light captured from the tip of the lens barrel 11101 is guided to the camera head 11102 and enters the lens unit 11401 .
- a lens unit 11401 is configured by combining a plurality of lenses including a zoom lens and a focus lens.
- the imaging unit 11402 is composed of an imaging element.
- the imaging device constituting the imaging unit 11402 may be one (so-called single-plate type) or plural (so-called multi-plate type).
- image signals corresponding to RGB may be generated by each image pickup element, and a color image may be obtained by synthesizing the image signals.
- the imaging unit 11402 may be configured to have a pair of imaging elements for respectively acquiring right-eye and left-eye image signals corresponding to 3D (Dimensional) display.
- the 3D display enables the operator 11131 to more accurately grasp the depth of the living tissue in the surgical site.
- a plurality of systems of lens units 11401 may be provided corresponding to each imaging element.
- the imaging unit 11402 does not necessarily have to be provided in the camera head 11102 .
- the imaging unit 11402 may be provided inside the lens barrel 11101 immediately after the objective lens.
- the drive unit 11403 is configured by an actuator, and moves the zoom lens and focus lens of the lens unit 11401 by a predetermined distance along the optical axis under control from the camera head control unit 11405 . Thereby, the magnification and focus of the image captured by the imaging unit 11402 can be appropriately adjusted.
- the communication unit 11404 is composed of a communication device for transmitting and receiving various information to and from the CCU 11201.
- the communication unit 11404 transmits the image signal obtained from the imaging unit 11402 as RAW data to the CCU 11201 via the transmission cable 11400 .
- the communication unit 11404 receives a control signal for controlling driving of the camera head 11102 from the CCU 11201 and supplies it to the camera head control unit 11405 .
- the control signal includes, for example, information to specify the frame rate of the captured image, information to specify the exposure value at the time of imaging, and/or information to specify the magnification and focus of the captured image. Contains information about conditions.
- the imaging conditions such as the frame rate, exposure value, magnification, and focus may be appropriately designated by the user, or may be automatically set by the control unit 11413 of the CCU 11201 based on the acquired image signal. good.
- the endoscope 11100 is equipped with so-called AE (Auto Exposure) function, AF (Auto Focus) function, and AWB (Auto White Balance) function.
- the camera head control unit 11405 controls driving of the camera head 11102 based on the control signal from the CCU 11201 received via the communication unit 11404.
- the communication unit 11411 is composed of a communication device for transmitting and receiving various information to and from the camera head 11102 .
- the communication unit 11411 receives image signals transmitted from the camera head 11102 via the transmission cable 11400 .
- the communication unit 11411 transmits a control signal for controlling driving of the camera head 11102 to the camera head 11102 .
- Image signals and control signals can be transmitted by electrical communication, optical communication, or the like.
- the image processing unit 11412 performs various types of image processing on the image signal, which is RAW data transmitted from the camera head 11102 .
- the control unit 11413 performs various controls related to imaging of the surgical site and the like by the endoscope 11100 and display of the captured image obtained by imaging the surgical site and the like. For example, the control unit 11413 generates control signals for controlling driving of the camera head 11102 .
- control unit 11413 causes the display device 11202 to display a captured image of the surgical site and the like based on the image signal that has undergone image processing by the image processing unit 11412 .
- the control unit 11413 may recognize various objects in the captured image using various image recognition techniques. For example, the control unit 11413 detects the shape, color, and the like of the edges of objects included in the captured image, thereby detecting surgical tools such as forceps, specific body parts, bleeding, mist during use of the energy treatment tool 11112, and the like. can recognize.
- the control unit 11413 may use the recognition result to display various types of surgical assistance information superimposed on the image of the surgical site. By superimposing and presenting the surgery support information to the operator 11131, the burden on the operator 11131 can be reduced and the operator 11131 can proceed with the surgery reliably.
- a transmission cable 11400 connecting the camera head 11102 and the CCU 11201 is an electrical signal cable compatible with electrical signal communication, an optical fiber compatible with optical communication, or a composite cable of these.
- wired communication is performed using the transmission cable 11400, but communication between the camera head 11102 and the CCU 11201 may be performed wirelessly.
- the technology according to the present disclosure can be preferably applied to the imaging unit 11402 provided in the camera head 11102 of the endoscope 11100 among the configurations described above.
- the technology according to the present disclosure it is possible to suppress a decrease in conversion efficiency of the imaging unit 11402, so that the endoscope 11100 with high image quality can be provided.
- the present disclosure is not limited to imaging devices, for example, and can also be applied to semiconductor devices, for example.
- Each pixel is provided with a photoelectric conversion unit, a charge storage unit in which signal charges generated in the photoelectric conversion unit are stored, and a transfer transistor for transferring the signal charges from the photoelectric conversion unit to the charge storage unit.
- a first semiconductor layer a second semiconductor layer provided with a pixel transistor for reading out the signal charge of the charge storage unit for each of the one or more pixels and laminated on the first semiconductor layer; a wiring layer provided between the first semiconductor layer and the second semiconductor layer and having a gate wiring connected to the gate of the transfer transistor for each pixel provided in an insulating layer;
- the pixel transistor has a first gate connected to the gate of the transfer transistor included in the first pixel in the first pixel and the second pixel, which are the two pixels adjacent to each other in plan view.
- a photoelectric conversion element arranged in a region between a wiring and a second gate wiring connected to the gate of the transfer transistor included in the second pixel.
- the pixel transistor includes an amplification transistor that generates a signal voltage corresponding to the level of the signal charge, a reset transistor that resets the potential of the charge storage section to a predetermined potential, a selection transistor that controls output timing of the signal voltage, and
- the photoelectric conversion element according to (1) which is at least one conversion transistor that controls sensitivity of the signal voltage to the amount of change in the signal charge.
- the first semiconductor layer has an element isolation portion that isolates the photoelectric conversion portion, the charge storage portion, and the transfer transistor for each pixel; (1) The photoelectric conversion device according to (1), wherein the pixel transistor is the amplification transistor, and is provided in a region of the element isolation section that faces a portion that partitions the first pixel and the second pixel. conversion element.
- the first gate wiring is connected to gates of the transfer transistors of each of the plurality of pixels including the first pixel; (4) The photoelectric conversion element according to (4), wherein the second gate wiring is connected to gates of the transfer transistors of the plurality of pixels including the second pixel.
- the photoelectric conversion element is Each pixel is provided with a photoelectric conversion unit, a charge storage unit in which signal charges generated in the photoelectric conversion unit are stored, and a transfer transistor for transferring the signal charges from the photoelectric conversion unit to the charge storage unit.
- a first semiconductor layer a second semiconductor layer provided with a pixel transistor for reading out the signal charge of the charge storage unit for each of the one or more pixels and laminated on the first semiconductor layer; a wiring layer provided between the first semiconductor layer and the second semiconductor layer and having a gate wiring connected to the gate of the transfer transistor for each pixel provided in an insulating layer;
- the pixel transistor has a first gate connected to the gate of the transfer transistor included in the first pixel in the first pixel and the second pixel, which are the two pixels adjacent to each other in a plan view.
- An electronic device arranged in a region between a wiring and a second gate wiring connected to the gate of the transfer transistor included in the second pixel.
- the pixel transistor is located between the first gate wiring and the second gate wiring in plan view. located in the area.
- the signal applied to the first gate wiring and the second gate wiring can be applied to the pixel transistor. less likely to interfere with As a result, deterioration of noise characteristics of the pixel transistor can be suppressed.
- the effects of the present technology are not necessarily limited to the effects described herein, and may be any of the effects described in this specification.
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Power Engineering (AREA)
- Electromagnetism (AREA)
- Condensed Matter Physics & Semiconductors (AREA)
- General Physics & Mathematics (AREA)
- Computer Hardware Design (AREA)
- Microelectronics & Electronic Packaging (AREA)
- Solid State Image Pick-Up Elements (AREA)
Abstract
Description
1.実施の形態(固体撮像素子)…図1~図7
2.変形例(固体撮像素子)…図8~図18
3.適用例(撮像システム)…図19
4.応用例
応用例1(移動体)…図20、図21
応用例2(手術システム)…図22、図23
[構成]
本開示の一実施の形態に係る固体撮像素子1について説明する。固体撮像素子1は、例えば、CMOS(Complementary Metal Oxide Semiconductor)イメージセンサ等からなる裏面照射型のイメージセンサである。固体撮像素子1は、被写体からの光を受光して光電変換し、画像信号を生成することで画像を撮像する。固体撮像素子1は、入射光に応じた画素信号を出力する。
次に、固体撮像素子1の製造方法について説明する。
次に、本実施の形態に係る固体撮像素子1の効果について説明する。
以下に、上記実施の形態に係る固体撮像素子1の変形例について説明する。
上記実施の形態において、配線層51における絶縁層46内に、例えば、図8、図9、図10に示したような導電層59が設けられていてもよい。なお、図9には、図8のSec1に対応する箇所の水平断面構成例が示されている。図10には、図8のSec2に対応する箇所の水平断面構成例が示されている。導電層59は、増幅トランジスタAMP(特に増幅トランジスタAMPのチャネル領域)と対向する領域に設けられている。これにより、増幅トランジスタAMPは、半導体基板11側からの信号が増幅トランジスタAMPに干渉する可能性が低減される。その結果、増幅トランジスタAMPのノイズ特性の悪化を抑制することができる。
第2の方向Hにおいて互いに隣接する2本の接続配線57のうち一方の接続配線57(第1のゲート配線)に着目する。このとき、上記実施の形態およびその変形例において、第1のゲート配線が、例えば、図11、図12に示したように、第1のゲート配線が接続されたセンサ画素12を含む複数のセンサ画素12の各々の転送トランジスタTGのゲート(転送ゲートTRG)に接続されていてもよい。また、第2の方向Hにおいて互いに隣接する2本の接続配線57のうち他方の接続配線57(第2のゲート配線)に着目する。このとき、上記実施の形態およびその変形例において、第2のゲート配線が、例えば、図11、図12に示したように、第2のゲート配線が接続されたセンサ画素12を含む複数のセンサ画素12の各々の転送トランジスタTGのゲート(転送ゲートTRG)に接続されていてもよい。これにより、実施の形態およびその変形例と比べて、第1基板10と第2基板20とを互いに電気的に接続する垂直配線(接続配線58)の数を減らすことができる。その結果、半導体基板21において、読み出し回路22を形成するための十分なスペースを確保することができる。
互いに隣接する2つの撮像画素のうちの、一方の撮像画素に含まれるセンサ画素12の転送トランジスタTRのゲートを第1のゲートと称し、他方の撮像画素に含まれるセンサ画素12の転送トランジスタTRのゲートを第2のゲートと称するとする。このとき、上記実施の形態およびその変形例において、接続配線57が、例えば、図13に示したように、第1のゲートと第2のゲートとを互いに接続するように構成されていてもよい。このようにした場合には、センサ画素12ごとに1本ずつ接続配線57を設けた場合と比べて、接続配線57の数を減らすことができる。その結果、半導体基板21において、読み出し回路22を形成するための十分なスペースを確保することができる。
上記変形例Dにおいて、導電層59が、例えば、図15に示したように、増幅トランジスタAMP全体と対向する領域に設けられていてもよい。このようにした場合には、半導体基板11側からの信号が増幅トランジスタAMPに干渉する可能性がより一層低減される。その結果、増幅トランジスタAMPのノイズ特性の悪化をより一層抑制することができる。
上記変形例Dにおいて、導電層59が、例えば、図16に示したように、接続配線54等の他の導電体と絶縁分離されていてもよい。このとき、導電層59は、フローティングとなっている。このようにした場合であっても、半導体基板11側からの信号が増幅トランジスタAMPに干渉する可能性が低減される。その結果、増幅トランジスタAMPのノイズ特性の悪化を抑制することができる。
上記実施の形態およびその変形例において、1つの読み出し回路22が、例えば、図17に示したように、1つのセンサ画素12だけに接続されていてもよい。このようにした場合であっても、上記実施の形態およびその変形例と同様に、接続配線57に印加された信号が、画素トランジスタに干渉する可能性が低減される。その結果、画素トランジスタのノイズ特性の悪化を抑制することができる。
上記実施の形態およびその変形例において、増幅トランジスタAMPがFinFETによって構成されていてもよい。増幅トランジスタAMPは、例えば、図18に示したように、半導体基板21に対する選択エッチングにより形成された開口の内側面内にチャネル領域、ソース領域およびドレイン領域を有している。つまり、増幅トランジスタAMPは、半導体基板21の表面と交差する面内にチャネル領域、ソース領域およびドレイン領域を有している。増幅トランジスタAMPは、さらに、チャネル領域に接するゲート絶縁膜82を有しており、このゲート絶縁膜82を介してチャネル領域と対向するゲート電極81を有している。このように、増幅トランジスタAMPがFinFETによって構成されていている場合であっても、上記実施の形態およびその変形例と同様に、接続配線57に印加された信号が、画素トランジスタに干渉する可能性が低減される。その結果、画素トランジスタのノイズ特性の悪化を抑制することができる。
図19は、上記実施の形態およびその変形例に係る固体撮像素子1を備えた撮像システム2の概略構成の一例を表したものである。
[応用例1]
本開示に係る技術(本技術)は、様々な製品へ応用することができる。例えば、本開示に係る技術は、自動車、電気自動車、ハイブリッド電気自動車、自動二輪車、自転車、パーソナルモビリティ、飛行機、ドローン、船舶、ロボット等のいずれかの種類の移動体に搭載される装置として実現されてもよい。
図22は、本開示に係る技術(本技術)が適用され得る内視鏡手術システムの概略的な構成の一例を示す図である。
(1)
画素ごとに、光電変換部と、前記光電変換部で発生した信号電荷が蓄積される電荷蓄積部と、前記信号電荷を前記光電変換部から前記電荷蓄積部に転送する転送トランジスタとが設けられた第1半導体層と、
1もしくは複数の前記画素ごとに、前記電荷蓄積部の前記信号電荷を読み出す画素トランジスタが設けられ、かつ、前記第1半導体層に積層された第2半導体層と、
前記第1半導体層と前記第2半導体層との間に設けられ、前記画素ごとに、前記転送トランジスタのゲートに接続されたゲート配線が絶縁層内に設けられた配線層と
を備え、
前記画素トランジスタが、平面視において、互いに隣接する2つの前記画素である第1の画素および第2の画素における、前記第1の画素に含まれる前記転送トランジスタのゲートに接続された第1のゲート配線と、前記第2の画素に含まれる前記転送トラジスタのゲートに接続された第2のゲート配線との間の領域に配置されている
光電変換素子。
(2)
前記画素トランジスタは、前記信号電荷のレベルに応じた信号電圧を生成する増幅トランジスタ、前記電荷蓄積部の電位を所定の電位にリセットするリセットトランジスタ、前記信号電圧の出力タイミングを制御する選択トランジスタ、および前記信号電荷の変化量に対する前記信号電圧の感度を制御する変換トランジスタのうち少なくとも1つである
(1)に記載の光電変換素子。
(3)
前記第1半導体層は、前記画素ごとに、前記光電変換部、前記電荷蓄積部および前記転送トランジスタを分離する素子分離部を有し、
前記画素トランジスタは、前記増幅トランジスタであり、前記素子分離部のうち、前記第1の画素と前記第2の画素とを区画する部分と対向する領域に設けられている
(1)に記載の光電変換素子。
(4)
前記第1のゲート配線および前記第2のゲート配線は、前記画素トランジスタを間にして互いに対向する方向と交差する方向に延在している
(3)に記載の光電変換素子。
(5)
前記第1半導体層は、前記画素トランジスタと対向する領域に導電層を更に有する
(3)に記載の光電変換素子。
(6)
前記配線層は、前記電荷蓄積部と前記画素トランジスタとを電気的に接続する垂直配線を有し、
前記導電層は、前記垂直配線に接続されている
(5)に記載の光電変換素子。
(7)
前記導電層は、フローティングとなっている
(5)に記載の光電変換素子。
(8)
前記第1のゲート配線は、前記第1の画素を含む複数の前記画素の各々の前記転送トランジスタのゲートに接続され、
前記第2のゲート配線は、前記第2の画素を含む複数の前記画素の各々の前記転送トランジスタのゲートに接続されている
(4)に記載の光電変換素子。
(9)
光電変換素子を備え、
前記光電変換素子は、
画素ごとに、光電変換部と、前記光電変換部で発生した信号電荷が蓄積される電荷蓄積部と、前記信号電荷を前記光電変換部から前記電荷蓄積部に転送する転送トランジスタとが設けられた第1半導体層と、
1もしくは複数の前記画素ごとに、前記電荷蓄積部の前記信号電荷を読み出す画素トランジスタが設けられ、かつ、前記第1半導体層に積層された第2半導体層と、
前記第1半導体層と前記第2半導体層との間に設けられ、前記画素ごとに、前記転送トランジスタのゲートに接続されたゲート配線が絶縁層内に設けられた配線層と
を有し、
前記画素トランジスタが、平面視において、互いに隣接する2つの前記画素である第1の画素および第2の画素における、前記第1の画素に含まれる前記転送トランジスタのゲートに接続された第1のゲート配線と、前記第2の画素に含まれる前記転送トラジスタのゲートに接続された第2のゲート配線との間の領域に配置されている
電子機器。
Claims (9)
- 画素ごとに、光電変換部と、前記光電変換部で発生した信号電荷が蓄積される電荷蓄積部と、前記信号電荷を前記光電変換部から前記電荷蓄積部に転送する転送トランジスタとが設けられた第1半導体層と、
1もしくは複数の前記画素ごとに、前記電荷蓄積部の前記信号電荷を読み出す画素トランジスタが設けられ、かつ、前記第1半導体層に積層された第2半導体層と、
前記第1半導体層と前記第2半導体層との間に設けられ、前記画素ごとに、前記転送トランジスタのゲートに接続されたゲート配線が絶縁層内に設けられた配線層と
を備え、
前記画素トランジスタが、平面視において、互いに隣接する2つの前記画素である第1の画素および第2の画素における、前記第1の画素に含まれる前記転送トランジスタのゲートに接続された第1のゲート配線と、前記第2の画素に含まれる前記転送トラジスタのゲートに接続された第2のゲート配線との間の領域に配置されている
光電変換素子。 - 前記画素トランジスタは、前記信号電荷のレベルに応じた信号電圧を生成する増幅トランジスタ、前記電荷蓄積部の電位を所定の電位にリセットするリセットトランジスタ、前記信号電圧の出力タイミングを制御する選択トランジスタ、および前記信号電荷の変化量に対する前記信号電圧の感度を制御する変換トランジスタのうち少なくとも1つである
請求項1に記載の光電変換素子。 - 前記第1半導体層は、前記画素ごとに、前記光電変換部、前記電荷蓄積部および前記転送トランジスタを分離する素子分離部を有し、
前記画素トランジスタは、前記増幅トランジスタであり、前記素子分離部のうち、前記第1の画素と前記第2の画素とを区画する部分と対向する領域に設けられている
請求項1に記載の光電変換素子。 - 前記第1のゲート配線および前記第2のゲート配線は、前記画素トランジスタを間にして互いに対向する方向と交差する方向に延在している
請求項3に記載の光電変換素子。 - 前記第1半導体層は、前記画素トランジスタと対向する領域に導電層を更に有する
請求項3に記載の光電変換素子。 - 前記配線層は、前記電荷蓄積部と前記画素トランジスタとを電気的に接続する垂直配線を有し、
前記導電層は、前記垂直配線に接続されている
請求項5に記載の光電変換素子。 - 前記導電層は、フローティングとなっている
請求項5に記載の光電変換素子。 - 前記第1のゲート配線は、前記第1の画素を含む複数の前記画素の各々の前記転送トランジスタのゲートに接続され、
前記第2のゲート配線は、前記第2の画素を含む複数の前記画素の各々の前記転送トランジスタのゲートに接続されている
請求項4に記載の光電変換素子。 - 光電変換素子を備え、
前記光電変換素子は、
画素ごとに、光電変換部と、前記光電変換部で発生した信号電荷が蓄積される電荷蓄積部と、前記信号電荷を前記光電変換部から前記電荷蓄積部に転送する転送トランジスタとが設けられた第1半導体層と、
1もしくは複数の前記画素ごとに、前記電荷蓄積部の前記信号電荷を読み出す画素トランジスタが設けられ、かつ、前記第1半導体層に積層された第2半導体層と、
前記第1半導体層と前記第2半導体層との間に設けられ、前記画素ごとに、前記転送トランジスタのゲートに接続されたゲート配線が絶縁層内に設けられた配線層と
を有し、
前記画素トランジスタが、平面視において、互いに隣接する2つの前記画素である第1の画素および第2の画素における、前記第1の画素に含まれる前記転送トランジスタのゲートに接続された第1のゲート配線と、前記第2の画素に含まれる前記転送トラジスタのゲートに接続された第2のゲート配線との間の領域に配置されている
電子機器。
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE112022001031.2T DE112022001031T5 (de) | 2021-02-12 | 2022-01-19 | Fotoelektrisches umwandlungselement und elektronische vorrichtung |
JP2022581283A JPWO2022172711A1 (ja) | 2021-02-12 | 2022-01-19 | |
KR1020237025017A KR20230138460A (ko) | 2021-02-12 | 2022-01-19 | 광전 변환 소자 및 전자 기기 |
US18/263,928 US20240088191A1 (en) | 2021-02-12 | 2022-01-19 | Photoelectric conversion device and electronic apparatus |
CN202280009143.1A CN116686077A (zh) | 2021-02-12 | 2022-01-19 | 光电转换元件及电子设备 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2021-020561 | 2021-02-12 | ||
JP2021020561 | 2021-02-12 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022172711A1 true WO2022172711A1 (ja) | 2022-08-18 |
Family
ID=82838707
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2022/001854 WO2022172711A1 (ja) | 2021-02-12 | 2022-01-19 | 光電変換素子および電子機器 |
Country Status (6)
Country | Link |
---|---|
US (1) | US20240088191A1 (ja) |
JP (1) | JPWO2022172711A1 (ja) |
KR (1) | KR20230138460A (ja) |
CN (1) | CN116686077A (ja) |
DE (1) | DE112022001031T5 (ja) |
WO (1) | WO2022172711A1 (ja) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2023189664A1 (ja) * | 2022-03-31 | 2023-10-05 | ソニーセミコンダクタソリューションズ株式会社 | 撮像装置及び半導体装置 |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2014033047A (ja) * | 2012-08-02 | 2014-02-20 | Nikon Corp | 固体撮像装置、及び撮像装置 |
JP2014060453A (ja) * | 2013-12-18 | 2014-04-03 | Sony Corp | 固体撮像装置及び電子機器 |
JP2020191334A (ja) * | 2019-05-20 | 2020-11-26 | ソニーセミコンダクタソリューションズ株式会社 | 固体撮像装置及び電子機器 |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2019130702A1 (ja) | 2017-12-27 | 2019-07-04 | ソニーセミコンダクタソリューションズ株式会社 | 撮像装置 |
JP7144657B2 (ja) | 2019-07-26 | 2022-09-30 | 豊田合成株式会社 | ステアリングホイール |
-
2022
- 2022-01-19 US US18/263,928 patent/US20240088191A1/en active Pending
- 2022-01-19 CN CN202280009143.1A patent/CN116686077A/zh active Pending
- 2022-01-19 KR KR1020237025017A patent/KR20230138460A/ko unknown
- 2022-01-19 WO PCT/JP2022/001854 patent/WO2022172711A1/ja active Application Filing
- 2022-01-19 JP JP2022581283A patent/JPWO2022172711A1/ja active Pending
- 2022-01-19 DE DE112022001031.2T patent/DE112022001031T5/de active Pending
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2014033047A (ja) * | 2012-08-02 | 2014-02-20 | Nikon Corp | 固体撮像装置、及び撮像装置 |
JP2014060453A (ja) * | 2013-12-18 | 2014-04-03 | Sony Corp | 固体撮像装置及び電子機器 |
JP2020191334A (ja) * | 2019-05-20 | 2020-11-26 | ソニーセミコンダクタソリューションズ株式会社 | 固体撮像装置及び電子機器 |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2023189664A1 (ja) * | 2022-03-31 | 2023-10-05 | ソニーセミコンダクタソリューションズ株式会社 | 撮像装置及び半導体装置 |
Also Published As
Publication number | Publication date |
---|---|
DE112022001031T5 (de) | 2023-11-23 |
CN116686077A (zh) | 2023-09-01 |
US20240088191A1 (en) | 2024-03-14 |
KR20230138460A (ko) | 2023-10-05 |
JPWO2022172711A1 (ja) | 2022-08-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11942502B2 (en) | Solid-state imaging device, method for manufacturing the same, and electronic apparatus | |
WO2020189534A1 (ja) | 撮像素子および半導体素子 | |
US20220254819A1 (en) | Solid-state imaging device and electronic apparatus | |
WO2021106732A1 (ja) | 撮像装置および電子機器 | |
WO2021100332A1 (ja) | 半導体装置、固体撮像装置及び電子機器 | |
US20220246653A1 (en) | Solid-state imaging element and solid-state imaging element manufacturing method | |
WO2022009627A1 (ja) | 固体撮像装置および電子機器 | |
WO2021124974A1 (ja) | 撮像装置 | |
WO2022172711A1 (ja) | 光電変換素子および電子機器 | |
WO2019181466A1 (ja) | 撮像素子、電子機器 | |
WO2022270039A1 (ja) | 固体撮像装置 | |
WO2023248925A1 (ja) | 撮像素子及び電子機器 | |
WO2024057814A1 (ja) | 光検出装置および電子機器 | |
WO2022249678A1 (ja) | 固体撮像装置およびその製造方法 | |
WO2023106316A1 (ja) | 受光装置 | |
WO2023106308A1 (ja) | 受光装置 | |
WO2024024269A1 (ja) | 固体撮像装置およびその製造方法 | |
WO2023248926A1 (ja) | 撮像素子及び電子機器 | |
WO2023047631A1 (ja) | 光検出素子及び光検出装置 | |
WO2024202748A1 (ja) | 光検出装置および電子機器 | |
WO2023058352A1 (ja) | 固体撮像装置 | |
WO2024127853A1 (ja) | 光検出装置及び電子機器 | |
US20240038807A1 (en) | Solid-state imaging device | |
WO2023188899A1 (ja) | 光検出装置及び電子機器 | |
WO2023079835A1 (ja) | 光電変換装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 22752543 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 202280009143.1 Country of ref document: CN |
|
ENP | Entry into the national phase |
Ref document number: 2022581283 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 18263928 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 112022001031 Country of ref document: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 22752543 Country of ref document: EP Kind code of ref document: A1 |