US7825970B2 - CMOS image sensor and image sensing method using the same - Google Patents

CMOS image sensor and image sensing method using the same Download PDF

Info

Publication number
US7825970B2
US7825970B2 US11/826,588 US82658807A US7825970B2 US 7825970 B2 US7825970 B2 US 7825970B2 US 82658807 A US82658807 A US 82658807A US 7825970 B2 US7825970 B2 US 7825970B2
Authority
US
United States
Prior art keywords
image sensor
cmos image
unit
overflow
floating diffusion
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US11/826,588
Other versions
US20080018765A1 (en
Inventor
Sung-Ho Choi
Jung-Chak Ahn
Yi-tae Kim
Young-Chan Kim
Hae-Kyung Kong
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics Co Ltd
Original Assignee
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Samsung Electronics Co Ltd filed Critical Samsung Electronics Co Ltd
Assigned to SAMSUNG ELECTRONICS CO., LTD. reassignment SAMSUNG ELECTRONICS CO., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: AHN, JUNG-CHAK, CHOI, SUNG-HO, KIM, YI-TAE, KIM, YOUNG-CHAN, Kong, Hae-kyung
Publication of US20080018765A1 publication Critical patent/US20080018765A1/en
Application granted granted Critical
Publication of US7825970B2 publication Critical patent/US7825970B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01LSEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
    • H01L27/00Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
    • H01L27/14Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
    • H01L27/144Devices controlled by radiation
    • H01L27/146Imager structures
    • H01L27/14643Photodiode arrays; MOS imagers
    • H01L27/14645Colour imagers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N25/00Circuitry of solid-state image sensors [SSIS]; Control thereof
    • H04N25/70SSIS architectures; Circuits associated therewith
    • H04N25/76Addressed sensors, e.g. MOS or CMOS sensors
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01LSEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
    • H01L27/00Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
    • H01L27/14Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
    • H01L27/144Devices controlled by radiation
    • H01L27/146Imager structures
    • H01L27/14601Structural or functional details thereof
    • H01L27/14603Special geometry or disposition of pixel-elements, address-lines or gate-electrodes
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01LSEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
    • H01L27/00Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
    • H01L27/14Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
    • H01L27/144Devices controlled by radiation
    • H01L27/146Imager structures
    • H01L27/14601Structural or functional details thereof
    • H01L27/14609Pixel-elements with integrated switching, control, storage or amplification elements
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01LSEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
    • H01L27/00Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
    • H01L27/14Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
    • H01L27/144Devices controlled by radiation
    • H01L27/146Imager structures
    • H01L27/14601Structural or functional details thereof
    • H01L27/1462Coatings
    • H01L27/14621Colour filter arrangements
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01LSEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
    • H01L27/00Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
    • H01L27/14Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
    • H01L27/144Devices controlled by radiation
    • H01L27/146Imager structures
    • H01L27/14601Structural or functional details thereof
    • H01L27/14625Optical elements or arrangements associated with the device
    • H01L27/14627Microlenses
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01LSEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
    • H01L27/00Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
    • H01L27/14Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
    • H01L27/144Devices controlled by radiation
    • H01L27/146Imager structures
    • H01L27/14601Structural or functional details thereof
    • H01L27/14641Electronic components shared by two or more pixel-elements, e.g. one amplifier shared by two pixel elements
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01LSEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
    • H01L27/00Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
    • H01L27/14Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
    • H01L27/144Devices controlled by radiation
    • H01L27/146Imager structures
    • H01L27/14643Photodiode arrays; MOS imagers
    • H01L27/14654Blooming suppression
    • H01L27/14656Overflow drain structures
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N25/00Circuitry of solid-state image sensors [SSIS]; Control thereof
    • H04N25/40Extracting pixel data from image sensors by controlling scanning circuits, e.g. by modifying the number of pixels sampled or to be sampled
    • H04N25/46Extracting pixel data from image sensors by controlling scanning circuits, e.g. by modifying the number of pixels sampled or to be sampled by combining or binning pixels
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N25/00Circuitry of solid-state image sensors [SSIS]; Control thereof
    • H04N25/70SSIS architectures; Circuits associated therewith
    • H04N25/76Addressed sensors, e.g. MOS or CMOS sensors
    • H04N25/77Pixel circuitry, e.g. memories, A/D converters, pixel amplifiers, shared circuits or shared components
    • H04N25/778Pixel circuitry, e.g. memories, A/D converters, pixel amplifiers, shared circuits or shared components comprising amplifiers shared between a plurality of pixels, i.e. at least one part of the amplifier must be on the sensor array itself

Definitions

  • Example embodiments may provide a CMOS image sensor, for example, a CMOS image sensor having a honeycomb structure in which photodiodes may have a hexagonal shape and two unit pixels may share a floating diffusion (FD).
  • CMOS image sensor for example, a CMOS image sensor having a honeycomb structure in which photodiodes may have a hexagonal shape and two unit pixels may share a floating diffusion (FD).
  • FD floating diffusion
  • a photodiode may convert incident light into an electric signal based on the light's wavelength. It is desirable to collect incident light with a quantum efficiency of 1 with respect to all wavelength bands, that is, all of incident light is collected.
  • a unit pixel in a related art CIS may include a single PD and four transistors.
  • the four transistors may include a transfer transistor that may transfer photo-generated charge collected at the PD to an FD, a reset transistor which may reset the FD by setting an electric potential of a node to a desired value and discharging the charge, a drive transistor which may function as a source follower buffer amplifier, and/or a selection transistor which may enable addressing through switching.
  • a PD which is influences characteristics of the CIS, may need to remain at a set size even if the size of an active pixel sensor is decreased. Sharing a circuit may overcome this problem.
  • unit pixels in a pixel array have a square shape and a degree of integration increases to achieve higher resolution, the length of a pixel array may increase, potentially resulting in increased parasitic resistance and/or parasitic capacitance.
  • a pixel pitch between unit pixels may also increase.
  • An image sensor having a honeycomb structure may prevent or reduce these problems.
  • a pixel architecture in which unit pixels share a circuit in the honeycomb structure may be desired.
  • Example embodiments may provide a CMOS image sensor with reduced size using a shared floating diffusion design in which unit pixels may share a circuit.
  • Example embodiments may also provide a CMOS image sensor capable of performing a pixel summation mode and/or draining overflow at a white level (high level intensity of illumination).
  • Example embodiments may provide a CMOS image sensor including a plurality of unit blocks each including two unit pixels.
  • Each unit block may include two photodiodes having a hexagonal shape, a floating diffusion shared by the two unit pixels, a first transfer transistor and/or a second transfer transistor between the floating diffusion and the two photodiodes, respectively, a reset transistor connected with the floating diffusion, a drive transistor whose gate may be connected with the floating diffusion, and/or a selection transistor connected with the drive transistor in series.
  • the photodiodes having the hexagonal shape may form a honeycomb structure.
  • the two unit pixels in the unit block may include a first unit pixel positioned on a k-th column and an n-th row in the honeycomb structure and a second unit pixel on the k-th column and an (n+2)-th row in the honeycomb structure, where “k” and “n” are positive integers.
  • the unit block may include two overflow transistors each connected with photodiodes to provide overflow drains.
  • One transistor of the two overflow transistors may be connected with a power supply voltage, and the other transistor of the two overflow transistors may be connected with a floating diffusion in an adjacent unit block.
  • an overflow transistor connected with a photodiode in a second unit pixel may be slightly turned on to provide an overflow drain in the second unit pixel
  • first and second transfer transistors in an adjacent unit block may be slightly turned on to provide overflow drains from the adjacent unit block.
  • Each of the unit pixels may include a microlens above one of the photodiodes to receive light and/or a color filter between the photodiode and the microlens to filter the light received by the microlens.
  • Color filters corresponding to the photodiodes in the unit block may filter the same color.
  • Color filters for a first color, a second color, and/or a third color may be sequentially arranged on an n-th row in a honeycomb structure, and color filters for the third color, the first color, and the second color may be sequentially arranged on an (n+1)-th row in the honeycomb structure, where “n” is a positive integer.
  • Color filters for a first color and a second color may be sequentially arranged on an n-th row in a honeycomb structure, and color filters for only a third color may be arranged on an (n+1)-th row in the honeycomb structure, where “n” a positive integer.
  • the first, second, and/or third colors may be blue, red, and/or green.
  • Example embodiments may provide a CMOS image sensor, in which photodiodes may form a honeycomb structure.
  • the CMOS image sensor may includes a plurality of unit blocks each including two unit pixels.
  • Each unit block may include two photodiodes having a hexagonal shape, a floating diffusion shared by the two unit pixels, a first transfer transistor and a second transfer transistor respectively between the floating diffusion and the two photodiodes, a reset transistor connected with the floating diffusion, a drive transistor whose gate may be connected with the floating diffusion, and/or a selection transistor connected with the drive transistor in series.
  • the CMOS image sensor may sum charge generated by the two photodiodes in the unit block and may output a voltage generated by the summed charge as an output voltage to support a pixel summation mode.
  • the unit block may include two overflow transistors each connected with two photodiodes to provide overflow drains.
  • One transistor of the two overflow transistors may be connected with a power supply voltage, and the other transistor of the two overflow transistors may be connected with a floating diffusion in an adjacent unit block.
  • first and second transfer transistors in an adjacent unit block may be slightly turned on to provide overflow drains in the adjacent unit block to support the pixel summation mode.
  • CMOS image sensor may include an image processor that receives data output from the CMOS image sensor and performs image processing of the received data.
  • Example embodiments may provide an image sensing method for a CMOS image sensor including a plurality of hexagonal unit blocks, each of which may include a first unit pixel and a second pixel unit that may include photodiodes.
  • Example embodiment image sensing methods may include selecting the first unit pixel, transferring charge generated by the photodiode in the first unit pixel to a first floating diffusion in a unit block including the first unit pixel, transferring a slight amount of charge, which may be generated by photodiodes in respective two unit pixels in an adjacent unit block connected to the first floating diffusion, to a second floating diffusion in the adjacent unit block to drain overflow, transferring a slight amount of charge, which may be generated by the photodiode in the second unit pixel, to a power supply voltage and/or a floating diffusion in another adjacent unit block to drain overflow, and/or reading charge from the first floating diffusion.
  • Example embodiments may provide an image sensing method for a CMOS image sensor including a plurality of hexagonal unit blocks, each of which may include a first unit pixel and a second pixel unit that may each include photodiodes.
  • Example embodiment image sensing methods may include selecting a unit block, transferring charge, which may be generated by two photodiodes in the selected unit block, to a first floating diffusion in the unit block, transferring a slight amount of charge, which may be generated by photodiodes in two unit pixels in an adjacent unit block connected to the first floating diffusion, to a second floating diffusion in the adjacent unit block, and reading charge from the first floating diffusion.
  • FIG. 1A illustrates a microlens mounted on related art rectangular-shaped photodiodes
  • FIG. 1B illustrates an example embodiment microlens mounted on photodiodes having a hexagonal shape
  • FIGS. 2A and 2B illustrate an example embodiment pixel pitch in a pixel array
  • FIGS. 2C and 2D illustrate a pixel pitch in a related art pixel array including photodiodes having a rectangular shape
  • FIG. 3 is a circuit diagram of an example embodiment CMOS image sensor
  • FIG. 4 illustrates the layout of an example embodiment CMOS image sensor
  • FIG. 5 illustrates the layout of an example embodiment CMOS image sensor
  • FIGS. 6A through 6D are timing charts of data readout in an example embodiment CMOS image sensor
  • FIGS. 7A and 7B are timing charts of a pixel summation mode of an example embodiment CMOS image sensor
  • FIGS. 8A and 8B illustrate interpolation processing in a normal mode of an example embodiment CMOS image sensor
  • FIGS. 9A through 9F illustrate an example embodiment method of calculating a pixel value during interpolation in a normal mode of a CMOS image sensor
  • FIG. 10 illustrates an example embodiment method of interpolation processing in a pixel summation mode of a CMOS image sensor
  • FIGS. 11A and 11B illustrate an example embodiment method of interpolation processing in a normal mode of a CMOS image sensor
  • FIGS. 12A through 12F illustrate an example embodiment method of calculating a pixel value during interpolation in a normal mode of a CMOS image sensor.
  • first, second, third, etc. may be used herein to describe various elements, components, regions, layers and/or sections, these elements, components, regions, layers and/or sections should not be limited by these terms. These terms are only used to distinguish one element, component, region, layer or section from another element, component, region, layer or section. Thus, a first element, component, region, layer or section discussed below could be termed a second element, component, region, layer or section without departing from the teachings of example embodiments.
  • FIG. 3 is a circuit diagram of an example embodiment CMOS image sensor 100 .
  • the CMOS image sensor 100 may include a plurality of unit blocks 10 and 20 that each may include two unit pixels.
  • the two unit pixels may be, for example, a unit pixel including a first photodiode (PD) 130 and/or a unit pixel including a third PD 140 .
  • the unit blocks 10 and 20 may have similar structures.
  • the unit block 10 may include two PDs 130 and 140 having a hexagonal shape, a floating diffusion (FD) 135 shared by the two unit pixels, a first transfer transistor 131 between the PD 130 and the FD 135 , a third transfer transistor 141 between the PD 140 and the FD 135 , a reset transistor 134 connected with the FD 135 , a drive transistor 137 with a gate connected with the FD 135 , and/or a selection transistor 136 connected in series with the drive transistor 137 .
  • the reset transistor 134 and the drive transistor 137 may be connected with power supply voltages 139 and 138 , respectively.
  • the CMOS image sensor 100 may be smaller than a related art CMOS image sensor having similar resolution.
  • the PDs 130 and 140 may receive light and generate photo-generated charge.
  • the first and third transfer transistors 131 and 141 may transfer photo-generated charge collected at the PDs 130 and 140 to the FD 135 .
  • the FD 135 may store the photo-generated charge transferred by the first and third transfer transistors 131 and 141 .
  • the reset transistor 134 may reset the FD 135 .
  • the drive transistor 137 may function as a source follower buffer amplifier.
  • the selection transistor 136 may enable addressing through switching. PDs having the hexagonal shape may form a honeycomb structure as illustrated in FIG. 4 or 5 .
  • FIG. 1A illustrates a related art microlens mounted on related art rectangular-shaped PDs.
  • FIG. 1B illustrates an example embodiment microlens mounted on PDs having a hexagonal shape.
  • dead space may be larger if a related art microlens is mounted on PDs having a rectangular shape as illustrated in FIG. 1A than if a microlens is mounted on PDs having a hexagonal shape as illustrated in FIG. 1B .
  • a hexagonal-shaped PD and a rectangular-shaped PD have the same area, dead space of the hexagonal-shaped PD may be about 13.5% while dead space of the rectangular-shaped PD may be about 21.5%. That is, the dead space of the hexagonal-shaped PD may be decreased by 8% compared to the dead space of the rectangular-shaped PD.
  • the fill-factor of a hexagonal-shaped PD may be greater than that of a rectangular-shaped PD. If a hexagonal-shaped PD is used, a pitch between pixels may be less than if the rectangular-shaped PD is used.
  • FIGS. 2A and 2B illustrate an example embodiment pixel pitch in a pixel array.
  • FIGS. 2C and 2D illustrate a pixel pitch in a conventional pixel array including photodiodes having a rectangular shape.
  • a pitch between pixels including a PD having a regular hexagonal shape is calculated as
  • a length of one side of a square is represented with “X”
  • the area of a PD having a square shape is calculated as X 2 .
  • a pitch between pixels including a PD having a square shape is expressed by X.
  • Table 1 shows the comparison of a pixel pitch between a PD having a regular hexagonal shape and a PD having a square shape when the PDs have the same area.
  • a pixel pitch for a PD having a regular hexagonal shape may be less than that for a PD having a square shape. If an image signal, for example, an image signal having a stripe pattern, is displayed, pixels including a PD having a regular hexagonal shape may display the image more clearly.
  • the dead space of a PD be reduced, and an image signal having a stripe pattern may be more clearly displayed as well because a pixel pitch in a hexagonal structure may be less than that in a square structure.
  • FIG. 4 illustrates the layout of an example embodiment CMOS image sensor 100 .
  • the circuit diagram illustrated in FIG. 3 corresponds to the layout illustrated in FIG. 4 .
  • the first unit pixel may be on a k-th (for example, 2nd or 5th) column, and an n-th (for example, 1st or 4th) row and the second unit pixel may be on the k-th column and an (n+2)-th (for example, 3rd or 6th) row, in the honeycomb structure.
  • the two unit pixels in each unit block may be on the same column at an interval of two rows.
  • “n” is a natural number less than m ⁇ 1 if a pixel array includes a total of “m” rows.
  • Each of the first and second unit pixels may include a microlens (not shown), which may be above the first or second unit pixel to receive light, and/or a color filter (not shown), which may be between the PD 130 or 140 and the microlens to filter the light received by the microlens.
  • Color filters corresponding to the respective PDs 130 and 140 having a hexagonal shape in the unit block 10 may filter the same colors, for example, green and/or red.
  • Two unit pixels in each unit block may include the same color filters, for example, green (G) color filters, red (R) color filters, or blue (B) color filters.
  • pixel summation may be accomplished through an FD shared by the two unit pixels.
  • a first, second, and third color filter for example, a blue, red, and green color filter, may be sequenced and/or repeated on a second row in the honeycomb structure.
  • Third, first, and second color filters for example, green, blue, and red color filters, may be sequenced and/or repeated on a third row in the honeycomb structure.
  • blue (B), red (R), and/or green (G) color filters may be sequenced and/or repeated on a second row and G, B, and/or R color filters may be sequenced and/or repeated on a third row.
  • a sequence of the color filters on the second row may not necessarily be B, R, and/or G. It may be desirable that a color filter on the first column and the second row, a color filter on the third column and the second row, and a color filter on the second column and the third row each filter different colors, for example, B for a PD 120 , R for a PD 160 , and G for a PD 140 .
  • FIG. 5 illustrates the layout of an example embodiment CMOS image sensor.
  • the layout illustrated in FIG. 5 is similar to that illustrated in FIG. 4 , with the exception of the arrangement of color filters.
  • first and second color filters for example, B and/or R color filters
  • a third color filter for example, a G color filter
  • Green color filters may be used more than any other color filters in the honeycomb structure illustrate in FIG. 5 because human eyes may be more sensitive to green and thus may perceive that a picture is clearer. Green color filters may absorb ultraviolet rays that may be harmful to human eyes, reduce the dazzling effects caused by intense light, and/or reduce eye fatigue.
  • the unit block 10 may include two overflow transistors 132 and 142 , which may be respectively connected with the PDs 130 and 140 to drain the overflow of charge.
  • the two overflow transistors 132 and 142 one may be connected with a power supply voltage 133 and the other may be connected with an FD in an adjacent unit block 30 .
  • CMOS image sensor If a CMOS image sensor is at a white level, a high level intensity of illumination, the overflow of photo-generated charge may occur at each PD. If the overflow occurs, the charge may flow into other unit pixels, disturbing the image display. For example, if photo-generated charge in the first PD 130 overflows to the first FD 135 if the first transfer transistor 131 is not turned on, data may be unclearly read from the unit pixel including the third PD 140 due to the overflow of the photo-generated charge.
  • the two overflow transistors 132 and 142 may be used to drain the overflow.
  • FIGS. 6A through 6D are timing charts of data readout in an example embodiment CMOS image sensor.
  • FIG. 6A is a timing chart of an operation in which data may be read from the unit pixel including the first PD 130 .
  • the overflow drain function will be described with reference to FIG. 6A below.
  • the first selection transistor 136 may be turned on by transiting a selection signal SEL 1 for controlling the first selection transistor 136 from a low level to a high level and transiting a reset signal RG 1 for controlling the first reset transistor 134 from a high level to a low level.
  • PDs that may be read by the first selection transistor 136 may be the two first and third PDs 130 and 140 . Only the first transfer transistor 131 connected with the first PD 130 may be turned on. To do so, the first transfer signal TG 1 for controlling the first transfer transistor 131 may be at a high level for a period of time.
  • the third overflow transistor 142 may be slightly turned on to prevent or reduce overflow of the third PD 140 .
  • a signal OG 3 having a voltage between a low level voltage and a high level voltage may be applied to the third overflow transistor 142 , as illustrated in FIG. 6A .
  • the voltage may be closer to the low level voltage than to the high level voltage.
  • the second and fourth transfer transistors 121 and 111 may be slightly turned on. For this operation, signals TG 2 and TG 4 applied respectively to a gate of the second and fourth transfer transistor 121 and 111 may have a voltage level similar to the signal OG 3 .
  • the third overflow transistor 142 in the unit pixel including the third PD 140 may be slightly turned on in the unit block 10 , and the two transfer transistors 111 and 121 may be slightly turned on in the adjacent unit block 20 connected to the first FD 135 .
  • FIGS. 6B through 6D are timing charts of operations in which data may be read from unit pixels, of the second, third, and fourth PDs 120 , 140 , and 110 . These operations may be similar to the operation described with reference to FIG. 6A .
  • FIGS. 7A and 7B are timing charts for a pixel summation mode of an example embodiment CMOS image sensor. If a preview function is performed in an optical system including an example embodiment CMOS image sensor, a high resolution set during actual photographing may not be needed, but quick data readout may be desired, and, therefore, the pixel summation mode may be used.
  • the first selection transistor 136 is turned on, photo-generated charge stored in the first FD 135 may be read, and the reset signal RG 1 for controlling the first reset transistor 134 may transit to the low level. Because photo-generated charge generated by the first and third PDs 130 and 140 may be transferred to the first FD 135 , the first and third transfer transistors 131 and 141 may be turned on. For this operation, the signal TG 1 applied to a gate of the first transfer transistor 131 and a signal TG 3 applied to a gate of the third transfer transistor 141 may be at a high level.
  • the two transfer transistors 111 and 121 in the adjacent unit block 20 connected with the first FD 135 may be slightly turned on to prevent or reduce overflow.
  • Data may be simultaneously read from two unit pixels in a unit block in the pixel summation mode, and it may be desirable for two color filters in the unit block filter to be the same color.
  • FIG. 7B is a timing chart of an operation in which data may be simultaneously read from the second and fourth rows. The operation may be similar to that described with reference to FIG. 7A .
  • FIGS. 8A and 8B illustrate an example embodiment method of interpolation processing in a normal mode of a CMOS image sensor.
  • Example embodiment methods may provide interpolation processing for achieving a resolution up to about two-fold higher than the number of real pixels.
  • FIG. 8A illustrates an example embodiment real pixel array in a CMOS image sensor.
  • FIG. 8B illustrates a virtual pixel array after interpolation processing is performed. As shown in FIG. 8B , the number of interpolated pixels may be double the number of real pixels, for example, 80 instead of 40.
  • a virtual pixel represented by a blank box, may be generated between real pixels, represented by a dotted box).
  • FIGS. 9A through 9F illustrate an example embodiment method of calculating data values, for example, RGB values, of each pixel illustrated in FIG. 8B .
  • a B data value of a real green pixel 11 may be generated based on data values of real blue pixels 12 , 13 , and 14 neighboring the real G pixel 11 .
  • the sum of the data values of the real blue pixels 12 , 13 , and 14 may calculated and then divided by 3.
  • An R data value of the real green pixel 11 may be generated based on data values of real red pixels 15 , 16 , and 17 neighboring the real green pixel 11 .
  • Example embodiment methods of calculating RGB values of a real blue pixel and a real red pixel may be similar to the method illustrated in FIG. 9A .
  • a G data value of a virtual pixel 21 between real green pixels 22 and 23 may be generated based on data values of the real green pixels 22 and 23 neighboring the virtual pixel 21 .
  • a B data value of the virtual pixel 21 may be generated based on data values of real blue pixels 24 , 25 , and 26 neighboring the virtual pixel 21 .
  • An R data value of the virtual pixel 21 may be generated based on data values of real red pixels 27 , 28 , and 29 neighboring the virtual pixel 21 .
  • Example embodiment methods of calculating RGB values of a virtual pixel between real blue pixels and a virtual pixel between real red pixels, which are illustrated in FIGS. 9E and 9F , respectively, may be similar to the method illustrated in FIG. 9D .
  • FIG. 10 illustrates an example embodiment method of interpolation processing in the pixel summation mode of a CMOS image sensor.
  • data of two unit pixels may be summed to generate a single data value in the pixel summation mode, in which only half of the number of all pixels in a real pixel array may exist. If the pixels obtained after pixel summation are subjected to the interpolation processing as illustrated in FIG. 8B , a resolution corresponding to the number of all pixels in the real pixel array may be generated.
  • FIGS. 11A and 11B illustrate an example embodiment method of interpolation processing in a normal mode of a CMOS image sensor having the color filter layout shown in FIG. 5 .
  • the method of the interpolation processing illustrated in FIGS. 11A and 11B may be similar to that illustrated in FIGS. 8A and 8B .
  • FIGS. 12A through 12F illustrate an example embodiment method of calculating RGB values of each pixel during interpolation in the normal mode of a CMOS image sensor.
  • a B data value of a real green pixel 31 may be generated based on data values of real blue pixels 32 and 33 neighboring the real green pixel 31 .
  • An R data value of the real green pixel 31 may be generated based on data values of real red, pixels 34 and 35 neighboring the real green pixel 31 .
  • a G data value of a blue green pixel 41 may be generated based on data values of real green pixels 42 , 43 , 44 , and 45 neighboring the real blue pixel 41 .
  • An R data value of the real blue pixel 41 may be generated based on data values of real red pixels 46 and 47 neighboring the real blue pixel 41 .
  • An example embodiment method of calculating RGB values of a real red pixel which is illustrated in FIG. 12C , may be similar to the method illustrated in FIG. 12B .
  • a G data value of a virtual pixel 61 or 66 between two real green pixels 63 and 65 or 68 and 69 and between a real blue pixel 62 or 67 and a real red pixel 64 may be generated based on data values of the real green pixels 63 and 65 or 68 and 69 neighboring the virtual pixel 61 or 66 .
  • a B data value of the virtual pixel 61 or 66 may be generated based on a data value of the real blue pixel 62 or 67 neighboring the virtual pixel 61 or 66 .
  • An R data value of the virtual pixel 61 or 66 may be generated based on a data value of the real red pixel 64 neighboring the virtual pixel 61 or 66 .
  • a G data value of a virtual pixel 51 between real green pixels 52 and 53 and between real blue pixels 54 and 55 may be generated based on data values of the real green pixels 52 and 53 neighboring the virtual pixel 51 .
  • a B data value of the virtual pixel 51 may be generated based on data values of the real blue pixels 54 and 55 neighboring the virtual pixel 51 .
  • An R data value of the virtual pixel 51 may be generated based on data values of real red pixels 56 , 57 , 58 , and 59 neighboring the virtual pixel 51 .
  • An example embodiment method of calculating RGB values of a virtual pixel between real green pixels and between real red pixels which is illustrated in FIG. 12F , may be similar to the method illustrated in FIG. 12E .
  • example embodiments provide devices that may use a shared floating diffusion design in which unit pixels may share a circuit, potentially increasing the degree of integration and resolution of those devices.
  • Example embodiments may provide a pixel summation mode and an overflow drain at a white level.
  • CMOS image sensor may be used in digital cameras, computer cameras, mobile terminals, and/or any other suitable application.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Power Engineering (AREA)
  • Electromagnetism (AREA)
  • Condensed Matter Physics & Semiconductors (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Hardware Design (AREA)
  • Microelectronics & Electronic Packaging (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Solid State Image Pick-Up Elements (AREA)
  • Transforming Light Signals Into Electric Signals (AREA)

Abstract

Example embodiments may provide a CMOS image sensor. The CMOS image sensor may include a plurality of unit blocks each including two unit pixels. Each unit block may include two photodiodes having a hexagonal shape, a floating diffusion shared by the two unit pixels, a first transfer transistor and a second transfer transistor between the floating diffusion and the two photodiodes, respectively, a reset transistor connected with the floating diffusion, a drive transistor with a gate connected with the floating diffusion, and/or a selection transistor connected to the drive transistor in series. Example embodiment CMOS image sensors may be used in digital cameras, mobile devices, computer cameras, or the like.

Description

PRIORITY STATEMENT
This application claims the priority of Korean Patent Application No. 2006-0067198, filed on Jul. 19, 2006, in the Korean Intellectual Property Office, the disclosure of which is incorporated herein in its entirety by reference.
BACKGROUND
1. Field
Example embodiments may provide a CMOS image sensor, for example, a CMOS image sensor having a honeycomb structure in which photodiodes may have a hexagonal shape and two unit pixels may share a floating diffusion (FD).
2. Description of the Related Art
In a related art charge coupled device (CCD) or CMOS image sensor (CIS), a photodiode (PD) may convert incident light into an electric signal based on the light's wavelength. It is desirable to collect incident light with a quantum efficiency of 1 with respect to all wavelength bands, that is, all of incident light is collected.
A unit pixel in a related art CIS may include a single PD and four transistors. The four transistors may include a transfer transistor that may transfer photo-generated charge collected at the PD to an FD, a reset transistor which may reset the FD by setting an electric potential of a node to a desired value and discharging the charge, a drive transistor which may function as a source follower buffer amplifier, and/or a selection transistor which may enable addressing through switching.
As size of related-art CIS's decrease, the size of an active pixel sensor receiving light may also decrease. A PD, which is influences characteristics of the CIS, may need to remain at a set size even if the size of an active pixel sensor is decreased. Sharing a circuit may overcome this problem.
If unit pixels in a pixel array have a square shape and a degree of integration increases to achieve higher resolution, the length of a pixel array may increase, potentially resulting in increased parasitic resistance and/or parasitic capacitance. A pixel pitch between unit pixels may also increase. An image sensor having a honeycomb structure may prevent or reduce these problems. A pixel architecture in which unit pixels share a circuit in the honeycomb structure may be desired.
SUMMARY
Example embodiments may provide a CMOS image sensor with reduced size using a shared floating diffusion design in which unit pixels may share a circuit.
Example embodiments may also provide a CMOS image sensor capable of performing a pixel summation mode and/or draining overflow at a white level (high level intensity of illumination).
Example embodiments may provide a CMOS image sensor including a plurality of unit blocks each including two unit pixels. Each unit block may include two photodiodes having a hexagonal shape, a floating diffusion shared by the two unit pixels, a first transfer transistor and/or a second transfer transistor between the floating diffusion and the two photodiodes, respectively, a reset transistor connected with the floating diffusion, a drive transistor whose gate may be connected with the floating diffusion, and/or a selection transistor connected with the drive transistor in series.
The photodiodes having the hexagonal shape may form a honeycomb structure.
The two unit pixels in the unit block may include a first unit pixel positioned on a k-th column and an n-th row in the honeycomb structure and a second unit pixel on the k-th column and an (n+2)-th row in the honeycomb structure, where “k” and “n” are positive integers.
The unit block may include two overflow transistors each connected with photodiodes to provide overflow drains.
One transistor of the two overflow transistors may be connected with a power supply voltage, and the other transistor of the two overflow transistors may be connected with a floating diffusion in an adjacent unit block.
If data is read from a first unit pixel among the two unit pixels in the unit block, an overflow transistor connected with a photodiode in a second unit pixel may be slightly turned on to provide an overflow drain in the second unit pixel, and first and second transfer transistors in an adjacent unit block may be slightly turned on to provide overflow drains from the adjacent unit block.
Each of the unit pixels may include a microlens above one of the photodiodes to receive light and/or a color filter between the photodiode and the microlens to filter the light received by the microlens.
Color filters corresponding to the photodiodes in the unit block may filter the same color.
Color filters for a first color, a second color, and/or a third color may be sequentially arranged on an n-th row in a honeycomb structure, and color filters for the third color, the first color, and the second color may be sequentially arranged on an (n+1)-th row in the honeycomb structure, where “n” is a positive integer.
Color filters for a first color and a second color may be sequentially arranged on an n-th row in a honeycomb structure, and color filters for only a third color may be arranged on an (n+1)-th row in the honeycomb structure, where “n” a positive integer.
The first, second, and/or third colors may be blue, red, and/or green.
Example embodiments may provide a CMOS image sensor, in which photodiodes may form a honeycomb structure. The CMOS image sensor may includes a plurality of unit blocks each including two unit pixels. Each unit block may include two photodiodes having a hexagonal shape, a floating diffusion shared by the two unit pixels, a first transfer transistor and a second transfer transistor respectively between the floating diffusion and the two photodiodes, a reset transistor connected with the floating diffusion, a drive transistor whose gate may be connected with the floating diffusion, and/or a selection transistor connected with the drive transistor in series. If the unit block is read, the CMOS image sensor may sum charge generated by the two photodiodes in the unit block and may output a voltage generated by the summed charge as an output voltage to support a pixel summation mode.
The unit block may include two overflow transistors each connected with two photodiodes to provide overflow drains. One transistor of the two overflow transistors may be connected with a power supply voltage, and the other transistor of the two overflow transistors may be connected with a floating diffusion in an adjacent unit block.
If the unit block is read, first and second transfer transistors in an adjacent unit block may be slightly turned on to provide overflow drains in the adjacent unit block to support the pixel summation mode.
For example, a digital camera, computer camera, or mobile terminal including an example embodiment CMOS image sensor may include an image processor that receives data output from the CMOS image sensor and performs image processing of the received data.
Example embodiments may provide an image sensing method for a CMOS image sensor including a plurality of hexagonal unit blocks, each of which may include a first unit pixel and a second pixel unit that may include photodiodes. Example embodiment image sensing methods may include selecting the first unit pixel, transferring charge generated by the photodiode in the first unit pixel to a first floating diffusion in a unit block including the first unit pixel, transferring a slight amount of charge, which may be generated by photodiodes in respective two unit pixels in an adjacent unit block connected to the first floating diffusion, to a second floating diffusion in the adjacent unit block to drain overflow, transferring a slight amount of charge, which may be generated by the photodiode in the second unit pixel, to a power supply voltage and/or a floating diffusion in another adjacent unit block to drain overflow, and/or reading charge from the first floating diffusion.
Example embodiments may provide an image sensing method for a CMOS image sensor including a plurality of hexagonal unit blocks, each of which may include a first unit pixel and a second pixel unit that may each include photodiodes. Example embodiment image sensing methods may include selecting a unit block, transferring charge, which may be generated by two photodiodes in the selected unit block, to a first floating diffusion in the unit block, transferring a slight amount of charge, which may be generated by photodiodes in two unit pixels in an adjacent unit block connected to the first floating diffusion, to a second floating diffusion in the adjacent unit block, and reading charge from the first floating diffusion.
BRIEF DESCRIPTION OF THE DRAWINGS
The above and other features and/or advantages of example embodiments will become more apparent by describing them in detail with reference to the attached drawings in which:
FIG. 1A illustrates a microlens mounted on related art rectangular-shaped photodiodes;
FIG. 1B illustrates an example embodiment microlens mounted on photodiodes having a hexagonal shape;
FIGS. 2A and 2B illustrate an example embodiment pixel pitch in a pixel array;
FIGS. 2C and 2D illustrate a pixel pitch in a related art pixel array including photodiodes having a rectangular shape;
FIG. 3 is a circuit diagram of an example embodiment CMOS image sensor;
FIG. 4 illustrates the layout of an example embodiment CMOS image sensor;
FIG. 5 illustrates the layout of an example embodiment CMOS image sensor;
FIGS. 6A through 6D are timing charts of data readout in an example embodiment CMOS image sensor;
FIGS. 7A and 7B are timing charts of a pixel summation mode of an example embodiment CMOS image sensor;
FIGS. 8A and 8B illustrate interpolation processing in a normal mode of an example embodiment CMOS image sensor;
FIGS. 9A through 9F illustrate an example embodiment method of calculating a pixel value during interpolation in a normal mode of a CMOS image sensor;
FIG. 10 illustrates an example embodiment method of interpolation processing in a pixel summation mode of a CMOS image sensor;
FIGS. 11A and 11B illustrate an example embodiment method of interpolation processing in a normal mode of a CMOS image sensor; and
FIGS. 12A through 12F illustrate an example embodiment method of calculating a pixel value during interpolation in a normal mode of a CMOS image sensor.
DETAILED DESCRIPTION
Detailed example embodiments are disclosed herein. However, specific structural and/or functional details disclosed herein are merely representative for purposes of describing example embodiments. The claims may, however, may be embodied in many alternate forms and should not be construed as limited to only example embodiments set forth herein.
It will be understood that when a component is referred to as being “on,” “connected to” or “coupled to” another component, it can be directly on, connected to or coupled to the other component or intervening components may be present. In contrast, when a component is referred to as being “directly on,” “directly connected to” or “directly coupled to” another component, there are no intervening components present. As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items.
It will be understood that, although the terms first, second, third, etc. may be used herein to describe various elements, components, regions, layers and/or sections, these elements, components, regions, layers and/or sections should not be limited by these terms. These terms are only used to distinguish one element, component, region, layer or section from another element, component, region, layer or section. Thus, a first element, component, region, layer or section discussed below could be termed a second element, component, region, layer or section without departing from the teachings of example embodiments.
Spatially relative terms, such as “beneath,” “below,” “lower,” “above,” “upper” and the like, may be used herein for ease of description to describe one component or feature's relationship to another component(s) or feature(s) as illustrated in the drawings. It will be understood that the spatially relative terms are intended to encompass different orientations of the device in use or operation in addition to the orientation depicted in the figures.
The terminology used herein is for the purpose of describing particular example embodiments only and is not intended to be limiting. As used herein, the singular forms “a,” “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, and/or components.
Unless otherwise defined, all terms (including technical and/or scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art to which example embodiments belong. It will be further understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the relevant art and will not be interpreted in an idealized or overly formal sense unless expressly so defined herein.
Reference will now be made to example embodiments, which are illustrated in the accompanying drawings, wherein like reference numerals refer to the like components throughout.
FIG. 3 is a circuit diagram of an example embodiment CMOS image sensor 100. The CMOS image sensor 100 may include a plurality of unit blocks 10 and 20 that each may include two unit pixels. The two unit pixels may be, for example, a unit pixel including a first photodiode (PD) 130 and/or a unit pixel including a third PD 140. The unit blocks 10 and 20 may have similar structures.
The unit block 10 may include two PDs 130 and 140 having a hexagonal shape, a floating diffusion (FD) 135 shared by the two unit pixels, a first transfer transistor 131 between the PD 130 and the FD 135, a third transfer transistor 141 between the PD 140 and the FD 135, a reset transistor 134 connected with the FD 135, a drive transistor 137 with a gate connected with the FD 135, and/or a selection transistor 136 connected in series with the drive transistor 137. The reset transistor 134 and the drive transistor 137 may be connected with power supply voltages 139 and 138, respectively. If the two unit pixels share the FD 135, the reset transistor 134, the drive transistor 137, and/or the selection transistor 136, the CMOS image sensor 100 may be smaller than a related art CMOS image sensor having similar resolution. The PDs 130 and 140 may receive light and generate photo-generated charge.
The first and third transfer transistors 131 and 141 may transfer photo-generated charge collected at the PDs 130 and 140 to the FD 135. The FD 135 may store the photo-generated charge transferred by the first and third transfer transistors 131 and 141.
The reset transistor 134 may reset the FD 135. The drive transistor 137 may function as a source follower buffer amplifier. The selection transistor 136 may enable addressing through switching. PDs having the hexagonal shape may form a honeycomb structure as illustrated in FIG. 4 or 5.
FIG. 1A illustrates a related art microlens mounted on related art rectangular-shaped PDs. FIG. 1B illustrates an example embodiment microlens mounted on PDs having a hexagonal shape.
As shown in FIGS. 1A and 1B, dead space may be larger if a related art microlens is mounted on PDs having a rectangular shape as illustrated in FIG. 1A than if a microlens is mounted on PDs having a hexagonal shape as illustrated in FIG. 1B. If a hexagonal-shaped PD and a rectangular-shaped PD have the same area, dead space of the hexagonal-shaped PD may be about 13.5% while dead space of the rectangular-shaped PD may be about 21.5%. That is, the dead space of the hexagonal-shaped PD may be decreased by 8% compared to the dead space of the rectangular-shaped PD.
If light emitted from an object is collected by an image pickup lens optical system and focused by a microlens onto a PD, the fill-factor of a hexagonal-shaped PD may be greater than that of a rectangular-shaped PD. If a hexagonal-shaped PD is used, a pitch between pixels may be less than if the rectangular-shaped PD is used.
FIGS. 2A and 2B illustrate an example embodiment pixel pitch in a pixel array. FIGS. 2C and 2D illustrate a pixel pitch in a conventional pixel array including photodiodes having a rectangular shape.
As shown in FIGS. 2A and 2B, if a length of one side of a regular hexagon is represented with “R”, the area of a PD having a regular hexagonal shape is
3 3 2 R 2 .
A pitch between pixels including a PD having a regular hexagonal shape is calculated as
3 2 R .
As shown in FIGS. 2C and 2D, if a length of one side of a square is represented with “X”, the area of a PD having a square shape is calculated as X2. A pitch between pixels including a PD having a square shape is expressed by X.
Table 1 shows the comparison of a pixel pitch between a PD having a regular hexagonal shape and a PD having a square shape when the PDs have the same area.
TABLE 1
Pixel pitch Pixel pitch Length (μm)
Area of (μm) for PD (μm) for PD having of one side
PD (μm2) having square shape regular hexagonal shape of hexagon
31.4 5.6 3.0 3.5
16 4 2.1 2.5
14.4 3.8 2.0 2.4
7.8 2.8 1.5 1.7
As shown in Table 1, for a particular area of PDs, a pixel pitch for a PD having a regular hexagonal shape may be less than that for a PD having a square shape. If an image signal, for example, an image signal having a stripe pattern, is displayed, pixels including a PD having a regular hexagonal shape may display the image more clearly.
If a honeycomb structure is used, the dead space of a PD be reduced, and an image signal having a stripe pattern may be more clearly displayed as well because a pixel pitch in a hexagonal structure may be less than that in a square structure.
FIG. 4 illustrates the layout of an example embodiment CMOS image sensor 100. The circuit diagram illustrated in FIG. 3 corresponds to the layout illustrated in FIG. 4. As shown in FIGS. 3 and 4, if among the two unit pixels in the unit block 10, the first unit pixel includes the PD 130 and the second unit pixel includes the PD 140, the first unit pixel may be on a k-th (for example, 2nd or 5th) column, and an n-th (for example, 1st or 4th) row and the second unit pixel may be on the k-th column and an (n+2)-th (for example, 3rd or 6th) row, in the honeycomb structure. The two unit pixels in each unit block may be on the same column at an interval of two rows. Here, “n” is a natural number less than m−1 if a pixel array includes a total of “m” rows.
Each of the first and second unit pixels may include a microlens (not shown), which may be above the first or second unit pixel to receive light, and/or a color filter (not shown), which may be between the PD 130 or 140 and the microlens to filter the light received by the microlens. Color filters corresponding to the respective PDs 130 and 140 having a hexagonal shape in the unit block 10 may filter the same colors, for example, green and/or red. Two unit pixels in each unit block may include the same color filters, for example, green (G) color filters, red (R) color filters, or blue (B) color filters.
If two unit pixels in a single unit block include the same color filters, pixel summation may be accomplished through an FD shared by the two unit pixels.
A first, second, and third color filter, for example, a blue, red, and green color filter, may be sequenced and/or repeated on a second row in the honeycomb structure. Third, first, and second color filters, for example, green, blue, and red color filters, may be sequenced and/or repeated on a third row in the honeycomb structure.
As shown in FIG. 4, blue (B), red (R), and/or green (G) color filters may be sequenced and/or repeated on a second row and G, B, and/or R color filters may be sequenced and/or repeated on a third row. A sequence of the color filters on the second row may not necessarily be B, R, and/or G. It may be desirable that a color filter on the first column and the second row, a color filter on the third column and the second row, and a color filter on the second column and the third row each filter different colors, for example, B for a PD 120, R for a PD 160, and G for a PD 140.
FIG. 5 illustrates the layout of an example embodiment CMOS image sensor. The layout illustrated in FIG. 5 is similar to that illustrated in FIG. 4, with the exception of the arrangement of color filters. In the honeycomb structure illustrated in FIG. 5, first and second color filters, for example, B and/or R color filters, may be sequenced and/or repeated on the second row and a third color filter, for example, a G color filter, may be repeated on the third row. It may be desirable for the third color to be green, but any color may be used.
Green color filters may be used more than any other color filters in the honeycomb structure illustrate in FIG. 5 because human eyes may be more sensitive to green and thus may perceive that a picture is clearer. Green color filters may absorb ultraviolet rays that may be harmful to human eyes, reduce the dazzling effects caused by intense light, and/or reduce eye fatigue.
As shown in FIG. 3, the unit block 10 may include two overflow transistors 132 and 142, which may be respectively connected with the PDs 130 and 140 to drain the overflow of charge. Among the two overflow transistors 132 and 142, one may be connected with a power supply voltage 133 and the other may be connected with an FD in an adjacent unit block 30.
If a CMOS image sensor is at a white level, a high level intensity of illumination, the overflow of photo-generated charge may occur at each PD. If the overflow occurs, the charge may flow into other unit pixels, disturbing the image display. For example, if photo-generated charge in the first PD 130 overflows to the first FD 135 if the first transfer transistor 131 is not turned on, data may be unclearly read from the unit pixel including the third PD 140 due to the overflow of the photo-generated charge.
In order to prevent or reduce such overflow, the two overflow transistors 132 and 142 may be used to drain the overflow. For such overflow drain function, it may be desirable for the overflow transistors 132 and 142 and second and fourth transfer transistors 121 and 111 in the unit block 20 adjacent to the unit block 10 to perform voltage control.
FIGS. 6A through 6D are timing charts of data readout in an example embodiment CMOS image sensor. FIG. 6A is a timing chart of an operation in which data may be read from the unit pixel including the first PD 130. The overflow drain function will be described with reference to FIG. 6A below.
If data is read from the unit pixel including the first PD 130 in the unit block 10, an entire row including the unit pixel including the first PD 130 may be selected, and the first selection transistor 136 may be turned on by transiting a selection signal SEL1 for controlling the first selection transistor 136 from a low level to a high level and transiting a reset signal RG1 for controlling the first reset transistor 134 from a high level to a low level. PDs that may be read by the first selection transistor 136 may be the two first and third PDs 130 and 140. Only the first transfer transistor 131 connected with the first PD 130 may be turned on. To do so, the first transfer signal TG1 for controlling the first transfer transistor 131 may be at a high level for a period of time. The third overflow transistor 142 may be slightly turned on to prevent or reduce overflow of the third PD 140. To slightly turn on the third overflow transistor 142, a signal OG3 having a voltage between a low level voltage and a high level voltage may be applied to the third overflow transistor 142, as illustrated in FIG. 6A. The voltage may be closer to the low level voltage than to the high level voltage. In order to prevent or reduce the overflow of the two PDs 110 and 120 in the adjacent unit block 20 connected with the first FD 135 in the unit block 10, the second and fourth transfer transistors 121 and 111 may be slightly turned on. For this operation, signals TG2 and TG4 applied respectively to a gate of the second and fourth transfer transistor 121 and 111 may have a voltage level similar to the signal OG3.
If the unit pixel including the first PD 130 is read, only photo-generated charge of the first PD 130 may be stored in the first FD 135. The third overflow transistor 142 in the unit pixel including the third PD 140 may be slightly turned on in the unit block 10, and the two transfer transistors 111 and 121 may be slightly turned on in the adjacent unit block 20 connected to the first FD 135.
FIGS. 6B through 6D are timing charts of operations in which data may be read from unit pixels, of the second, third, and fourth PDs 120, 140, and 110. These operations may be similar to the operation described with reference to FIG. 6A.
FIGS. 7A and 7B are timing charts for a pixel summation mode of an example embodiment CMOS image sensor. If a preview function is performed in an optical system including an example embodiment CMOS image sensor, a high resolution set during actual photographing may not be needed, but quick data readout may be desired, and, therefore, the pixel summation mode may be used.
As shown in FIG. 7A, if the first selection transistor 136 is turned on, photo-generated charge stored in the first FD 135 may be read, and the reset signal RG1 for controlling the first reset transistor 134 may transit to the low level. Because photo-generated charge generated by the first and third PDs 130 and 140 may be transferred to the first FD 135, the first and third transfer transistors 131 and 141 may be turned on. For this operation, the signal TG1 applied to a gate of the first transfer transistor 131 and a signal TG3 applied to a gate of the third transfer transistor 141 may be at a high level. The two transfer transistors 111 and 121 in the adjacent unit block 20 connected with the first FD 135 may be slightly turned on to prevent or reduce overflow.
Data may be simultaneously read from two unit pixels in a unit block in the pixel summation mode, and it may be desirable for two color filters in the unit block filter to be the same color.
FIG. 7B is a timing chart of an operation in which data may be simultaneously read from the second and fourth rows. The operation may be similar to that described with reference to FIG. 7A.
FIGS. 8A and 8B illustrate an example embodiment method of interpolation processing in a normal mode of a CMOS image sensor. Example embodiment methods may provide interpolation processing for achieving a resolution up to about two-fold higher than the number of real pixels. FIG. 8A illustrates an example embodiment real pixel array in a CMOS image sensor. FIG. 8B illustrates a virtual pixel array after interpolation processing is performed. As shown in FIG. 8B, the number of interpolated pixels may be double the number of real pixels, for example, 80 instead of 40. During interpolation processing, a virtual pixel, represented by a blank box, may be generated between real pixels, represented by a dotted box).
FIGS. 9A through 9F illustrate an example embodiment method of calculating data values, for example, RGB values, of each pixel illustrated in FIG. 8B. As shown in FIG. 9A, a B data value of a real green pixel 11 may be generated based on data values of real blue pixels 12, 13, and 14 neighboring the real G pixel 11. For example, the sum of the data values of the real blue pixels 12, 13, and 14 may calculated and then divided by 3. An R data value of the real green pixel 11 may be generated based on data values of real red pixels 15, 16, and 17 neighboring the real green pixel 11.
Example embodiment methods of calculating RGB values of a real blue pixel and a real red pixel, which are illustrated in FIGS. 9B and 9C, respectively, may be similar to the method illustrated in FIG. 9A.
As shown in FIG. 9D, a G data value of a virtual pixel 21 between real green pixels 22 and 23 may be generated based on data values of the real green pixels 22 and 23 neighboring the virtual pixel 21. A B data value of the virtual pixel 21 may be generated based on data values of real blue pixels 24, 25, and 26 neighboring the virtual pixel 21. An R data value of the virtual pixel 21 may be generated based on data values of real red pixels 27, 28, and 29 neighboring the virtual pixel 21.
Example embodiment methods of calculating RGB values of a virtual pixel between real blue pixels and a virtual pixel between real red pixels, which are illustrated in FIGS. 9E and 9F, respectively, may be similar to the method illustrated in FIG. 9D.
FIG. 10 illustrates an example embodiment method of interpolation processing in the pixel summation mode of a CMOS image sensor. As shown in FIG. 10, data of two unit pixels may be summed to generate a single data value in the pixel summation mode, in which only half of the number of all pixels in a real pixel array may exist. If the pixels obtained after pixel summation are subjected to the interpolation processing as illustrated in FIG. 8B, a resolution corresponding to the number of all pixels in the real pixel array may be generated.
FIGS. 11A and 11B illustrate an example embodiment method of interpolation processing in a normal mode of a CMOS image sensor having the color filter layout shown in FIG. 5. The method of the interpolation processing illustrated in FIGS. 11A and 11B may be similar to that illustrated in FIGS. 8A and 8B.
FIGS. 12A through 12F illustrate an example embodiment method of calculating RGB values of each pixel during interpolation in the normal mode of a CMOS image sensor. As shown in FIG. 12A, a B data value of a real green pixel 31 may be generated based on data values of real blue pixels 32 and 33 neighboring the real green pixel 31. An R data value of the real green pixel 31 may be generated based on data values of real red, pixels 34 and 35 neighboring the real green pixel 31.
As shown in FIG. 12B, a G data value of a blue green pixel 41 may be generated based on data values of real green pixels 42, 43, 44, and 45 neighboring the real blue pixel 41. An R data value of the real blue pixel 41 may be generated based on data values of real red pixels 46 and 47 neighboring the real blue pixel 41.
An example embodiment method of calculating RGB values of a real red pixel, which is illustrated in FIG. 12C, may be similar to the method illustrated in FIG. 12B.
As shown in FIG. 12D, a G data value of a virtual pixel 61 or 66 between two real green pixels 63 and 65 or 68 and 69 and between a real blue pixel 62 or 67 and a real red pixel 64 may be generated based on data values of the real green pixels 63 and 65 or 68 and 69 neighboring the virtual pixel 61 or 66. A B data value of the virtual pixel 61 or 66 may be generated based on a data value of the real blue pixel 62 or 67 neighboring the virtual pixel 61 or 66. An R data value of the virtual pixel 61 or 66 may be generated based on a data value of the real red pixel 64 neighboring the virtual pixel 61 or 66.
As shown in FIG. 12E, a G data value of a virtual pixel 51 between real green pixels 52 and 53 and between real blue pixels 54 and 55 may be generated based on data values of the real green pixels 52 and 53 neighboring the virtual pixel 51. A B data value of the virtual pixel 51 may be generated based on data values of the real blue pixels 54 and 55 neighboring the virtual pixel 51. An R data value of the virtual pixel 51 may be generated based on data values of real red pixels 56, 57, 58, and 59 neighboring the virtual pixel 51.
An example embodiment method of calculating RGB values of a virtual pixel between real green pixels and between real red pixels, which is illustrated in FIG. 12F, may be similar to the method illustrated in FIG. 12E.
As described above, example embodiments provide devices that may use a shared floating diffusion design in which unit pixels may share a circuit, potentially increasing the degree of integration and resolution of those devices. Example embodiments may provide a pixel summation mode and an overflow drain at a white level.
For example, an example embodiment CMOS image sensor may be used in digital cameras, computer cameras, mobile terminals, and/or any other suitable application.
While example embodiments have been particularly shown and described with reference to example embodiments thereof, it will be understood by those of ordinary skill in the art that various changes in form and details may be made therein without departing from the spirit and scope of example embodiments as defined by the following claims.

Claims (21)

1. A CMOS image sensor having a plurality of unit blocks, each unit block of the plurality of unit blocks comprising:
a plurality of photodiodes having a hexagonal shape, the plurality of photodiodes forming a honeycomb structure;
two unit pixels including,
a first unit pixel on a k-th column and an n-th row in the honeycomb structure, “k” and “n” being positive integers, and
a second unit pixel on the k-th column and an (n+2)-th row in the honeycomb structure;
a floating diffusion shared by the two unit pixels;
a first transfer transistor and a second transfer transistor, each between the floating diffusion and one of the photodiodes;
a reset transistor connected to the floating diffusion;
a drive transistor with a gate connected to the floating diffusion; and
a selection transistor connected with the drive transistor in series.
2. The CMOS image sensor of claim 1, wherein each unit block includes two overflow transistors, wherein a first and a second overflow transistor of the two overflow transistors are respectively connected to a first and a second photodiode, wherein the first and the second overflow transistors are configured to provide overflow drains.
3. The CMOS image sensor of claim 2, wherein the first overflow transistor of the two overflow transistors is connected with a power supply voltage and the second overflow transistor of the two overflow transistors is connected to a floating diffusion in an adjacent unit block of the plurality of unit blocks.
4. The CMOS image sensor of claim 3, wherein one of the two overflow transistors is configured to partially turn on and drain overflow in the second unit pixel during a data reading operation and wherein the first and second transfer transistors in the adjacent unit block are slightly turned on and drain overflow from the adjacent unit block.
5. The CMOS image sensor of claim 1, wherein each of the unit pixels includes a microlens above one of the photodiodes configured to receive light and a color filter between the photodiode and the microlens configured to filter the light received by the microlens.
6. The CMOS image sensor of claim 5, wherein the color filter filters the same color as the photodiodes in the unit block.
7. The CMOS image sensor of claim 6, wherein color filters for a first color, a second color, and a third color are sequenced and repeated on an n-th row in a honeycomb structure and color filters for the third color, the first color, and the second color are sequenced and repeated on an (n+1)-th row in the honeycomb structure, wherein “n” is a positive integer.
8. The CMOS image sensor of claim 6, wherein color filters for a first color and a second color are sequenced and repeated on an n-th row in a honeycomb structure and color filters for a third color are sequenced and repeated on an (n+1)-th row in the honeycomb structure, wherein “n” is a positive integer.
9. The CMOS image sensor of claim 7, wherein the first, second, and third colors are blue, red, and green, respectively.
10. The CMOS image sensor of claim 8, wherein the first, second, and third colors are blue, red, and green, respectively.
11. A CMOS image sensor having a plurality of unit blocks, each unit block of the plurality of unit blocks comprising:
two unit pixels;
two photodiodes having a hexagonal shape;
a floating diffusion shared by the two unit pixels;
a first transfer transistor and a second transfer transistor between the floating diffusion and the two photodiodes, respectively;
a reset transistor connected to the floating diffusion;
a drive transistor with a gate connected to the floating diffusion; and
a selection transistor connected with the drive transistor in series, wherein a plurality of the two photodiodes having the hexagonal shape form a honeycomb structure and wherein the CMOS image sensor sums charge generated by the two photodiodes in each unit block of the plurality of unit blocks and outputs a voltage generated by the summed charge as an output voltage during a reading operation on each unit block of the plurality of unit blocks to support a pixel summation mode.
12. The CMOS image sensor of claim 11, wherein each unit block of the plurality of unit blocks further includes two overflow transistors, wherein a first and a second overflow transistor of the two overflow transistors are respectively connected to a first and a second photodiode of the two photodiodes, wherein the first and the second overflow transistors are configured to provide overflow drains, and wherein one transistor of the two overflow transistors is connected with a power supply voltage, and another transistor of the two overflow transistors is connected with a floating diffusion in an adjacent unit block of the plurality of unit blocks.
13. The CMOS image sensor of claim 12, wherein the first and second transfer transistors in the adjacent unit block are configured to slightly turned on to and drain overflow during a reading operation in the adjacent unit block to support the pixel summation mode.
14. A digital camera comprising:
the CMOS image sensor of claim 1; and
an image processor receiving data output from the CMOS image sensor and performing image processing of the received data.
15. A digital camera comprising:
the CMOS image sensor of claim 11; and
an image processor receiving data output from the CMOS image sensor and performing image processing of the received data.
16. A computer camera comprising:
the CMOS image sensor of claim 1; and
an image processor receiving data output from the CMOS image sensor and performing image processing of the received data.
17. A computer camera comprising:
the CMOS image sensor of claim 11; and
an image processor receiving data output from the CMOS image sensor and performing image processing of the received data.
18. A mobile terminal comprising:
the CMOS image sensor of claim 1; and
an image processor receiving data output from the CMOS image sensor and performing image processing of the received data.
19. A mobile terminal comprising:
the CMOS image sensor of claim 11; and
an image processor receiving data output from the CMOS image sensor and performing image processing of the received data.
20. An image sensing method for a CMOS image sensor comprising:
selecting a first unit pixel;
transferring charge generated by a photodiode in the first unit pixel to a first floating diffusion in a unit block including the first unit pixel;
transferring an amount of charge generated by photodiodes in two unit pixels in a first adjacent unit block connected to the first floating diffusion to a second floating diffusion in the first adjacent unit block to drain overflow;
transferring an amount of charge generated by a photodiode in a second unit pixel to a power supply voltage or a floating diffusion in a second adjacent unit block to drain overflow; and
reading charge from the first floating diffusion.
21. An image sensing method for a CMOS image sensor comprising:
selecting a unit block;
transferring charge generated by two photodiodes in the selected unit block to a first floating diffusion in the unit block;
transferring an amount of charge generated by photodiodes in two unit pixels in an adjacent unit block connected to the first floating diffusion to a second floating diffusion in the adjacent unit block; and
reading charge from the first floating diffusion.
US11/826,588 2006-07-19 2007-07-17 CMOS image sensor and image sensing method using the same Active 2028-03-28 US7825970B2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR10-2006-0067198 2006-07-19
KR1020060067198A KR100818724B1 (en) 2006-07-19 2006-07-19 CMOS image sensor and sensing method thereof

Publications (2)

Publication Number Publication Date
US20080018765A1 US20080018765A1 (en) 2008-01-24
US7825970B2 true US7825970B2 (en) 2010-11-02

Family

ID=38971067

Family Applications (1)

Application Number Title Priority Date Filing Date
US11/826,588 Active 2028-03-28 US7825970B2 (en) 2006-07-19 2007-07-17 CMOS image sensor and image sensing method using the same

Country Status (3)

Country Link
US (1) US7825970B2 (en)
KR (1) KR100818724B1 (en)
CN (1) CN101110439B (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110278653A1 (en) * 2009-02-04 2011-11-17 Rosnes Corporation Solid-state image pickup device
US8547458B2 (en) 2009-02-04 2013-10-01 Rosnes Corporation Solid-state image pickup device
US20140307140A1 (en) * 2011-12-27 2014-10-16 Fujifilm Corporation Color imaging apparatus
US20150268392A1 (en) * 2014-03-20 2015-09-24 Kabushiki Kaisha Toshiba Filter-array-equipped microlens and solid-state imaging device
US9793310B2 (en) 2015-03-11 2017-10-17 Samsung Electronics Co., Ltd. Image sensor devices using offset pixel patterns
US9953574B2 (en) 2015-04-28 2018-04-24 Microsoft Technology Licensing, Llc Sub-pixel compensation
US11018173B2 (en) 2018-06-18 2021-05-25 Samsung Electronics Co., Ltd. Image sensor
US11323643B2 (en) 2018-09-12 2022-05-03 Dolby Laboratories Licensing Corporation CMOS sensor architecture for temporal dithered sampling

Families Citing this family (35)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7825966B2 (en) * 2007-06-29 2010-11-02 Omnivision Technologies, Inc. High dynamic range sensor with blooming drain
JP4952601B2 (en) * 2008-02-04 2012-06-13 日本テキサス・インスツルメンツ株式会社 Solid-state imaging device
US7781716B2 (en) * 2008-03-17 2010-08-24 Eastman Kodak Company Stacked image sensor with shared diffusion regions in respective dropped pixel positions of a pixel array
US8035711B2 (en) 2008-05-22 2011-10-11 Panavision Imaging, Llc Sub-pixel array optical sensor
US20100149393A1 (en) * 2008-05-22 2010-06-17 Panavision Imaging, Llc Increasing the resolution of color sub-pixel arrays
US8130302B2 (en) * 2008-11-07 2012-03-06 Aptina Imaging Corporation Methods and apparatus providing selective binning of pixel circuits
JP4760915B2 (en) * 2009-01-08 2011-08-31 ソニー株式会社 Solid-state image sensor
US8350940B2 (en) * 2009-06-08 2013-01-08 Aptina Imaging Corporation Image sensors and color filter arrays for charge summing and interlaced readout modes
US20110205384A1 (en) * 2010-02-24 2011-08-25 Panavision Imaging, Llc Variable active image area image sensor
FR2968877A1 (en) * 2010-12-14 2012-06-15 St Microelectronics Crolles 2 Image sensor for detecting color on surface of substrate semiconductor, has pixels arranged in rows adjacent to each other based on pixel width, where pixels of adjacent rows are being offset relative to each other than half of pixel length
JP5868975B2 (en) * 2011-08-14 2016-02-24 富士フイルム株式会社 Radiation image capturing apparatus and radiation image capturing method
KR101868145B1 (en) * 2011-10-06 2018-06-18 엘지디스플레이 주식회사 Stereoscopic image display
JP6035744B2 (en) * 2012-01-10 2016-11-30 凸版印刷株式会社 Solid-state image sensor
CN102595057B (en) * 2012-02-27 2014-09-24 北京思比科微电子技术股份有限公司 CMOS (complementary metal-oxide-semiconductor transistor) image sensor pixel and control time sequence thereof
US10832616B2 (en) 2012-03-06 2020-11-10 Samsung Display Co., Ltd. Pixel arrangement structure for organic light emitting diode display
KR101615332B1 (en) 2012-03-06 2016-04-26 삼성디스플레이 주식회사 Pixel arrangement structure for organic light emitting display device
US20140048749A1 (en) * 2012-08-16 2014-02-20 Nthdegree Technologies Worldwide Inc. Conductive Ink Composition
US10134788B2 (en) 2013-09-17 2018-11-20 Omnivision Technologies, Inc. Dual VPIN HDR image sensor pixel
CN104410847B (en) * 2014-12-05 2016-08-31 林立果 A kind of chromatic filter and color image sensor
US9743015B2 (en) * 2015-05-22 2017-08-22 Samsung Electronics Co., Ltd. Image capturing apparatus and method of controlling the same
KR102414030B1 (en) * 2015-11-18 2022-06-29 에스케이하이닉스 주식회사 Image sensor
JP6969539B2 (en) * 2016-03-31 2021-11-24 ソニーグループ株式会社 Solid-state image sensor, solid-state image sensor drive method, and electronic equipment
CN107295277A (en) * 2016-04-13 2017-10-24 合肥芯福传感器技术有限公司 Increase honeycomb pixel array, imaging sensor and the acquisition method of picture based on dislocation
KR102524400B1 (en) * 2016-07-04 2023-04-24 에스케이하이닉스 주식회사 Image Sensors Having a Plurality of Photodiodes Sharing One Color Filter and One Micro-lens
DE102016212771A1 (en) * 2016-07-13 2018-01-18 Robert Bosch Gmbh Method and device for scanning a light sensor
JP6311771B2 (en) * 2016-10-31 2018-04-18 凸版印刷株式会社 Solid-state image sensor
KR102406996B1 (en) * 2017-04-07 2022-06-08 삼성전자주식회사 Image Sensor
US20180301484A1 (en) * 2017-04-17 2018-10-18 Semiconductor Components Industries, Llc Image sensors with high dynamic range and autofocusing hexagonal pixels
US10931902B2 (en) 2018-05-08 2021-02-23 Semiconductor Components Industries, Llc Image sensors with non-rectilinear image pixel arrays
CN110223994A (en) * 2019-06-05 2019-09-10 芯盟科技有限公司 Pixel group and imaging sensor
CN110324540A (en) * 2019-06-10 2019-10-11 芯盟科技有限公司 The forming method and electronic equipment of a kind of imaging sensor, imaging sensor
CN110379824A (en) * 2019-07-08 2019-10-25 Oppo广东移动通信有限公司 A kind of cmos image sensor and image processing method, storage medium
US11469264B2 (en) * 2020-01-30 2022-10-11 Omnivision Technologies, Inc. Flare-blocking image sensor
US11393861B2 (en) * 2020-01-30 2022-07-19 Omnivision Technologies, Inc. Flare-suppressing image sensor
CN115866423A (en) * 2021-09-22 2023-03-28 华为技术有限公司 Image sensor, camera module, electronic equipment and image processing method

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6657665B1 (en) 1998-12-31 2003-12-02 Eastman Kodak Company Active Pixel Sensor with wired floating diffusions and shared amplifier
KR20040006486A (en) 2002-07-12 2004-01-24 주식회사 하이닉스반도체 Method for aligning color filter array
JP2004056048A (en) 2002-07-24 2004-02-19 Microsignal Kk Solid imaging device
US20040218078A1 (en) 2003-04-30 2004-11-04 Won-Ho Lee Complementary metal oxide semiconductor image sensor with multi-floating diffusion region
US20050110093A1 (en) * 2003-11-26 2005-05-26 Altice Peter P.Jr. Anti-blooming storage pixel
US20060118837A1 (en) * 2004-12-07 2006-06-08 Sung-Ho Choi Active pixel sensor having honeycomb structure

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4497688B2 (en) 1999-09-27 2010-07-07 富士フイルム株式会社 Solid-state imaging device
JP4139672B2 (en) 2002-11-01 2008-08-27 大日本印刷株式会社 Interpolation apparatus and interpolation method of captured image by single-plate color digital camera

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6657665B1 (en) 1998-12-31 2003-12-02 Eastman Kodak Company Active Pixel Sensor with wired floating diffusions and shared amplifier
KR20040006486A (en) 2002-07-12 2004-01-24 주식회사 하이닉스반도체 Method for aligning color filter array
JP2004056048A (en) 2002-07-24 2004-02-19 Microsignal Kk Solid imaging device
US20040218078A1 (en) 2003-04-30 2004-11-04 Won-Ho Lee Complementary metal oxide semiconductor image sensor with multi-floating diffusion region
KR20040093947A (en) 2003-04-30 2004-11-09 매그나칩 반도체 유한회사 Cmos image sensor with multi floating diffusion region
JP2004336006A (en) 2003-04-30 2004-11-25 Hynix Semiconductor Inc Unit pixel of cmos image sensor having a plurality of floating diffusion regions
US20050110093A1 (en) * 2003-11-26 2005-05-26 Altice Peter P.Jr. Anti-blooming storage pixel
US20060118837A1 (en) * 2004-12-07 2006-06-08 Sung-Ho Choi Active pixel sensor having honeycomb structure
KR20060063244A (en) 2004-12-07 2006-06-12 삼성전자주식회사 Active pixel sensor of honeycomb structure

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
English translation of Korean Office Action dated Jul. 18, 2007.
Korean Office Action dated Jul. 18, 2007.

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110278653A1 (en) * 2009-02-04 2011-11-17 Rosnes Corporation Solid-state image pickup device
US8547458B2 (en) 2009-02-04 2013-10-01 Rosnes Corporation Solid-state image pickup device
US8576312B2 (en) * 2009-02-04 2013-11-05 Rosnes Corporation Solid-state image pickup device with particular pixel arrangement
US20140307140A1 (en) * 2011-12-27 2014-10-16 Fujifilm Corporation Color imaging apparatus
US9036061B2 (en) * 2011-12-27 2015-05-19 Fujifilm Corporation Color imaging apparatus
US20150268392A1 (en) * 2014-03-20 2015-09-24 Kabushiki Kaisha Toshiba Filter-array-equipped microlens and solid-state imaging device
US9793310B2 (en) 2015-03-11 2017-10-17 Samsung Electronics Co., Ltd. Image sensor devices using offset pixel patterns
US9953574B2 (en) 2015-04-28 2018-04-24 Microsoft Technology Licensing, Llc Sub-pixel compensation
US10650750B2 (en) 2015-04-28 2020-05-12 Microsoft Technology Licensing, Llc Sub-pixel compensation
US11018173B2 (en) 2018-06-18 2021-05-25 Samsung Electronics Co., Ltd. Image sensor
US11323643B2 (en) 2018-09-12 2022-05-03 Dolby Laboratories Licensing Corporation CMOS sensor architecture for temporal dithered sampling

Also Published As

Publication number Publication date
KR100818724B1 (en) 2008-04-01
CN101110439B (en) 2010-12-15
KR20080007937A (en) 2008-01-23
US20080018765A1 (en) 2008-01-24
CN101110439A (en) 2008-01-23

Similar Documents

Publication Publication Date Title
US7825970B2 (en) CMOS image sensor and image sensing method using the same
US20240006427A1 (en) Imaging device and imaging system
US9866771B2 (en) Solid-state imaging device, signal processing method of solid-state imaging device, and electronic apparatus
CN110113546B (en) Imaging system and method for combining and reading out adjacent pixel units in pixel array
RU2490715C1 (en) Image capturing device
US9560325B2 (en) Imaging device camera system and driving method of the same
US7916195B2 (en) Solid-state imaging device, imaging apparatus and camera
JP4768305B2 (en) Solid-state imaging device
US20100110243A1 (en) Solid state imaging device
US20110128425A1 (en) Cmos image sensor with selectable hard-wired binning
US20130033631A1 (en) Solid-state imaging device and imaging device
WO2013172205A1 (en) Imaging device and imaging method, electronic apparatus, as well as program
CN111741242A (en) Image sensor and method of operating the same
CN113242344B (en) Image sensor, camera assembly and mobile terminal
KR20120140609A (en) Solid-state imaging device, method of driving the same, and electronic system
US11616934B2 (en) Image sensor
US20210104562A1 (en) Image sensor, focus adjustment device, and imaging device
US9007498B2 (en) Solid state imaging device and imaging apparatus having a plurality of addition read modes
JP5526342B2 (en) Solid-state imaging device
JP2007166486A (en) Solid-state imaging apparatus
US11463640B1 (en) Bitline control supporting binning mode phase detection autofocus photodiodes
CN110557587B (en) Image sensor and method of operating the same
US11736821B1 (en) Four shared pixel with phase detection and horizontal and vertical binning readout modes
US20240056699A1 (en) Imaging device and electronic apparatus
US20220384513A1 (en) Image sensor

Legal Events

Date Code Title Description
AS Assignment

Owner name: SAMSUNG ELECTRONICS CO., LTD., KOREA, REPUBLIC OF

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CHOI, SUNG-HO;AHN, JUNG-CHAK;KIM, YI-TAE;AND OTHERS;REEL/FRAME:019642/0537

Effective date: 20070611

STCF Information on status: patent grant

Free format text: PATENTED CASE

FPAY Fee payment

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552)

Year of fee payment: 8

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 12