US20180007289A1 - Image sensor, imaging device, mobile terminal and imaging method - Google Patents

Image sensor, imaging device, mobile terminal and imaging method Download PDF

Info

Publication number
US20180007289A1
US20180007289A1 US15/544,537 US201615544537A US2018007289A1 US 20180007289 A1 US20180007289 A1 US 20180007289A1 US 201615544537 A US201615544537 A US 201615544537A US 2018007289 A1 US2018007289 A1 US 2018007289A1
Authority
US
United States
Prior art keywords
filter
merged
photosensitive
pixel
pixels
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US15/544,537
Other versions
US10594962B2 (en
Inventor
Jian Kang
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong Oppo Mobile Telecommunications Corp Ltd
Original Assignee
Guangdong Oppo Mobile Telecommunications Corp Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong Oppo Mobile Telecommunications Corp Ltd filed Critical Guangdong Oppo Mobile Telecommunications Corp Ltd
Assigned to GUANGDONG OPPO MOBILE TELECOMMUNICATIONS CORP., LTD. reassignment GUANGDONG OPPO MOBILE TELECOMMUNICATIONS CORP., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: KANG, Jian
Publication of US20180007289A1 publication Critical patent/US20180007289A1/en
Application granted granted Critical
Publication of US10594962B2 publication Critical patent/US10594962B2/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • H04N5/357
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01LSEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
    • H01L27/00Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
    • H01L27/14Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
    • H01L27/144Devices controlled by radiation
    • H01L27/146Imager structures
    • H01L27/14601Structural or functional details thereof
    • H01L27/14603Special geometry or disposition of pixel-elements, address-lines or gate-electrodes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N25/00Circuitry of solid-state image sensors [SSIS]; Control thereof
    • H04N25/60Noise processing, e.g. detecting, correcting, reducing or removing noise
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/10Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths
    • H04N23/12Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths with one sensor only
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01LSEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
    • H01L27/00Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
    • H01L27/14Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
    • H01L27/144Devices controlled by radiation
    • H01L27/146Imager structures
    • H01L27/14601Structural or functional details thereof
    • H01L27/1462Coatings
    • H01L27/14621Colour filter arrangements
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01LSEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
    • H01L27/00Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
    • H01L27/14Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
    • H01L27/144Devices controlled by radiation
    • H01L27/146Imager structures
    • H01L27/14601Structural or functional details thereof
    • H01L27/14625Optical elements or arrangements associated with the device
    • H01L27/14627Microlenses
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01LSEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
    • H01L27/00Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
    • H01L27/14Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
    • H01L27/144Devices controlled by radiation
    • H01L27/146Imager structures
    • H01L27/14643Photodiode arrays; MOS imagers
    • H01L27/14645Colour imagers
    • HELECTRICITY
    • H01ELECTRIC ELEMENTS
    • H01LSEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
    • H01L27/00Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate
    • H01L27/14Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components sensitive to infrared radiation, light, electromagnetic radiation of shorter wavelength or corpuscular radiation and specially adapted either for the conversion of the energy of such radiation into electrical energy or for the control of electrical energy by such radiation
    • H01L27/144Devices controlled by radiation
    • H01L27/146Imager structures
    • H01L27/14683Processes or apparatus peculiar to the manufacture or treatment of these devices or parts thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/57Mechanical or electrical details of cameras or camera modules specially adapted for being embedded in other devices
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N25/00Circuitry of solid-state image sensors [SSIS]; Control thereof
    • H04N25/10Circuitry of solid-state image sensors [SSIS]; Control thereof for transforming different wavelengths into image signals
    • H04N25/11Arrangement of colour filter arrays [CFA]; Filter mosaics
    • H04N25/13Arrangement of colour filter arrays [CFA]; Filter mosaics characterised by the spectral characteristics of the filter elements
    • H04N25/133Arrangement of colour filter arrays [CFA]; Filter mosaics characterised by the spectral characteristics of the filter elements including elements passing panchromatic light, e.g. filters passing white light
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N25/00Circuitry of solid-state image sensors [SSIS]; Control thereof
    • H04N25/10Circuitry of solid-state image sensors [SSIS]; Control thereof for transforming different wavelengths into image signals
    • H04N25/11Arrangement of colour filter arrays [CFA]; Filter mosaics
    • H04N25/13Arrangement of colour filter arrays [CFA]; Filter mosaics characterised by the spectral characteristics of the filter elements
    • H04N25/135Arrangement of colour filter arrays [CFA]; Filter mosaics characterised by the spectral characteristics of the filter elements based on four or more different wavelength filter elements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N25/00Circuitry of solid-state image sensors [SSIS]; Control thereof
    • H04N25/40Extracting pixel data from image sensors by controlling scanning circuits, e.g. by modifying the number of pixels sampled or to be sampled
    • H04N25/46Extracting pixel data from image sensors by controlling scanning circuits, e.g. by modifying the number of pixels sampled or to be sampled by combining or binning pixels
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N25/00Circuitry of solid-state image sensors [SSIS]; Control thereof
    • H04N25/60Noise processing, e.g. detecting, correcting, reducing or removing noise
    • H04N25/62Detection or reduction of noise due to excess charges produced by the exposure, e.g. smear, blooming, ghost image, crosstalk or leakage between pixels
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N25/00Circuitry of solid-state image sensors [SSIS]; Control thereof
    • H04N25/70SSIS architectures; Circuits associated therewith
    • H04N25/71Charge-coupled device [CCD] sensors; Charge-transfer registers specially adapted for CCD sensors
    • H04N25/75Circuitry for providing, modifying or processing image signals from the pixel array
    • H04N5/378
    • H04N9/07

Definitions

  • the present disclosure relates to imaging technologies, and more particularly to an image sensor, an imaging device, a mobile terminal, and an imaging method of the image sensor.
  • the imaging sensor of the imaging device may have a problem of larger noise and lower resolution in an image generated in low light.
  • the present disclosure aims to solve at least one of the problems existing in the related art to at least some extent.
  • the image sensor includes: a photosensitive pixel array and a filter arranged on the photosensitive pixel array.
  • the filter comprises a filer unit array having a plurality of filter units, wherein each filter unit covers N photosensitive pixels, some of the filter units at least include white filter areas.
  • the white filter areas cover at least one of the N photosensitive pixels.
  • a merged pixel is formed by the N photosensitive pixels covered by the same filter unit, wherein N is a positive integer.
  • the amount of light entering into the image sensor can be increased through embedding the white filter areas into some of the filter units, so that a higher signal to noise ratio, higher brightness, higher sharpness and less noise image can be obtained in low light.
  • the filter unit array includes R filter units, G filter units, and B filter units, wherein the G filter units at least include the white filter areas, and the white filter areas cover at least one of the N photosensitive pixels covered by the G filter units.
  • each filter unit includes 2*2 photosensitive pixels, wherein the white filter areas cover one photosensitive pixel covered by the G filter unit, and the G filter unit further includes green filter areas, wherein the green filter areas cover the other three photosensitive pixels.
  • the white filter areas cover two photosensitive pixels covered by the G filter unit, and the G filter unit further includes green filter areas, wherein the green filter areas cover the other two photosensitive pixels.
  • the white filter areas cover three photosensitive pixels covered by the G filter unit, and the G filter unit further includes green filter areas, wherein the green filter areas cover the other one photosensitive pixel.
  • the white filter areas cover all of the four photosensitive pixels covered by the G filter unit.
  • the image sensor further includes a control module, which is configured to control the photosensitive pixel array to be exposed by row.
  • the image sensor further includes a register.
  • the image sensor further includes an ADC (Analog-to-digital converter) array which includes a plurality of ADCs, wherein each of the ADCs connects to one photosensitive pixel.
  • ADC Analog-to-digital converter
  • the image sensor further includes a micro-lens array, each micro-lens corresponds to one photosensitive pixel.
  • an imaging device including the image sensor.
  • the imaging device including the image sensor according to embodiments of the present disclosure includes the above-identified image sensor and an image signal processor, wherein the image signal processor is configured to read and to process the outputs of the photosensitive pixel array of the image sensor in order to obtain the pixel value of the merged pixels and further to obtain merged images.
  • the amount of light entering into the image sensor can be increased by the provided image sensor mentioned above, so that a higher signal to noise ratio, higher brightness, higher sharpness and less noise image can be obtained in low light.
  • the image signal processor is further configured to add the outputs of the N photosensitive pixels corresponding to the same merged pixel together and to make the result as the pixel value of the merged pixel.
  • the image signal processor is further configured to add the outputs of the photosensitive pixels corresponding to the white filter areas together to obtain a first pixel value of the merged pixel, and to add the outputs of the photosensitive pixels corresponding to the non-white filter areas together to obtain a second pixel value of the merged pixel.
  • the noise of merged pixels is less than the sum of noise of all the photosensitive pixels before being merged, it could improve the signal to noise ratio, brightness, and sharpness, and reduce noise of an image in low light via the merged pixels.
  • embodiments of a third aspect of the present disclosure provide a mobile terminal.
  • the mobile terminal includes the above-identified imaging device for a mobile terminal.
  • the amount of light entering into the imaging device can be increased by the provided imaging device mentioned above, so that a higher signal to noise ratio, higher brightness, higher sharpness and less noise image can be obtained in low light, and further, as the noise of merged pixels is less than the sum of noise of all the photosensitive pixels, it could improve the signal to noise ratio, brightness, and sharpness, and reduce noise of an image in low light via the merged pixels.
  • the mobile terminal includes a cell phone.
  • the imaging device includes front cameras.
  • the mobile terminal further comprises a CPU and an external memory, which are connected with the imaging device separately, wherein the CPU is configured to control the external memory to store the merged images.
  • the mobile terminal further includes a display device, which is connected with the imaging device, wherein the CPU is configured to control the display device to display the merged images.
  • embodiments of a fourth aspect of the present disclosure provide an imaging method of the image sensor.
  • the imaging method includes: reading outputs of the photosensitive unit array; calculating pixel value of the merged pixel according to the outputs of the photosensitive pixels of the same merged pixel to obtain a merged image.
  • the amount of light entering into the imaging device can be increased by the provided image sensor mentioned above, so that a higher signal to noise ratio, higher brightness, higher sharpness and less noise image can be obtained in low light.
  • each filter unit includes 2*2 photosensitive pixels.
  • the step of calculating pixel value of the merged pixel according to the outputs of the photosensitive pixels of the same merged pixel further comprises: adding the outputs of the N photosensitive pixels corresponding to the same merged pixel together and making the result as the pixel value of the merged pixel.
  • the pixel value of the merged pixel when the filter units include the white filter areas and the non-white filter areas, the pixel value of the merged pixel includes a first pixel value corresponding to the white filter areas and a second pixel value corresponding to the non-white filter areas.
  • the step of calculating pixel value of the merged pixel according to the outputs of the photosensitive pixels of the same merged pixel further comprises: adding the outputs of the photosensitive pixels corresponding to the white filter areas together to obtain the first pixel value of the merged pixel; and adding the outputs of the photosensitive pixels corresponding to the non-white filter areas together to obtain the second pixel value of the merged pixel.
  • the noise of merged pixels is less than the sum of noise of all the photosensitive pixels, it could improve the signal to noise ratio, brightness, and sharpness, and reduce noise of an image in low light via the merged pixels.
  • each ADC connects to one photosensitive pixel
  • the imaging method further includes: converting analog signal outputs of the photosensitive pixel into digital signal outputs; and calculating the pixel value of the merged pixel according to the digital signal outputs of the photosensitive pixels of the same merged pixel.
  • FIG. 1 is a schematic side view of an exemplary image sensor according to an embodiment of the present disclosure
  • FIG. 2 a -2 d are schematic views of an exemplary filter units of the image sensor according to an embodiment of the present disclosure
  • FIG. 3 is a schematic view of an exemplary Bayer Color Filter Array according to an embodiment of the present disclosure
  • FIG. 4 a -4 d are schematic views of an exemplary filter unit array of the image sensor according to an embodiment of the present disclosure
  • FIG. 5 a -5 d are schematic views of structure of an exemplary image sensor according to an embodiment of the present disclosure.
  • FIG. 6 is a schematic block diagram of an exemplary image sensor according to an embodiment of the present disclosure.
  • FIG. 7 is a schematic circuit diagram of an exemplary photosensitive pixel of the image sensor according to an embodiment of the present disclosure.
  • FIG. 8 is a schematic block diagram of an exemplary image sensor according to an embodiment of the present disclosure.
  • FIG. 9 is a schematic view of structure of an exemplary image sensor according to an embodiment of the present disclosure.
  • FIG. 10 is a schematic block diagram of an exemplary imaging device according to an embodiment of the present disclosure.
  • FIG. 11 is a schematic block diagram of an exemplary mobile terminal according to an embodiment of the present disclosure.
  • FIG. 12 is a schematic block diagram of an exemplary mobile terminal according to an embodiment of the present disclosure.
  • FIG. 13 is a schematic flow chart of an exemplary imaging method according to an embodiment of the present disclosure.
  • FIG. 14 is a schematic flow chart of an exemplary reading step of the imaging method according to an embodiment of the present disclosure.
  • FIG. 15 is a schematic flow chart of an exemplary imaging method according to an embodiment of the present disclosure.
  • a structure in which a first feature is “on” a second feature may include an embodiment in which the first feature directly contacts the second feature, and may also include an embodiment in which an additional feature is formed between the first feature and the second feature so that the first feature does not directly contact the second feature.
  • the terms “mounted,” “connected,” and “coupled” and variations thereof are used broadly and encompass such as mechanical or electrical mountings, connections and couplings, also can be inner mountings, connections and couplings of two components, and further can be direct and indirect mountings, connections, and couplings, which can be understood by those skilled in the art according to the detail embodiment of the present disclosure.
  • An image sensor, an imaging device, a mobile terminal and an imaging method, according to embodiments of the present disclosure, may be described below with reference to FIGS. 1-15 .
  • the image sensor 10 includes a photosensitive pixel array 11 and a filter 13 .
  • the filter 13 is arranged on the photosensitive pixel array 11 and includes a filter unit array 131 .
  • the filter unit array 131 includes a plurality of filter units 1311 .
  • Each filter unit 1311 covers N photosensitive pixels 111 .
  • Some of the filter units 1311 at least include a white filter area 1313 .
  • the white filter area 1313 covers at least one of the N photosensitive pixel 111 .
  • the N photosensitive pixels 111 covered by the same filter unit 1311 forms a merged pixel, and N is a positive integer.
  • External light illuminates the photosensitive part 1111 of a photosensitive pixel 111 to generate electric signal, which is the output of a photosensitive pixel 111 .
  • the white filter areas 1313 is configured to allow the light pass through instead of being filtered.
  • the white filter areas 1313 may refers to either the areas covered by the transparent filters or the areas with no filters, the areas with no filters is the hollow of the filter 13 .
  • the other filter units 1311 except the filter units 1311 with the white filter area 1313 include non-white filter areas, such as green filter area, red filter area or blue filter area.
  • the filter units 1311 with white filter area 1313 further include the non-white filter area.
  • some filter units 1311 consist of the white filter area and the non-white filter area at the same time.
  • the two areas cover N photosensitive pixels together.
  • the non-white filter area is configured to obtain color information of the merged pixels
  • the white filter area is configured to obtain white light information.
  • the white filter area may allow the light pass through. Such that, the brightness is higher which is outputted by the photosensitive with the better effect of light transmission of the white filter area.
  • White filter area is further configured to obtain the brightness information of the merged pixels in low light, and the noise of the brightness information is less.
  • the amount of light entering into the image sensor can be increased through embedding the white filter areas into some of the filter units, so that a higher signal to noise ratio, higher brightness, higher sharpness and less noise image can be obtained in low light.
  • the filter unit array shown in FIG. 3 is arranged according to the Bayer Color Filter Array (Bayer Pattern).
  • the Bayer pattern includes a filter structure 1317 , and each filter structure 1317 includes filter units 1311 2 by 2, which are green, red, blue and green filter units.
  • the filter unit array 131 includes R (red) filter units 1311 , G (green) filter units 1311 and B (Blue) filter units 1311 , wherein, the G filter units 1311 at least include the white filter areas 1313 .
  • the white filter areas cover at least one of the N photosensitive pixel 111 covered by the G filter units.
  • one photosensitive pixel corresponds to one filter unit.
  • the filter unit array 131 may configure to use Bayer pattern, which includes the filter structure 1317 .
  • Each filter structure 1317 includes R, B and G filter units 1311 .
  • one filter unit 1311 corresponds to N photosensitive pixels, which is different from the traditional structure.
  • each G filter unit 1311 includes a white filter area 1315 which is corresponding to at least one of the N photosensitive pixels 111 .
  • the G filter unit 1311 further includes a green filter area 1315 , wherein the green filter area 1315 corresponds to the other photosensitive pixels of N.
  • the R filter units 1311 only include the red filter area. In other words, one red filter unit 1311 cover 4 photosensitive pixels corresponding to one R filter unit 1311 .
  • the B filter units 1311 only include the blue filter area. In other words, one blue filter unit 1311 cover 4 photosensitive pixels corresponding to one B filter unit 1311 .
  • each filter unit 1311 includes 2*2 photosensitive pixels, in other words, each filter unit 1311 covers 2*2 photosensitive pixels to form a merged pixel.
  • n*m are the positive integer respectively.
  • the resolution of the image is limited. For example, when the pixel value of the photosensitive pixel array 11 is 16M (mega), if the structure of 2*2 photosensitive pixels is used, a 4M-resolution image will be obtained, and if the structure of 4*4 is used, a 1M-resolution image will be obtained.
  • the structure of 2*2 photosensitive pixels is a better arrangement, which can enhance brightness and sharpness of the image under the premise of sacrificing less resolution as possible.
  • a white filter area 1313 covers one photosensitive pixel covered by the G filter unit 1311 , wherein, the G filter unit 1311 further includes a green filter area 1315 and the green filter area 1315 covers the other 3 of 4 photosensitive pixels.
  • a white filter area 1313 covers 2 photosensitive pixels covered by the G filter unit 1311 , wherein, the G filter unit 1311 further includes a green filter area 1315 and the green filter area 1315 covers the other 2 of 4 photosensitive pixels.
  • FIGS. 4 a and 5 a a white filter area 1313 covers one photosensitive pixel covered by the G filter unit 1311 , wherein, the G filter unit 1311 further includes a green filter area 1315 and the green filter area 1315 covers the other 3 of 4 photosensitive pixels.
  • a white filter area 1313 covers 2 photosensitive pixels covered by the G filter unit 1311 , wherein, the G filter unit 1311 further includes a green filter area 1315 and the green filter area 1315 covers the other 2 of 4 photosensitive pixels.
  • a white filter area 1313 covers 3 photosensitive pixels covered by the G filter unit 1311 , wherein, the G filter unit 1311 further includes a green filter area 1315 and the green filter area 1315 covers the other 1 of 4 photosensitive pixels. Or as shown in FIGS. 4 d and 5 d, a white filter area 1313 covers 4 photosensitive pixels covered by the G filter unit 1311 .
  • N photosensitive pixels 111 of a merged pixel can be covered by the non-white filter area 1315 which is the green filter area and the white filter area 1313 together, or N photosensitive pixels 111 of a merged pixel can be only covered by the white filter area 1313 .
  • N photosensitive pixels 111 of a merged pixel in a R filter unit 1313 , can be only covered by the non-white filter area 1313 , which is the red filter area.
  • N photosensitive pixels 111 of a merged pixel in a B filter unit 1313 , can be only covered by the non-white filter area 1313 , which is the blue filter area.
  • the image sensor further includes a control module 17 , which is configured to control the photosensitive pixel array 11 to expose by row.
  • the control module 17 is connected with a row selection logic unit 171 and a column selection logic unit 173 , such that the control module 17 is configured to process the outputs of the photosensitive pixels 111 by row.
  • Exposing and outputting by row can be realized easier in hardware aspect.
  • the image sensor includes a register 19 .
  • reading outputs, buffering and merging the photosensitive units can be realized by adopting the register. Reading outputs, buffering and merging the photosensitive units can be realized easier and processed faster by hardware.
  • the image sensor 10 includes a row selection logic unit 171 and a column selection logic unit 173 , wherein the control module 17 is connected with a row selection logic unit 171 and a column selection logic unit 173 , row selection logic unit 171 and the column selection logic unit 173 connects with each time gate 1115 corresponding to the photosensitive pixel 111 .
  • the control module 17 is configured to control the row selection logic unit 171 and the column selection logic unit 173 to conduct the time gate 1115 of the photosensitive pixel 111 in specific position.
  • the control module 17 is configured to collect outputs of the first and the second row the photosensitive pixels and to store the outputs into the register 19 .
  • the circuit of the image sensor is configured to process the outputs of the four photosensitive pixels in position 1 - 1 , 1 - 2 , 2 - 1 and 2 - 2 and to obtain the pixel value of the merged pixel, wherein the number in the left of the position means the row of the photosensitive pixels and the number in the right means the column.
  • the circuit of the image sensor is configured to process the outputs of the 4 photosensitive pixels in position 1 - 3 , 1 - 4 , 2 - 3 and 2 - 4 to obtain the pixel value of the corresponding merged pixel.
  • the outputs of the third and the fourth row and the fifth and the sixth row of the photosensitive pixels will be processed in the way which is mentioned above, and the processing will not be accomplished until outputs of all photosensitive pixels have been processed.
  • the image sensor 10 further includes ADCs 21 , wherein each ADC 21 is configured to connect to one photosensitive 111 and to convert analog signal outputs of the photosensitive 111 to digital signal outputs.
  • a photosensitive pixel 111 includes a photodiode 1113 .
  • the photodiode 1113 is configured to convert the light to charge, and the charge is proportional to light intensity.
  • the time gate 1115 is configured to conduct the circuit according to the control signals of the row selection logic unit 171 and the column selection logic unit 173 .
  • the source follower 1117 is configured to convert a charge signal generated by exposure of the photodiode 1113 to a voltage signal.
  • the ADC 211 is configured to convert the voltage to a digital signal so that to transmit the digital to following circuit to process.
  • the processing way mentioned above converts the outputs of the photosensitive pixels to the digital signal, so that the digital signal can be processed in the following digital circuit or by the program in a chip.
  • the output information of every photosensitive pixel can be conserved.
  • the information of the 16M pixels which means the image before being merged can be conserved, and then a 4M pixels merged image or other solution images can be obtained by processing on basis of the 16M pixels information.
  • the probability of defective pixels in the generated image is lower.
  • the noise of output of the processing way mentioned above is smaller, and signal to noise ratio is higher.
  • the image sensor 10 includes a micro-lens array 23 which is arranged on the filter 13 , wherein each micro-lens is configured to correspond to a photosensitive pixel 111 .
  • each micro-lens 231 corresponds to a photosensitive pixel 111 , which means to correspond with the size and the position.
  • each filter unit 1311 corresponds to 2*2 photosensitive pixels 111 and 2*2 micro-lenses 191 .
  • the image sensor in the embodiments of the present disclosure embedding the white filter areas into some of the filter units, so that the brightness information of the merged pixels can be obtained in low light, and the noise of the brightness information is less.
  • the pixel value of a merged image generated by the merged pixels includes both color information and less noise brightness information. Both of the brightness and the sharpness of the merged image are better and the noise of the image is less.
  • Embodiments of the present disclosure provide an imaging device 100 .
  • an imaging device 100 including the image sensor 10 includes the above-identified image sensor 10 and an image signal processor 50 connected with the image sensor 10 , wherein the image signal processor 50 is configured to read and to process the outputs of the photosensitive pixel array 11 to obtain the pixel value of the merged pixels and further obtain the merged images.
  • the image sensor 10 may include the control module 17 , the row selection logic unit 171 , the column selection logic unit 173 , the ADC array 21 and the register 19 .
  • the ADC array 21 is configured to convert the outputs of the photosensitive pixel array 11 to digital signals.
  • the digital signals are being stored into the register 19 by row and transmitted to the image signal processor 50 to process.
  • a merged image can be generated by processing the outputs of all the photosensitive pixels.
  • the image signal processor 50 is configured to calculate the pixel value of merged pixels according to the output of the photosensitive pixels of the same merged pixel in order to generate the merged image.
  • the image signal process 50 is further configured to add together outputs of N photosensitive pixels of the same merged pixel and to take the result as the pixel value of the merged pixel.
  • the image signal processor 50 is further configured to add the outputs of the photosensitive pixels corresponding to the white filter areas together to obtain a first pixel value of the merged pixel, and to add the outputs of the photosensitive pixels corresponding to the non-white filter areas together to obtain a second pixel value of the merged pixel.
  • pixel value can be obtained by adding the outputs of the photosensitive pixels covered by the same color filter in every filter unit.
  • the first pixel value of the merged pixel can be obtained by adding the outputs of the 2 photosensitive pixels covered by the white filter area together, which are in the position of upper left corner and bottom right corner of a G filter unit.
  • the second pixel value of the merged pixel can be obtained by adding together the outputs of the 2 photosensitive pixels covered by the green filter area.
  • the pixel value of the merged pixel can be obtained by adding together the 4 photosensitive pixels covered by the blue filter area.
  • the pixel value of the merged pixel can be obtained by adding together the 4 photosensitive pixels covered by the red filter area.
  • the image signal processor 50 is configured to generated a merged image according to the first pixel value of the merged pixel of G filter unit, the second pixel value of the merged pixel of the G filter unit, the pixel value of the merged pixel of B filter unit and the pixel value of the merged pixel of the R filter unit.
  • adding together the outputs of a plurality of photosensitive pixels can make signal to noise ratio of the merged pixel higher.
  • the output of each photosensitive pixel is S
  • the noise of each photosensitive pixel is Ns
  • the merged pixel includes N photosensitive pixels, such that the pixel value of the merged pixel is N*S
  • the noise of the merged pixel is
  • N is a positive integer which is greater than or equal to 1. It is known that, when N is greater than 1, the noise of merged pixel is less than the sum of the noise of photosensitive pixels before being merged. For example, in an embodiment of the present disclosure, when N equals to 4, the noise of the merged pixel equals to Ns/2, and less than the sum of the noise of the 4 photosensitive pixels 4*Ns. In another hand, the output of the merged pixel equals to the sum of outputs of the 4 photosensitive. Such that, the signal to noise ratio of the merged image has been improved and the noise of the merged image has been reduced and the sharpness of the merged image has been improved in the meantime.
  • the imaging device embedding the white filter areas into some of the filter units, so that the brightness information of the merged pixels can be obtained in lower light, and the noise of the brightness information is lower.
  • the pixel value of a merged image generated by the merged pixels includes both color information and lower noise brightness information. Both of the brightness and the sharpness of the merged image are better and the noise of the image is less.
  • signal to noise ratio and brightness of images in low light can be improved and the noise of the image can be reduced through merging pixels by the image signal processor because of the noise of the merged pixel is less than the sum of the noise of the photosensitive pixels before being merged.
  • a mobile terminal applies an imaging device is further provided in the present disclosure.
  • the mobile terminal includes the imaging device according to embodiments of the present disclosure. Such that, the mobile terminal is configured to take photos and to generate a merged image with integrity of the color, higher signal to noise ratio and higher sharpness in low light.
  • the mobile terminal may be a cell phone.
  • the imaging device may be a front camera of a cell phone.
  • the front camera is often used for taking self-portrait, and the requirement of the self-portrait with higher sharpness and normal solution can be satisfied with the mobile terminal of the present disclosure.
  • the mobile terminal 200 includes a CPU 81 and an external memory 83 , which are connected with the imaging device 100 separately, wherein the CPU 81 is configured to control the external memory 83 to store the merged images.
  • the external memory 83 includes a SM (Smart Media) card and a CF (Compact Flash) card.
  • the mobile terminal 200 includes a display device 85 , which is connected with the imaging device 100 , wherein the CPU 81 is configured to control the display device 85 to display the merged images. Such that, the images taken by the mobile terminal 200 can be displayed on the display device for viewing by users.
  • the display device may be a LED monitor.
  • the mobile terminal is configured to take photos and to generate a merged image with complete color, higher signal to noise ratio and higher sharpness in low light.
  • the imaging device is used as the front camera, a higher brightness and sharpness and less noise self-portrait image can be generated.
  • An imaging method of the image sensor is further provided in the present disclosure.
  • the imaging method of the present disclosure includes:
  • the image sensor includes a photosensitive pixel array and a filter arranged on the array.
  • the filter includes a filter unit array.
  • the filter unit array includes a plurality of filter units, wherein each unit covers N photosensitive pixels. Some of the filter units include white filter areas, wherein the whiter filter areas cover at least one of the N photosensitive pixels.
  • a merged pixel can be formed by the N photosensitive pixels covered by the same filter unit, wherein N is positive integer. External light illuminates the photosensitive part of a photosensitive pixel to generate electric signal, which is the output of a photosensitive pixel.
  • the pixel value of a merged image generated by the merged pixels includes both color information and lower noise brightness information. Both of the brightness and the sharpness of the merged image are better and the noise of the image is less.
  • S 2 further includes: adding the outputs of the N photosensitive pixels corresponding to the same merged pixel together and making the result as the pixel value of the merged pixel.
  • the pixel value of the merged pixel includes a first pixel value corresponding to the white filter areas and a second pixel value corresponding to the non-white filter areas.
  • S 2 further includes:
  • S 2 includes:
  • reading outputs, buffering and merging the photosensitive units can be realized by adopting the register. Reading outputs, buffering and merging the photosensitive units can be realized easier in hardware and processed faster.
  • each ADC connects to one photosensitive pixel
  • the imaging method further includes:
  • the image signal process which is a digital signal process (DSP) can process the output of the image sensor directly, and secondly, the information of the image can be conserved better than the circuit processes the analog signal outputs of the image sensor.
  • DSP digital signal process
  • the imaging method of the present disclosure can conserve the information of 16 M pixel, and then obtain merged images with 4M pixel or other resolutions by processing the 16M pixel.
  • the imaging method as the present disclosure embedding the white filter areas into some of the filter units, so that the brightness information of the merged pixels can be obtained in lower light, and the noise of the brightness information is lower.
  • the pixel value of a merged image generated by the merged pixels includes both color information and lower noise brightness information. Both of the brightness and the sharpness of the merged image are better and the noise of the image is less. Further, Since the noise of the merged pixel is less than the sum of noise of the photosensitive pixels before being merged, the signal to noise ratio and brightness of images in low light can be improved and the noise of the image can be reduced through merging pixels by the image signal processor
  • the unexpanded portion of the imaging method and the mobile terminal of the embodiment of the present disclosure can be referred to the corresponding portion of the image sensor or the imaging device of the mentioned above, and will not be expanded in detail here.
  • the computer readable medium comprise but are not limited to: an electronic connection (an electronic device) with one or more wires, a portable computer enclosure (a magnetic device), a random-access memory (RAM), a read only memory (ROM), an erasable programmable read-only memory (EPROM or a flash memory), an optical fiber device and a portable compact disk read-only memory (CDROM).
  • the computer readable medium may even be a paper or other appropriate medium capable of printing programs thereon, this is because, for example, the paper or other appropriate medium may be optically scanned and then edited, decrypted or processed with other appropriate methods when necessary to obtain the programs in an electric manner, and then the programs may be stored in the computer memories.
  • each part of the present disclosure may be realized by the hardware, software, firmware or their combination.
  • a plurality of steps or methods may be realized by the software or firmware stored in the memory and executed by the appropriate instruction execution system.
  • the steps or methods may be realized by one or a combination of the following techniques known in the art: a discrete logic circuit having a logic gate circuit for realizing a logic function of a data signal, an application-specific integrated circuit having an appropriate combination logic gate circuit, a programmable gate array (PGA), a field programmable gate array (FPGA), etc.
  • each function cell of the embodiments of the present disclosure may be integrated in a processing module, or these cells may be separate physical existence, or two or more cells are integrated in a processing module.
  • the integrated module may be realized in a form of hardware or in a form of software function modules. When the integrated module is realized in a form of software function module and is sold or used as a standalone product, the integrated module may be stored in a computer readable storage medium.
  • the storage medium mentioned above may be read-only memories, magnetic disks or CD, etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Power Engineering (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Condensed Matter Physics & Semiconductors (AREA)
  • Computer Hardware Design (AREA)
  • Microelectronics & Electronic Packaging (AREA)
  • General Physics & Mathematics (AREA)
  • Electromagnetism (AREA)
  • Spectroscopy & Molecular Physics (AREA)
  • Transforming Light Signals Into Electric Signals (AREA)
  • Color Television Image Signal Generators (AREA)
  • Solid State Image Pick-Up Elements (AREA)

Abstract

The present disclosure discloses an image sensor, an imaging device, a mobile terminal and an imaging method. The image sensor comprises a photosensitive pixel array and a filer arranged on the photosensitive pixel array. The filter comprises a filer unit array comprised a plurality of filter units, wherein each filter unit covers N photosensitive pixels, and some of the filter units comprise white filter areas. The white filter areas cover at least one of the N photosensitive pixels of the N photosensitive pixels, wherein a merged pixel is formed by the N photosensitive pixels covered by the same filter unit, wherein N is a positive integer.

Description

    CROSS-REFERENCE TO RELATED APPLICATION
  • This application claims priority to and the benefit of Chinese Patent Application No. 201510963465.1, filed with the State Intellectual Property Office of P. R. China on Dec. 18, 2015. The entire contents of the above-identified application are incorporated herein by reference.
  • FIELD
  • The present disclosure relates to imaging technologies, and more particularly to an image sensor, an imaging device, a mobile terminal, and an imaging method of the image sensor.
  • BACKGROUND
  • In the related art, the imaging sensor of the imaging device may have a problem of larger noise and lower resolution in an image generated in low light.
  • SUMMARY
  • The present disclosure aims to solve at least one of the problems existing in the related art to at least some extent.
  • In order to achieve the above purposes, embodiments of a first aspect of the present disclosure provide an image sensor. The image sensor includes: a photosensitive pixel array and a filter arranged on the photosensitive pixel array. The filter comprises a filer unit array having a plurality of filter units, wherein each filter unit covers N photosensitive pixels, some of the filter units at least include white filter areas. The white filter areas cover at least one of the N photosensitive pixels. A merged pixel is formed by the N photosensitive pixels covered by the same filter unit, wherein N is a positive integer.
  • According to the image sensor provided in embodiments of the present disclosure, the amount of light entering into the image sensor can be increased through embedding the white filter areas into some of the filter units, so that a higher signal to noise ratio, higher brightness, higher sharpness and less noise image can be obtained in low light.
  • According to some embodiments of the present disclosure, the filter unit array includes R filter units, G filter units, and B filter units, wherein the G filter units at least include the white filter areas, and the white filter areas cover at least one of the N photosensitive pixels covered by the G filter units.
  • According to some embodiments of the present disclosure, each filter unit includes 2*2 photosensitive pixels, wherein the white filter areas cover one photosensitive pixel covered by the G filter unit, and the G filter unit further includes green filter areas, wherein the green filter areas cover the other three photosensitive pixels. Or the white filter areas cover two photosensitive pixels covered by the G filter unit, and the G filter unit further includes green filter areas, wherein the green filter areas cover the other two photosensitive pixels. Or the white filter areas cover three photosensitive pixels covered by the G filter unit, and the G filter unit further includes green filter areas, wherein the green filter areas cover the other one photosensitive pixel. Or the white filter areas cover all of the four photosensitive pixels covered by the G filter unit.
  • According to some embodiments of the present disclosure, the image sensor further includes a control module, which is configured to control the photosensitive pixel array to be exposed by row.
  • According to some embodiments of the present disclosure, the image sensor further includes a register. The control module is configured to collect outputs of the photosensitive pixels of row k and row k+1 in turn which have been exposed and to store the outputs into the register, wherein k=2n−1, n is a positive integer, and k+1 is less than or equal to the amount of the rows of the photosensitive pixel array.
  • According to some embodiments of the present disclosure, the image sensor further includes an ADC (Analog-to-digital converter) array which includes a plurality of ADCs, wherein each of the ADCs connects to one photosensitive pixel.
  • According to some embodiments of the present disclosure, the image sensor further includes a micro-lens array, each micro-lens corresponds to one photosensitive pixel.
  • In order to achieve the above purposes, embodiments of a second aspect of the present disclosure provide an imaging device including the image sensor. The imaging device including the image sensor according to embodiments of the present disclosure includes the above-identified image sensor and an image signal processor, wherein the image signal processor is configured to read and to process the outputs of the photosensitive pixel array of the image sensor in order to obtain the pixel value of the merged pixels and further to obtain merged images.
  • With the imaging device provided in embodiments of the present disclosure, the amount of light entering into the image sensor can be increased by the provided image sensor mentioned above, so that a higher signal to noise ratio, higher brightness, higher sharpness and less noise image can be obtained in low light.
  • According to some embodiments of the present disclosure, when the filter units only include white filter areas or non-white filter areas, the image signal processor is further configured to add the outputs of the N photosensitive pixels corresponding to the same merged pixel together and to make the result as the pixel value of the merged pixel.
  • According to some embodiments of the present disclosure, when the filter units include the white filter areas and the non-white filter areas, the image signal processor is further configured to add the outputs of the photosensitive pixels corresponding to the white filter areas together to obtain a first pixel value of the merged pixel, and to add the outputs of the photosensitive pixels corresponding to the non-white filter areas together to obtain a second pixel value of the merged pixel.
  • As the noise of merged pixels is less than the sum of noise of all the photosensitive pixels before being merged, it could improve the signal to noise ratio, brightness, and sharpness, and reduce noise of an image in low light via the merged pixels.
  • In order to achieve the above purposes, embodiments of a third aspect of the present disclosure provide a mobile terminal. The mobile terminal, according to embodiments of the present disclosure, includes the above-identified imaging device for a mobile terminal.
  • With the mobile terminal provided in embodiments of the present disclosure, the amount of light entering into the imaging device can be increased by the provided imaging device mentioned above, so that a higher signal to noise ratio, higher brightness, higher sharpness and less noise image can be obtained in low light, and further, as the noise of merged pixels is less than the sum of noise of all the photosensitive pixels, it could improve the signal to noise ratio, brightness, and sharpness, and reduce noise of an image in low light via the merged pixels.
  • According to some embodiments of the present disclosure, the mobile terminal includes a cell phone.
  • According to some embodiments of the present disclosure, the imaging device includes front cameras.
  • According to some embodiments of the present disclosure, the mobile terminal further comprises a CPU and an external memory, which are connected with the imaging device separately, wherein the CPU is configured to control the external memory to store the merged images.
  • According to some embodiments of the present disclosure, the mobile terminal further includes a display device, which is connected with the imaging device, wherein the CPU is configured to control the display device to display the merged images.
  • In order to achieve the above purposes, embodiments of a fourth aspect of the present disclosure provide an imaging method of the image sensor. The imaging method, according to embodiments of the present disclosure, includes: reading outputs of the photosensitive unit array; calculating pixel value of the merged pixel according to the outputs of the photosensitive pixels of the same merged pixel to obtain a merged image.
  • According to the imaging method provided in embodiments of the present disclosure, the amount of light entering into the imaging device can be increased by the provided image sensor mentioned above, so that a higher signal to noise ratio, higher brightness, higher sharpness and less noise image can be obtained in low light.
  • According to some embodiments of the present disclosure, each filter unit includes 2*2 photosensitive pixels. The step of calculating pixel value of the merged pixel according to the outputs of the photosensitive pixels of the same merged pixel further comprises: collecting and storing the outputs of the photosensitive pixels of row k and row k+1 into the register, wherein k=2n−1, n is a positive integer, and k+1 is less than or equal to the amount of the rows of the photosensitive pixel array; and extracting the outputs of the photosensitive pixels of row k and row k+1 from the register to obtain the pixel value of the merged pixels.
  • According to some embodiments of the present disclosure, when the filter units only include the white filter areas or the non-white filter areas, the step of calculating pixel value of the merged pixel according to the outputs of the photosensitive pixels of the same merged pixel further comprises: adding the outputs of the N photosensitive pixels corresponding to the same merged pixel together and making the result as the pixel value of the merged pixel.
  • According to some embodiments of the present disclosure, when the filter units include the white filter areas and the non-white filter areas, the pixel value of the merged pixel includes a first pixel value corresponding to the white filter areas and a second pixel value corresponding to the non-white filter areas. The step of calculating pixel value of the merged pixel according to the outputs of the photosensitive pixels of the same merged pixel further comprises: adding the outputs of the photosensitive pixels corresponding to the white filter areas together to obtain the first pixel value of the merged pixel; and adding the outputs of the photosensitive pixels corresponding to the non-white filter areas together to obtain the second pixel value of the merged pixel.
  • As the noise of merged pixels is less than the sum of noise of all the photosensitive pixels, it could improve the signal to noise ratio, brightness, and sharpness, and reduce noise of an image in low light via the merged pixels.
  • According to some embodiments of the present disclosure, each ADC connects to one photosensitive pixel, the imaging method further includes: converting analog signal outputs of the photosensitive pixel into digital signal outputs; and calculating the pixel value of the merged pixel according to the digital signal outputs of the photosensitive pixels of the same merged pixel.
  • Additional aspects and advantages of embodiments of the present disclosure will be given in part in the following descriptions, become apparent in part from the following descriptions, or be learned from the practice of the embodiments of the present disclosure.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • These and other aspects and advantages of embodiments of the present disclosure will become apparent and more readily appreciated from the following descriptions made with reference to the accompanying drawings, in which:
  • FIG. 1 is a schematic side view of an exemplary image sensor according to an embodiment of the present disclosure;
  • FIG. 2a-2d are schematic views of an exemplary filter units of the image sensor according to an embodiment of the present disclosure;
  • FIG. 3 is a schematic view of an exemplary Bayer Color Filter Array according to an embodiment of the present disclosure;
  • FIG. 4a-4d are schematic views of an exemplary filter unit array of the image sensor according to an embodiment of the present disclosure;
  • FIG. 5a-5d are schematic views of structure of an exemplary image sensor according to an embodiment of the present disclosure;
  • FIG. 6 is a schematic block diagram of an exemplary image sensor according to an embodiment of the present disclosure;
  • FIG. 7 is a schematic circuit diagram of an exemplary photosensitive pixel of the image sensor according to an embodiment of the present disclosure;
  • FIG. 8 is a schematic block diagram of an exemplary image sensor according to an embodiment of the present disclosure;
  • FIG. 9 is a schematic view of structure of an exemplary image sensor according to an embodiment of the present disclosure;
  • FIG. 10 is a schematic block diagram of an exemplary imaging device according to an embodiment of the present disclosure;
  • FIG. 11 is a schematic block diagram of an exemplary mobile terminal according to an embodiment of the present disclosure;
  • FIG. 12 is a schematic block diagram of an exemplary mobile terminal according to an embodiment of the present disclosure;
  • FIG. 13 is a schematic flow chart of an exemplary imaging method according to an embodiment of the present disclosure;
  • FIG. 14 is a schematic flow chart of an exemplary reading step of the imaging method according to an embodiment of the present disclosure; and
  • FIG. 15 is a schematic flow chart of an exemplary imaging method according to an embodiment of the present disclosure.
  • DETAILED DESCRIPTION
  • Reference will be made in detail to embodiments of the present disclosure. Embodiments of the present disclosure will be shown in drawings, in which the same or similar elements and the elements having same or similar functions are denoted by like reference numerals throughout the descriptions. The embodiments described herein according to drawings are explanatory and illustrative, not construed to limit the present disclosure.
  • Various embodiments and examples are provided in the following description to implement different structures of the present disclosure. In order to simplify the present disclosure, certain elements and settings will be described. However, these elements and settings are only by way of example and are not intended to limit the present disclosure. In addition, reference numerals may be repeated in different examples in the present disclosure. This repeating is for the purpose of simplification and clarity and does not refer to relations between different embodiments and/or settings. Furthermore, examples of different processes and materials are provided in the present disclosure. However, it would be appreciated by those skilled in the art that other processes and/or materials may be also applied. Moreover, a structure in which a first feature is “on” a second feature may include an embodiment in which the first feature directly contacts the second feature, and may also include an embodiment in which an additional feature is formed between the first feature and the second feature so that the first feature does not directly contact the second feature.
  • In the description of the present disclosure, it should be noted that, unless specified or limited otherwise, the terms “mounted,” “connected,” and “coupled” and variations thereof are used broadly and encompass such as mechanical or electrical mountings, connections and couplings, also can be inner mountings, connections and couplings of two components, and further can be direct and indirect mountings, connections, and couplings, which can be understood by those skilled in the art according to the detail embodiment of the present disclosure.
  • An image sensor, an imaging device, a mobile terminal and an imaging method, according to embodiments of the present disclosure, may be described below with reference to FIGS. 1-15.
  • In some embodiments, as shown in FIG. 1 and FIGS. 2a -2 d, the image sensor 10 includes a photosensitive pixel array 11 and a filter 13.
  • Wherein, the filter 13 is arranged on the photosensitive pixel array 11 and includes a filter unit array 131. The filter unit array 131 includes a plurality of filter units 1311. Each filter unit 1311 covers N photosensitive pixels 111. Some of the filter units 1311 at least include a white filter area 1313. The white filter area 1313 covers at least one of the N photosensitive pixel 111. The N photosensitive pixels 111 covered by the same filter unit 1311 forms a merged pixel, and N is a positive integer. External light illuminates the photosensitive part 1111 of a photosensitive pixel 111 to generate electric signal, which is the output of a photosensitive pixel 111.
  • It should be noted that, the white filter areas 1313 is configured to allow the light pass through instead of being filtered. Such that, the white filter areas 1313 may refers to either the areas covered by the transparent filters or the areas with no filters, the areas with no filters is the hollow of the filter 13.
  • It should be further noted that, the other filter units 1311 except the filter units 1311 with the white filter area 1313 include non-white filter areas, such as green filter area, red filter area or blue filter area. When the number of the photosensitive pixels covered by the white filter area 1313 is less than N, the filter units 1311 with white filter area 1313 further include the non-white filter area. In other words, some filter units 1311 consist of the white filter area and the non-white filter area at the same time. The two areas cover N photosensitive pixels together. Wherein, the non-white filter area is configured to obtain color information of the merged pixels, and the white filter area is configured to obtain white light information. In other words, the white filter area may allow the light pass through. Such that, the brightness is higher which is outputted by the photosensitive with the better effect of light transmission of the white filter area. White filter area is further configured to obtain the brightness information of the merged pixels in low light, and the noise of the brightness information is less.
  • With the image sensor provided in embodiments of the present disclosure, the amount of light entering into the image sensor can be increased through embedding the white filter areas into some of the filter units, so that a higher signal to noise ratio, higher brightness, higher sharpness and less noise image can be obtained in low light.
  • In some embodiments, the filter unit array shown in FIG. 3 is arranged according to the Bayer Color Filter Array (Bayer Pattern). Wherein, the Bayer pattern includes a filter structure 1317, and each filter structure 1317 includes filter units 1311 2 by 2, which are green, red, blue and green filter units.
  • With the Bayer pattern, algorithm for the traditional Bayer pattern can be used to process the image signal instead of changing the structure of hardware.
  • In some embodiments, as shown in FIGS. 4a -4 d, the filter unit array 131 includes R (red) filter units 1311, G (green) filter units 1311 and B (Blue) filter units 1311, wherein, the G filter units 1311 at least include the white filter areas 1313. The white filter areas cover at least one of the N photosensitive pixel 111 covered by the G filter units.
  • Specifically, in traditional filter unit array structure, one photosensitive pixel corresponds to one filter unit. In some embodiments of the present disclosure, the filter unit array 131 may configure to use Bayer pattern, which includes the filter structure 1317. Each filter structure 1317 includes R, B and G filter units 1311. In this embodiment, one filter unit 1311 corresponds to N photosensitive pixels, which is different from the traditional structure.
  • Wherein, each G filter unit 1311 includes a white filter area 1315 which is corresponding to at least one of the N photosensitive pixels 111. When the number of photosensitive pixels 111 covered by the white filter area 1315 is less than N, the G filter unit 1311 further includes a green filter area 1315, wherein the green filter area 1315 corresponds to the other photosensitive pixels of N. Furthermore, the R filter units 1311 only include the red filter area. In other words, one red filter unit 1311 cover 4 photosensitive pixels corresponding to one R filter unit 1311. Similarly, the B filter units 1311 only include the blue filter area. In other words, one blue filter unit 1311 cover 4 photosensitive pixels corresponding to one B filter unit 1311.
  • As shown in FIGS. 4a-4d and FIGS. 5a -5 d, each filter unit 1311 includes 2*2 photosensitive pixels, in other words, each filter unit 1311 covers 2*2 photosensitive pixels to form a merged pixel.
  • In some embodiments, concerning the structure of the merged pixel, in addition to the structure of 2*2, there may be the structure of 3*3, 4*4, and any n*m (n and m are the positive integer respectively). It is to be understood that, since the number of the photosensitive pixels 111 arranged on the photosensitive pixel array 11 is limited, if each merged pixel includes too many photosensitive pixels, the resolution of the image is limited. For example, when the pixel value of the photosensitive pixel array 11 is 16M (mega), if the structure of 2*2 photosensitive pixels is used, a 4M-resolution image will be obtained, and if the structure of 4*4 is used, a 1M-resolution image will be obtained. Therefore, the structure of 2*2 photosensitive pixels is a better arrangement, which can enhance brightness and sharpness of the image under the premise of sacrificing less resolution as possible. At the same time, it is easy to realize to merge the photosensitive pixels and to read out the outputs of the photosensitive pixels in hardware aspect with the structure of 2*2 photosensitive pixels.
  • As shown in FIGS. 4a and 5a , a white filter area 1313 covers one photosensitive pixel covered by the G filter unit 1311, wherein, the G filter unit 1311 further includes a green filter area 1315 and the green filter area 1315 covers the other 3 of 4 photosensitive pixels. Or as shown in FIGS. 4b and 5b , a white filter area 1313 covers 2 photosensitive pixels covered by the G filter unit 1311, wherein, the G filter unit 1311 further includes a green filter area 1315 and the green filter area 1315 covers the other 2 of 4 photosensitive pixels. Or as shown in FIGS. 4c and 5c , a white filter area 1313 covers 3 photosensitive pixels covered by the G filter unit 1311, wherein, the G filter unit 1311 further includes a green filter area 1315 and the green filter area 1315 covers the other 1 of 4 photosensitive pixels. Or as shown in FIGS. 4d and 5 d, a white filter area 1313 covers 4 photosensitive pixels covered by the G filter unit 1311.
  • Such that, in a G filter unit 1313, N photosensitive pixels 111 of a merged pixel can be covered by the non-white filter area 1315 which is the green filter area and the white filter area 1313 together, or N photosensitive pixels 111 of a merged pixel can be only covered by the white filter area 1313. In some embodiments, in a R filter unit 1313, N photosensitive pixels 111 of a merged pixel can be only covered by the non-white filter area 1313, which is the red filter area. In some embodiments, in a B filter unit 1313, N photosensitive pixels 111 of a merged pixel can be only covered by the non-white filter area 1313, which is the blue filter area.
  • In some embodiment, as shown in FIG. 6, the image sensor further includes a control module 17, which is configured to control the photosensitive pixel array 11 to expose by row. In detail, the control module 17 is connected with a row selection logic unit 171 and a column selection logic unit 173, such that the control module 17 is configured to process the outputs of the photosensitive pixels 111 by row.
  • Exposing and outputting by row can be realized easier in hardware aspect.
  • Further, as shown in FIG. 6, the image sensor includes a register 19, The control module 17 is configured to collect outputs of the photosensitive pixels 111 of row k and row k+1 in turn which have been exposed and to store the outputs into the register 19, wherein k=2n−1, n is a positive integer, and k+1 is less than or equal to the amount of the rows of the photosensitive pixel array 11.
  • Such that, reading outputs, buffering and merging the photosensitive units can be realized by adopting the register. Reading outputs, buffering and merging the photosensitive units can be realized easier and processed faster by hardware.
  • In detail as shown in FIGS. 6 and 7, the image sensor 10 includes a row selection logic unit 171 and a column selection logic unit 173, wherein the control module 17 is connected with a row selection logic unit 171 and a column selection logic unit 173, row selection logic unit 171 and the column selection logic unit 173 connects with each time gate 1115 corresponding to the photosensitive pixel 111. The control module 17 is configured to control the row selection logic unit 171 and the column selection logic unit 173 to conduct the time gate 1115 of the photosensitive pixel 111 in specific position.
  • The control module 17 is configured to collect outputs of the first and the second row the photosensitive pixels and to store the outputs into the register 19. The circuit of the image sensor is configured to process the outputs of the four photosensitive pixels in position 1-1, 1-2, 2-1 and 2-2 and to obtain the pixel value of the merged pixel, wherein the number in the left of the position means the row of the photosensitive pixels and the number in the right means the column.
  • Then, the circuit of the image sensor is configured to process the outputs of the 4 photosensitive pixels in position 1-3, 1-4, 2-3 and 2-4 to obtain the pixel value of the corresponding merged pixel.
  • And also, the outputs of the third and the fourth row and the fifth and the sixth row of the photosensitive pixels will be processed in the way which is mentioned above, and the processing will not be accomplished until outputs of all photosensitive pixels have been processed.
  • In some embodiments, as shown in FIGS. 7 and 8, the image sensor 10 further includes ADCs 21, wherein each ADC 21 is configured to connect to one photosensitive 111 and to convert analog signal outputs of the photosensitive 111 to digital signal outputs.
  • In some embodiments, as shown in FIG. 7, a photosensitive pixel 111 includes a photodiode 1113. The photodiode 1113 is configured to convert the light to charge, and the charge is proportional to light intensity. The time gate 1115 is configured to conduct the circuit according to the control signals of the row selection logic unit 171 and the column selection logic unit 173. When the circuit is conducted, the source follower 1117 is configured to convert a charge signal generated by exposure of the photodiode 1113 to a voltage signal. The ADC 211 is configured to convert the voltage to a digital signal so that to transmit the digital to following circuit to process.
  • The processing way mentioned above converts the outputs of the photosensitive pixels to the digital signal, so that the digital signal can be processed in the following digital circuit or by the program in a chip. The output information of every photosensitive pixel can be conserved. In one embodiment, for an image sensor with 16M pixels, the imaging method of the present disclosure, the information of the 16M pixels which means the image before being merged can be conserved, and then a 4M pixels merged image or other solution images can be obtained by processing on basis of the 16M pixels information. The probability of defective pixels in the generated image is lower. Furthermore, the noise of output of the processing way mentioned above is smaller, and signal to noise ratio is higher.
  • In some embodiment, as shown in FIG. 9, the image sensor 10 includes a micro-lens array 23 which is arranged on the filter 13, wherein each micro-lens is configured to correspond to a photosensitive pixel 111.
  • In detail, each micro-lens 231 corresponds to a photosensitive pixel 111, which means to correspond with the size and the position. In some embodiments, each filter unit 1311 corresponds to 2*2 photosensitive pixels 111 and 2*2 micro-lenses 191. With the development of technology, in order to obtain a higher resolution image, the number of the photosensitive pixels 111 becomes more and more, and the density of arrangement of the photosensitive pixels is higher. The photosensitive pixel 111 is getting smaller and smaller and its light reception capability will be affected. And the area of photosensitive part 1111 of the photosensitive pixel 111 is limited. The micro-lens 191 can gather light together on the photosensitive part 1111, so that the light reception capability of the photosensitive pixel 111 and the quality of images can be promoted.
  • In summary, according to the image sensor in the embodiments of the present disclosure, embedding the white filter areas into some of the filter units, so that the brightness information of the merged pixels can be obtained in low light, and the noise of the brightness information is less. The pixel value of a merged image generated by the merged pixels includes both color information and less noise brightness information. Both of the brightness and the sharpness of the merged image are better and the noise of the image is less.
  • Embodiments of the present disclosure provide an imaging device 100.
  • In some embodiments, as shown in FIG. 10, an imaging device 100 including the image sensor 10 according to embodiments of the present disclosure includes the above-identified image sensor 10 and an image signal processor 50 connected with the image sensor 10, wherein the image signal processor 50 is configured to read and to process the outputs of the photosensitive pixel array 11 to obtain the pixel value of the merged pixels and further obtain the merged images.
  • In detail, the image sensor 10 may include the control module 17, the row selection logic unit 171, the column selection logic unit 173, the ADC array 21 and the register 19. The ADC array 21 is configured to convert the outputs of the photosensitive pixel array 11 to digital signals. The digital signals are being stored into the register 19 by row and transmitted to the image signal processor 50 to process. A merged image can be generated by processing the outputs of all the photosensitive pixels.
  • Such that, the image signal processor 50 is configured to calculate the pixel value of merged pixels according to the output of the photosensitive pixels of the same merged pixel in order to generate the merged image.
  • In detail, in some embodiments, when the filter units only include the white filter area or the non-white filter area, the image signal process 50 is further configured to add together outputs of N photosensitive pixels of the same merged pixel and to take the result as the pixel value of the merged pixel.
  • And also, when the filter units include both of the white filter areas and the non-white filter areas, the image signal processor 50 is further configured to add the outputs of the photosensitive pixels corresponding to the white filter areas together to obtain a first pixel value of the merged pixel, and to add the outputs of the photosensitive pixels corresponding to the non-white filter areas together to obtain a second pixel value of the merged pixel.
  • In other words, pixel value can be obtained by adding the outputs of the photosensitive pixels covered by the same color filter in every filter unit. As shown in FIG. 4b , in each filter structure, the first pixel value of the merged pixel can be obtained by adding the outputs of the 2 photosensitive pixels covered by the white filter area together, which are in the position of upper left corner and bottom right corner of a G filter unit. And the second pixel value of the merged pixel can be obtained by adding together the outputs of the 2 photosensitive pixels covered by the green filter area. In the B filter unit on the bottom left corner, the pixel value of the merged pixel can be obtained by adding together the 4 photosensitive pixels covered by the blue filter area. In the R filter unit on the upper right corner, the pixel value of the merged pixel can be obtained by adding together the 4 photosensitive pixels covered by the red filter area.
  • The image signal processor 50 is configured to generated a merged image according to the first pixel value of the merged pixel of G filter unit, the second pixel value of the merged pixel of the G filter unit, the pixel value of the merged pixel of B filter unit and the pixel value of the merged pixel of the R filter unit. Such that, adding together the outputs of a plurality of photosensitive pixels can make signal to noise ratio of the merged pixel higher. For example, in an embodiment of the present disclosure, assuming that the output of each photosensitive pixel is S, the noise of each photosensitive pixel is Ns, the merged pixel includes N photosensitive pixels, such that the pixel value of the merged pixel is N*S, and the noise of the merged pixel is
  • N * Ns 2 N ,
  • wherein, N is a positive integer which is greater than or equal to 1. It is known that, when N is greater than 1, the noise of merged pixel is less than the sum of the noise of photosensitive pixels before being merged. For example, in an embodiment of the present disclosure, when N equals to 4, the noise of the merged pixel equals to Ns/2, and less than the sum of the noise of the 4 photosensitive pixels 4*Ns. In another hand, the output of the merged pixel equals to the sum of outputs of the 4 photosensitive. Such that, the signal to noise ratio of the merged image has been improved and the noise of the merged image has been reduced and the sharpness of the merged image has been improved in the meantime. In summary, the imaging device according to the embodiments of the present disclosure, embedding the white filter areas into some of the filter units, so that the brightness information of the merged pixels can be obtained in lower light, and the noise of the brightness information is lower. The pixel value of a merged image generated by the merged pixels includes both color information and lower noise brightness information. Both of the brightness and the sharpness of the merged image are better and the noise of the image is less. In the meanwhile, signal to noise ratio and brightness of images in low light can be improved and the noise of the image can be reduced through merging pixels by the image signal processor because of the noise of the merged pixel is less than the sum of the noise of the photosensitive pixels before being merged.
  • A mobile terminal applies an imaging device is further provided in the present disclosure.
  • In an embodiment of the present disclosure, the mobile terminal includes the imaging device according to embodiments of the present disclosure. Such that, the mobile terminal is configured to take photos and to generate a merged image with integrity of the color, higher signal to noise ratio and higher sharpness in low light.
  • The mobile terminal may be a cell phone.
  • In an embodiment of the present disclosure, the imaging device may be a front camera of a cell phone. The front camera is often used for taking self-portrait, and the requirement of the self-portrait with higher sharpness and normal solution can be satisfied with the mobile terminal of the present disclosure.
  • Further, as shown in FIG. 11, the mobile terminal 200 includes a CPU 81 and an external memory 83, which are connected with the imaging device 100 separately, wherein the CPU 81 is configured to control the external memory 83 to store the merged images.
  • Such that, the merged images can be stored for later viewing, using or transferring. The external memory 83 includes a SM (Smart Media) card and a CF (Compact Flash) card.
  • Further, as shown in FIG. 12, the mobile terminal 200 includes a display device 85, which is connected with the imaging device 100, wherein the CPU 81 is configured to control the display device 85 to display the merged images. Such that, the images taken by the mobile terminal 200 can be displayed on the display device for viewing by users. The display device may be a LED monitor.
  • In summary, the mobile of the present disclosure, the mobile terminal is configured to take photos and to generate a merged image with complete color, higher signal to noise ratio and higher sharpness in low light. Particularly, when the imaging device is used as the front camera, a higher brightness and sharpness and less noise self-portrait image can be generated.
  • An imaging method of the image sensor is further provided in the present disclosure.
  • As shown in FIG. 13, the imaging method of the present disclosure includes:
  • S1: Reading outputs of the photosensitive pixel array of the image sensor.
  • Wherein, the image sensor includes a photosensitive pixel array and a filter arranged on the array. The filter includes a filter unit array. The filter unit array includes a plurality of filter units, wherein each unit covers N photosensitive pixels. Some of the filter units include white filter areas, wherein the whiter filter areas cover at least one of the N photosensitive pixels. A merged pixel can be formed by the N photosensitive pixels covered by the same filter unit, wherein N is positive integer. External light illuminates the photosensitive part of a photosensitive pixel to generate electric signal, which is the output of a photosensitive pixel.
  • Such that, embedding the white filter areas into some of the filter units, so that the brightness information of the merged pixels can be obtained in lower light, and the noise of the brightness information is less. The pixel value of a merged image generated by the merged pixels includes both color information and lower noise brightness information. Both of the brightness and the sharpness of the merged image are better and the noise of the image is less.
  • S2: Calculating pixel value of the merged pixel according to the outputs of the photosensitive pixels of the same merged pixel to obtain a merged image.
  • In detail, in some embodiments, when the filter units only include white filter area or the non-white filter area, S2 further includes: adding the outputs of the N photosensitive pixels corresponding to the same merged pixel together and making the result as the pixel value of the merged pixel.
  • Further, in some embodiments of the present disclosure, when the filter units include the white filter areas and the non-white filter areas, the pixel value of the merged pixel includes a first pixel value corresponding to the white filter areas and a second pixel value corresponding to the non-white filter areas. S2 further includes:
  • Adding the outputs of the photosensitive pixels corresponding to the white filter areas together to obtain the first pixel value of the merged pixel; and adding the outputs of the photosensitive pixels corresponding to the non-white filter areas together to obtain the second pixel value of the merged pixel.
  • With the sum of noise of merged pixels is less than that of photosensitive pixels before being merged, an image with higher signal to noise ratio, brightness, and sharpness, and less noise can be obtained in low light via the merged pixels.
  • Further, in an embodiment of the present disclosure, as shown in FIG. 14, S2 includes:
  • S21: Collecting the outputs of the photosensitive pixels of row k and row k+1, and storing them into the register, wherein k=2n−1, n is a positive integer, and k+1 is less than or equal to the amount of the rows of the photosensitive pixel array; and
  • S22: Extracting the outputs of the photosensitive pixels of row k and row k+1 from the register to obtain the pixel value of the merged pixels.
  • Such that, reading outputs, buffering and merging the photosensitive units can be realized by adopting the register. Reading outputs, buffering and merging the photosensitive units can be realized easier in hardware and processed faster.
  • Further, in an embodiment of the present disclosure, as shown in FIG. 15, each ADC connects to one photosensitive pixel, the imaging method further includes:
  • S31: Converting analog signal outputs of the photosensitive pixel into digital signal outputs; and
  • S32: Calculating the pixel value of the merged pixel according to the digital signal outputs of the photosensitive pixels of the same merged pixel.
  • Such that, firstly, the image signal process which is a digital signal process (DSP) can process the output of the image sensor directly, and secondly, the information of the image can be conserved better than the circuit processes the analog signal outputs of the image sensor. For example, in an embodiment of the present disclosure, for an image sensor with 16M pixel value, the imaging method of the present disclosure can conserve the information of 16 M pixel, and then obtain merged images with 4M pixel or other resolutions by processing the 16M pixel.
  • In summary, the imaging method as the present disclosure, embedding the white filter areas into some of the filter units, so that the brightness information of the merged pixels can be obtained in lower light, and the noise of the brightness information is lower. The pixel value of a merged image generated by the merged pixels includes both color information and lower noise brightness information. Both of the brightness and the sharpness of the merged image are better and the noise of the image is less. Further, Since the noise of the merged pixel is less than the sum of noise of the photosensitive pixels before being merged, the signal to noise ratio and brightness of images in low light can be improved and the noise of the image can be reduced through merging pixels by the image signal processor
  • The unexpanded portion of the imaging method and the mobile terminal of the embodiment of the present disclosure can be referred to the corresponding portion of the image sensor or the imaging device of the mentioned above, and will not be expanded in detail here.
  • It should be noted that the relational terms herein, such as “first” and “second”, are used only for differentiating one entity or operation, from another entity or operation, which, however do not necessarily require or imply that there should be any real relationship or sequence. Moreover, the terms “comprise”, “include” or any other variations thereof are meant to cover non-exclusive including, so that the process, method, article or device comprising a series of elements do not only comprise those elements, but also comprise other elements that are not explicitly listed or also comprise the inherent elements of the process, method, article or device. In the case that there are no more restrictions, an element qualified by the statement “comprises a . . . ” does not exclude the presence of additional identical elements in the process, method, article or device that comprises the said element.
  • The logic and/or steps described in other manners herein or shown in the flow chart, for example, a particular sequence table of executable instructions for realizing the logical function, may be specifically achieved in any computer readable medium to be used by the instruction execution system, device or equipment (such as the system based on computers, the system comprising processors or other systems capable of obtaining the instruction from the instruction execution system, device and equipment and executing the instruction), or to be used in combination with the instruction execution system, device and equipment. As to the specification, “the computer readable medium” may be any device adaptive for including, storing, communicating, propagating or transferring programs to be used by or in combination with the instruction execution system, device or equipment. More specific examples of the computer readable medium comprise but are not limited to: an electronic connection (an electronic device) with one or more wires, a portable computer enclosure (a magnetic device), a random-access memory (RAM), a read only memory (ROM), an erasable programmable read-only memory (EPROM or a flash memory), an optical fiber device and a portable compact disk read-only memory (CDROM). In addition, the computer readable medium may even be a paper or other appropriate medium capable of printing programs thereon, this is because, for example, the paper or other appropriate medium may be optically scanned and then edited, decrypted or processed with other appropriate methods when necessary to obtain the programs in an electric manner, and then the programs may be stored in the computer memories.
  • It should be understood that each part of the present disclosure may be realized by the hardware, software, firmware or their combination. In the above embodiments, a plurality of steps or methods may be realized by the software or firmware stored in the memory and executed by the appropriate instruction execution system. For example, if it is realized by the hardware, likewise in another embodiment, the steps or methods may be realized by one or a combination of the following techniques known in the art: a discrete logic circuit having a logic gate circuit for realizing a logic function of a data signal, an application-specific integrated circuit having an appropriate combination logic gate circuit, a programmable gate array (PGA), a field programmable gate array (FPGA), etc.
  • Those skilled in the art shall understand that all or parts of the steps in the above exemplifying method of the present disclosure may be achieved by commanding the related hardware with programs. The programs may be stored in a computer readable storage medium, and the programs comprise one or a combination of the steps in the method embodiments of the present disclosure when run on a computer.
  • In addition, each function cell of the embodiments of the present disclosure may be integrated in a processing module, or these cells may be separate physical existence, or two or more cells are integrated in a processing module. The integrated module may be realized in a form of hardware or in a form of software function modules. When the integrated module is realized in a form of software function module and is sold or used as a standalone product, the integrated module may be stored in a computer readable storage medium.
  • The storage medium mentioned above may be read-only memories, magnetic disks or CD, etc.
  • Reference throughout this specification to “an embodiment,” “some embodiments,” “one embodiment”, “another example,” “an example,” “a specific example,” or “some examples,” means that a particular feature, structure, material, or characteristic described in connection with the embodiment or example is included in at least one embodiment or example of the present disclosure. Thus, the appearances of the phrases such as “in some embodiments,” “in one embodiment”, “in an embodiment”, “in another example,” “in an example,” “in a specific example,” or “in some examples,” in various places throughout this specification are not necessarily referring to the same embodiment or example of the present disclosure. Furthermore, the particular features, structures, materials, or characteristics may be combined in any suitable manner in one or more embodiments or examples.
  • Although explanatory embodiments have been shown and described, it would be appreciated by those skilled in the art that the above embodiments cannot be construed to limit the present disclosure, and changes, alternatives, and modifications can be made in the embodiments without departing from scope of the present disclosure.

Claims (20)

1. An image sensor, comprising:
a photosensitive pixel array; and
a filter arranged on the photosensitive array, wherein the filter comprises a filter unit array comprised a plurality of filter units; wherein each filter unit covers N photosensitive pixels, and some of the filter units comprise white filter areas, wherein the white filter areas cover at least one of the N photosensitive pixels, wherein a merged pixel is formed by the N photosensitive pixels covered by the same filter unit, wherein N is a positive integer.
2. The image sensor of claim 1, wherein the filter unit array comprises R filter units, G filter units and B filter units, wherein the G filter units comprise white filter areas, and a white filter area covers at least one of the N photosensitive pixels covered by a G filter unit.
3. The image sensor of claim 2, wherein each filter unit comprises 2*2 photosensitive pixels, wherein each G filter unit comprises 2*2 photosensitive pixels, wherein the G filter unit comprises a white filter area and a green filter area; wherein the white filter area covers one of the four photosensitive pixels of the G filter unit and the green filter area covers the other three; or
wherein the white filter area covers two of the four photosensitive pixels and the green filter area covers the other two; or
wherein the white filter area covers three of the four photosensitive pixels and the green filter area covers the other one; or
wherein the white filter area covers all of the four photosensitive pixels.
4. The image sensor of claim 1, wherein the image sensor further comprises a control module, and wherein the control module is configured to control the photosensitive pixel array to be exposed by row.
5. The image sensor of claim 4, wherein the image sensor further comprises a register; wherein the control module is configured to collect outputs of the photosensitive pixels of row k and row k+1 in turn which have been exposed and to store the outputs into the register, wherein k=2n−1, n is a positive integer, and k+1 is less than or equal to the amount of the rows of the photosensitive pixel array.
6. The image sensor of claim 1, wherein the image sensor further comprises an ADC array which comprises a plurality of ADCs; and
wherein each of the ADCs connects to one photosensitive pixel.
7. The image sensor of claim 1, wherein the image sensor further comprises a micro-lens array; wherein the micro-lens array comprises a plurality of micro-lens and wherein each micro-lens corresponds to one photosensitive pixel.
8. An imaging device, comprising:
an image sensor, wherein the image sensor comprises a photosensitive pixel array and a filter arranged on the photosensitive array, wherein the filter comprises a filter unit array comprised a plurality of filter units; wherein each filter unit covers N photosensitive pixels, and some of the filter units comprise white filter areas, wherein the white filter areas cover at least one of the N photosensitive pixels, wherein a merged pixel is formed by the N photosensitive pixels covered by the same filter unit, wherein N is a positive integer; and
an image signal processor, wherein the image signal processor is configured to read and to process outputs of the photosensitive pixel array of the image sensor to obtain pixel values of merged pixels and further to obtain merged images.
9. The imaging device of claim 8, wherein when the filter units only comprise white filter areas or only comprise non-white filter areas, the image signal processor is further configured to add outputs of the N photosensitive pixels corresponding to the same merged pixel together and to make the result as a pixel value of the merged pixel.
10. The imaging device of claim 8, wherein the filter units comprise white filter areas and non-white filter areas, the image signal processor is further configured to add the outputs of the photosensitive pixels corresponding to the white filter areas together to obtain first pixel values of merged pixels, and to add the outputs of the photosensitive pixels corresponding to the non-white filter areas together to obtain second pixel values of merged pixels.
11. A mobile terminal, comprising:
an imaging device, wherein the the imaging device comprises an image sensor and an image signal processor, the image sensor comprises a photosensitive pixel array and a filter arranged on the photosensitive array, the filter comprises a filter unit array comprised a plurality of filter units, each filter unit covers N photosensitive pixels, and some of the filter units comprise white filter areas, wherein the white filter areas cover at least one of the N photosensitive pixels, a merged pixel is formed by the N photosensitive pixels covered by the same filter unit, N is a positive integer, the image signal processor is configured to read and to process outputs of the photosensitive pixel array of the image sensor to obtain pixel values of merged pixels and further to obtain merged images.
12. The mobile terminal of claim 11, wherein the mobile terminal may be a cell phone.
13. The mobile terminal of claim 12, wherein the imaging device may be a front camera of the mobile terminal.
14. The mobile terminal of claim 11, wherein further comprising a CPU and an external memory, wherein the CPU is configured to control the external memory to store the merged images.
15. The mobile terminal of claim 14, further comprising a display device, wherein the CPU is configured to control the display device to display the merged images.
16. An imaging method of the image sensor of claim 1, comprising:
reading outputs of the photosensitive unit array; and
calculating pixel values of merged pixels according to the outputs of the photosensitive pixels of the merged pixel in order to obtain a merged image.
17. The imaging method of claim 16, wherein each filter unit comprises 2*2 photosensitive pixels; wherein the step of calculating pixel values of merged pixels according to the outputs of the photosensitive pixels of the merged pixel in order to obtain a merged image further comprises:
collecting and storing the outputs of the photosensitive pixels of row k and row k+1 into the register, wherein k=2n−1, n is a positive integer, and k+1 is less than or equal to the amount of the rows of the photosensitive pixel array; and
extracting the outputs of the photosensitive pixels of row k and row k+1 from the register to obtain pixel values of merged pixels.
18. The imaging method of claim 16, wherein when the filter units only comprise white filter areas or non-white filter areas, step of calculating pixel values of merged pixels according to the outputs of the photosensitive pixels of the merged pixel in order to obtain a merged image further comprises:
adding the outputs of the N photosensitive pixels corresponding to the same merged pixel together and making the result as the pixel value of merged pixel.
19. The imaging method of claim 16, wherein when the filter units comprise white filter areas and non-white filter areas, the pixel values of merged pixels comprise first pixel values corresponding to white filter areas and second pixel values corresponding to non-white filter areas, the step of calculating pixel values of merged pixels according to the outputs of the photosensitive pixels of the merged pixel in order to obtain a merged image further comprises:
adding the outputs of the photosensitive pixels corresponding to white filter areas together to obtain first pixel values of merged pixels; and
adding the outputs of the photosensitive pixels corresponding to non-white filter areas together to obtain second pixel values of merged pixels.
20. The imaging method of claim 16, wherein each ADC connects to one photosensitive pixel, the imaging method further comprises:
converting analog signal outputs of the photosensitive pixel into digital signal outputs; and
calculating the pixel value of the merged pixel according to the digital signal outputs of the photosensitive pixels of the same merged pixel.
US15/544,537 2015-12-18 2016-09-22 Image sensor, imaging device, mobile terminal and imaging method for producing high resolution image Active US10594962B2 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
CN201510963465.1 2015-12-18
CN201510963465 2015-12-18
CN201510963465.1A CN105516697B (en) 2015-12-18 2015-12-18 Imaging sensor, imaging device, mobile terminal and imaging method
PCT/CN2016/099753 WO2017101546A1 (en) 2015-12-18 2016-09-22 Image sensor, imaging device, mobile terminal and imaging method

Publications (2)

Publication Number Publication Date
US20180007289A1 true US20180007289A1 (en) 2018-01-04
US10594962B2 US10594962B2 (en) 2020-03-17

Family

ID=55724292

Family Applications (1)

Application Number Title Priority Date Filing Date
US15/544,537 Active US10594962B2 (en) 2015-12-18 2016-09-22 Image sensor, imaging device, mobile terminal and imaging method for producing high resolution image

Country Status (11)

Country Link
US (1) US10594962B2 (en)
EP (1) EP3242479B1 (en)
JP (1) JP6325755B2 (en)
KR (1) KR102083292B1 (en)
CN (1) CN105516697B (en)
AU (1) AU2016369789B2 (en)
MY (1) MY184809A (en)
SG (1) SG11201706246XA (en)
TW (1) TWI617196B (en)
WO (1) WO2017101546A1 (en)
ZA (1) ZA201706230B (en)

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105516697B (en) * 2015-12-18 2018-04-17 广东欧珀移动通信有限公司 Imaging sensor, imaging device, mobile terminal and imaging method
CN105578072A (en) 2015-12-18 2016-05-11 广东欧珀移动通信有限公司 Imaging method, imaging device and electronic device
CN106341670B (en) 2016-11-29 2017-09-22 广东欧珀移动通信有限公司 Control method, control device and electronic installation
CN106454054B (en) 2016-11-29 2019-03-19 Oppo广东移动通信有限公司 Control method, control device and electronic device
CN106504218B (en) 2016-11-29 2019-03-12 Oppo广东移动通信有限公司 Control method, control device and electronic device
CN107040724B (en) * 2017-04-28 2020-05-15 Oppo广东移动通信有限公司 Dual-core focusing image sensor, focusing control method thereof and imaging device
CN108269243B (en) * 2018-01-18 2021-08-31 福州鑫图光电有限公司 Image signal-to-noise ratio enhancing method and terminal
WO2019153327A1 (en) * 2018-02-12 2019-08-15 深圳市汇顶科技股份有限公司 Image acquisition method and apparatus
CN111835977B (en) * 2019-04-18 2021-11-02 北京小米移动软件有限公司 Image sensor, image generation method and device, electronic device, and storage medium
CN111756972A (en) * 2020-05-15 2020-10-09 深圳市汇顶科技股份有限公司 Image sensor and electronic device
CN113676708B (en) * 2021-07-01 2023-11-14 Oppo广东移动通信有限公司 Image generation method, device, electronic equipment and computer readable storage medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150138407A1 (en) * 2012-05-14 2015-05-21 Sony Corporation Imaging device, imaging method, electronic device, and program
US20150350583A1 (en) * 2014-06-03 2015-12-03 Semiconductor Components Industries, Llc Imaging systems having image sensor pixel arrays with sub-pixel resolution capabilities

Family Cites Families (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1324363C (en) * 2002-05-04 2007-07-04 三星电子株式会社 LCD device and filtering color picec array board
US8139130B2 (en) * 2005-07-28 2012-03-20 Omnivision Technologies, Inc. Image sensor with improved light sensitivity
EP1961214A4 (en) * 2005-10-13 2011-11-16 Rjs Technology Inc System and method for a high performance color filter mosaic array
KR100976284B1 (en) * 2007-06-07 2010-08-16 가부시끼가이샤 도시바 Image pickup device
TWI413242B (en) * 2007-08-10 2013-10-21 Hon Hai Prec Ind Co Ltd Solid image sensor
US7745779B2 (en) 2008-02-08 2010-06-29 Aptina Imaging Corporation Color pixel arrays having common color filters for multiple adjacent pixels for use in CMOS imagers
JP4683121B2 (en) * 2008-12-08 2011-05-11 ソニー株式会社 Solid-state imaging device, signal processing method for solid-state imaging device, and imaging device
US8237831B2 (en) 2009-05-28 2012-08-07 Omnivision Technologies, Inc. Four-channel color filter array interpolation
US8134115B2 (en) * 2009-06-23 2012-03-13 Nokia Corporation Color filters for sub-diffraction limit-sized light sensors
US20110013056A1 (en) * 2009-07-17 2011-01-20 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Color filters and demosaicing techniques for digital imaging
KR20110040402A (en) 2009-10-14 2011-04-20 삼성전자주식회사 Color filter array, image sensor having the same, and interpolation method thereof
US8345132B2 (en) 2010-07-23 2013-01-01 Omnivision Technologies, Inc. Image sensor with dual element color filter array and three channel color output
US8610813B2 (en) * 2011-05-31 2013-12-17 Omnivision Technologies, Inc. System and method for extending depth of field in a lens system by use of color-dependent wavefront coding
JP2013021660A (en) * 2011-07-14 2013-01-31 Sony Corp Image processing apparatus, image pickup apparatus, image processing method, and program
WO2013100033A1 (en) * 2011-12-28 2013-07-04 富士フイルム株式会社 Image processing device and method, and imaging device
US9165526B2 (en) * 2012-02-28 2015-10-20 Shenzhen Yunyinggu Technology Co., Ltd. Subpixel arrangements of displays and method for rendering the same
US9191635B2 (en) 2012-03-19 2015-11-17 Semiconductor Components Industries, Llc Imaging systems with clear filter pixels
JP5927068B2 (en) * 2012-07-06 2016-05-25 富士フイルム株式会社 Color image sensor
WO2014007281A1 (en) * 2012-07-06 2014-01-09 富士フイルム株式会社 Colour imaging element and imaging device
US9154750B2 (en) * 2013-05-28 2015-10-06 Omnivision Technologies, Inc. Correction of image sensor fixed-pattern noise (FPN) due to color filter pattern
US9692992B2 (en) * 2013-07-01 2017-06-27 Omnivision Technologies, Inc. Color and infrared filter array patterns to reduce color aliasing
TWI644568B (en) * 2013-07-23 2018-12-11 新力股份有限公司 Camera element, camera method and camera program
US10136107B2 (en) 2013-11-21 2018-11-20 Semiconductor Components Industries, Llc Imaging systems with visible light sensitive pixels and infrared light sensitive pixels
EP2887655A1 (en) * 2013-12-20 2015-06-24 Swiss Timing Ltd. Adaptive colour filter for digital sensor
CN105578078B (en) * 2015-12-18 2018-01-19 广东欧珀移动通信有限公司 Imaging sensor, imaging device, mobile terminal and imaging method
CN105578006B (en) * 2015-12-18 2018-02-13 广东欧珀移动通信有限公司 Imaging method, imaging device and electronic installation
CN105578071B (en) * 2015-12-18 2018-03-20 广东欧珀移动通信有限公司 Imaging method, imaging device and the electronic installation of imaging sensor
CN105516697B (en) * 2015-12-18 2018-04-17 广东欧珀移动通信有限公司 Imaging sensor, imaging device, mobile terminal and imaging method
CN105578066B (en) * 2015-12-18 2018-01-19 广东欧珀移动通信有限公司 Imaging method, imaging device and electronic installation

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150138407A1 (en) * 2012-05-14 2015-05-21 Sony Corporation Imaging device, imaging method, electronic device, and program
US20150350583A1 (en) * 2014-06-03 2015-12-03 Semiconductor Components Industries, Llc Imaging systems having image sensor pixel arrays with sub-pixel resolution capabilities

Also Published As

Publication number Publication date
AU2016369789B2 (en) 2019-06-27
EP3242479B1 (en) 2022-03-23
TWI617196B (en) 2018-03-01
EP3242479A1 (en) 2017-11-08
TW201724845A (en) 2017-07-01
EP3242479A4 (en) 2018-04-25
MY184809A (en) 2021-04-23
CN105516697B (en) 2018-04-17
KR102083292B1 (en) 2020-03-02
JP2018509813A (en) 2018-04-05
JP6325755B2 (en) 2018-05-16
CN105516697A (en) 2016-04-20
KR20170122772A (en) 2017-11-06
US10594962B2 (en) 2020-03-17
AU2016369789A1 (en) 2017-08-24
SG11201706246XA (en) 2017-08-30
WO2017101546A1 (en) 2017-06-22
ZA201706230B (en) 2019-01-30

Similar Documents

Publication Publication Date Title
US10594962B2 (en) Image sensor, imaging device, mobile terminal and imaging method for producing high resolution image
US10257447B2 (en) Imaging method, imaging device, and electronic device
US10764522B2 (en) Image sensor, output method, phase focusing method, imaging device, and terminal
US9979883B2 (en) Imaging method for image sensor, imaging apparatus, and electronic device
US11108943B2 (en) Image sensor, focusing control method, and electronic device
US10270988B2 (en) Method for generating high-dynamic range image, camera device, terminal and imaging method
US10893187B2 (en) Dual-core focusing image sensor, control-focusing method therefor, and mobile terminal
TWI615029B (en) Image sensor, terminal having same and imaging method using same
CN111201780B (en) Imaging apparatus and method, and image processing apparatus and method
US10313612B2 (en) Image sensor, control method, and electronic device
TW201724843A (en) Image sensor, terminal having same and imaging method using same
US9843746B2 (en) Image sensor combining high dynamic range techniques
TWI615030B (en) Image sensor, terminal device having same sensor, imaging method using same, mobile terminal device and computer readable storage media
CN110891137A (en) Image sensor, electronic device, image processing method, and storage medium

Legal Events

Date Code Title Description
AS Assignment

Owner name: GUANGDONG OPPO MOBILE TELECOMMUNICATIONS CORP., LTD., CHINA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:KANG, JIAN;REEL/FRAME:043048/0454

Effective date: 20170703

Owner name: GUANGDONG OPPO MOBILE TELECOMMUNICATIONS CORP., LT

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:KANG, JIAN;REEL/FRAME:043048/0454

Effective date: 20170703

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4