WO2009119417A1 - 撮像装置、表示撮像装置、電子機器および物体の検出方法 - Google Patents
撮像装置、表示撮像装置、電子機器および物体の検出方法 Download PDFInfo
- Publication number
- WO2009119417A1 WO2009119417A1 PCT/JP2009/055328 JP2009055328W WO2009119417A1 WO 2009119417 A1 WO2009119417 A1 WO 2009119417A1 JP 2009055328 W JP2009055328 W JP 2009055328W WO 2009119417 A1 WO2009119417 A1 WO 2009119417A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- imaging
- light
- display
- photoelectric conversion
- charge
- Prior art date
Links
- 238000003384 imaging method Methods 0.000 title claims abstract description 350
- 238000000034 method Methods 0.000 title claims description 7
- 230000007613 environmental effect Effects 0.000 claims abstract description 23
- 238000006243 chemical reaction Methods 0.000 claims description 178
- 239000003990 capacitor Substances 0.000 claims description 30
- 238000007599 discharging Methods 0.000 claims description 24
- 238000009825 accumulation Methods 0.000 claims description 16
- 238000005286 illumination Methods 0.000 claims description 12
- 239000004973 liquid crystal related substance Substances 0.000 claims description 7
- 239000011159 matrix material Substances 0.000 claims description 5
- 230000001360 synchronised effect Effects 0.000 claims description 4
- 238000004519 manufacturing process Methods 0.000 abstract description 11
- 230000015654 memory Effects 0.000 abstract description 9
- 230000004048 modification Effects 0.000 description 77
- 238000012986 modification Methods 0.000 description 77
- 238000010586 diagram Methods 0.000 description 27
- 238000000605 extraction Methods 0.000 description 19
- 238000001514 detection method Methods 0.000 description 17
- 230000006870 function Effects 0.000 description 15
- 239000000203 mixture Substances 0.000 description 10
- 230000008859 change Effects 0.000 description 8
- 230000000052 comparative effect Effects 0.000 description 7
- 230000008878 coupling Effects 0.000 description 6
- 238000010168 coupling process Methods 0.000 description 6
- 238000005859 coupling reaction Methods 0.000 description 6
- 102220008337 rs1437698471 Human genes 0.000 description 4
- 239000004065 semiconductor Substances 0.000 description 3
- 230000035945 sensitivity Effects 0.000 description 3
- 102220472121 Protein Wnt-2_C62A_mutation Human genes 0.000 description 2
- 230000007423 decrease Effects 0.000 description 2
- 230000000452 restraining effect Effects 0.000 description 2
- 230000007704 transition Effects 0.000 description 2
- 230000005540 biological transmission Effects 0.000 description 1
- 230000000903 blocking effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000000593 degrading effect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000005684 electric field Effects 0.000 description 1
- 238000005401 electroluminescence Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 239000011521 glass Substances 0.000 description 1
- 230000005484 gravity Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000001678 irradiating effect Effects 0.000 description 1
- 230000003071 parasitic effect Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 239000010409 thin film Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/0412—Digitisers structurally integrated in a display
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/0416—Control or interface arrangements specially adapted for digitisers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/0304—Detection arrangements using opto-electronic means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/042—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/042—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means
- G06F3/0421—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means by interrupting or reflecting a light beam, e.g. optical touch-screen
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G3/00—Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes
- G09G3/20—Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes for presentation of an assembly of a number of characters, e.g. a page, by composing the assembly by combination of individual elements arranged in a matrix no fixed position being assigned to or needed to be assigned to the individual characters or partial characters
- G09G3/34—Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes for presentation of an assembly of a number of characters, e.g. a page, by composing the assembly by combination of individual elements arranged in a matrix no fixed position being assigned to or needed to be assigned to the individual characters or partial characters by control of light from an independent source
- G09G3/3406—Control of illumination source
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/70—SSIS architectures; Circuits associated therewith
- H04N25/71—Charge-coupled device [CCD] sensors; Charge-transfer registers specially adapted for CCD sensors
- H04N25/75—Circuitry for providing, modifying or processing image signals from the pixel array
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N25/00—Circuitry of solid-state image sensors [SSIS]; Control thereof
- H04N25/70—SSIS architectures; Circuits associated therewith
- H04N25/76—Addressed sensors, e.g. MOS or CMOS sensors
- H04N25/78—Readout circuits for addressed sensors, e.g. output amplifiers or A/D converters
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/041—Indexing scheme relating to G06F3/041 - G06F3/045
- G06F2203/04101—2.5D-digitiser, i.e. digitiser detecting the X/Y position of the input means, finger or stylus, also when it does not touch, but is proximate to the digitiser's interaction surface and also measures the distance of the input means within a short range in the Z direction, possibly with a separate measurement setup
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2354/00—Aspects of interface with display user
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2360/00—Aspects of the architecture of display systems
- G09G2360/14—Detecting light within display terminals, e.g. using a single or a plurality of photosensors
- G09G2360/141—Detecting light within display terminals, e.g. using a single or a plurality of photosensors the light conveying information used for selecting or modulating the light emitting or modulating element
- G09G2360/142—Detecting light within display terminals, e.g. using a single or a plurality of photosensors the light conveying information used for selecting or modulating the light emitting or modulating element the light being detected by light detection means within each pixel
Definitions
- the present invention relates to an imaging device and a display imaging device for acquiring information such as the position of an object in contact with or close to a panel, an electronic apparatus provided with such a display imaging device, and an object for acquiring such information. It relates to a detection method.
- touch panels Although there are various types of touch panels, ones that are widely used include those that detect capacitance. This type is adapted to detect a change in surface charge of the panel by touching the touch panel with a finger to detect the position of an object or the like. Therefore, the user can operate intuitively by using such a touch panel.
- the present applicant also has a display unit (display imaging panel) having a display function for displaying an image and an imaging function (detection function) for imaging (detecting) an object in, for example, Patent Document 1 and Patent Document 2.
- a display device is proposed.
- Patent Document 1 When the display device described in Patent Document 1 is used, for example, when an object such as a finger is brought into contact with or in proximity to a display and imaging panel, reflected light by the irradiation light from the display and imaging panel reflected by the object It is also possible to detect the position or the like of an object based on a captured image by using Therefore, by using this display device, it is possible to detect the position or the like of the object with a simple configuration without separately providing a component such as a touch panel on the display imaging panel.
- the voltage value Va corresponds to the brightness of the ambient light L0 except for the part touched by the finger f in the display area 101, and the finger f in the display area 101 At the touched portion, the voltage value Vb corresponding to the reflectance for reflecting the irradiation light Lon from the backlight 105 is reduced to the surface of the object (finger f) touched at that time.
- the light reception output voltage Voff 101 in the state in which the backlight 105 is turned off is the same as the voltage value Va corresponding to the brightness of the ambient light L0 except for the part touched by the finger f. At the point where the finger f is touched, the ambient light L0 is blocked, and the voltage value Vc is very low.
- the light reception output voltage Von 201 in the state where the backlight 105 is lit is as shown in FIG.
- the voltage value Vc of a very low level is obtained because there is no ambient light L0 except for the part touched by the finger f in the display area 101, and the finger f in the display area 101 is touched
- the voltage value Vb corresponding to the reflectance for reflecting the irradiation light Lon from the backlight 105 is raised.
- the light reception output voltage Voff2 in the state in which the backlight 105 is turned off remains unchanged at the very low level of the voltage value Vc at the place touched by the finger f and the other places. .
- the light reception output voltage is largely different depending on whether the ambient light L0 is present or not.
- the voltage Vb when the backlight 105 is on and the voltage Vc when the backlight 105 is off regardless of the presence or absence of the ambient light L0. It is almost the same condition. Therefore, by detecting the difference between the voltage at the time of lighting of the backlight 105 and the voltage at the time of extinguishing, an object is in contact or proximity with a point where there is a certain difference or more, such as the difference between the voltage Vb and the voltage Vc. For example, as in the difference image C shown in FIG. 57, it is considered possible to detect the position of an object without being affected by external light or fixed noise. .
- the present invention has been made in view of such problems, and an object of the present invention is to provide an imaging device, a display imaging device, and an object that can stably detect an object regardless of usage conditions while suppressing manufacturing costs. It is an object of the present invention to provide a detection method and an electronic device provided with such a display and imaging device.
- An imaging apparatus includes an imaging panel having a plurality of imaging pixels, and an imaging panel having an irradiation light source for a close object, and an imaging drive for acquiring an imaging signal from each imaging pixel by performing imaging driving on each imaging pixel.
- image processing means for acquiring object information including at least one of the position, the shape, and the size of the proximity object by using a captured image based on an imaging signal obtained from each imaging pixel.
- the imaging drive means when the irradiation light is irradiated from the irradiation light source, the imaging drive means accumulates the charging charge in each imaging pixel according to the total light amount of the reflected light by the irradiation light and the environmental light, An imaging signal is obtained from each imaging pixel by performing imaging drive so that discharge charges are emitted from each imaging pixel according to the light amount of environmental light when the irradiation light is not irradiated from the light source. ing.
- the term "proximity object" is meant to include not only objects that are literally close, but also objects that are in contact.
- the display-and-image-pickup device includes a display-and-image-pickup panel including a plurality of display pixels and a plurality of image pickup pixels, a display driving unit for displaying an image by performing display driving on each display pixel, The position, shape, or size of a close object using an imaging drive unit that acquires an imaging signal from each imaging pixel by performing imaging drive on the pixels, and a captured image based on the imaging signal obtained from each imaging pixel And image processing means for acquiring object information including at least one of the above.
- the imaging drive panel is emitting illumination light from the display imaging panel to the close object, the charge is stored in each imaging pixel in accordance with the total amount of light reflected by the illumination light and the ambient light.
- imaging By performing imaging drive so that discharge charges are emitted from each imaging pixel according to the light amount of ambient light while being accumulated and irradiation light is not emitted from the display imaging panel, imaging is performed from each imaging pixel It is supposed to get a signal.
- irradiation light is the meaning including not only the display light irradiated from a display imaging panel but the light irradiated from another light source (for example, an infrared light source etc.) on a display imaging panel.
- An electronic apparatus includes the display and imaging device having an image display function and an imaging function.
- the method when irradiating light to a nearby object by an imaging panel formed of a plurality of imaging pixels, the method according to the total light amount of reflected light by the irradiation light and environmental light And charge the imaging pixels, and discharge the discharge charges from the imaging pixels according to the light amount of the environmental light when the irradiation light is not irradiated, thereby performing imaging driving on the imaging pixels. Capture an imaging signal from each imaging pixel, and use the imaging image based on the imaging signal obtained from each imaging pixel to acquire object information including at least one of the position, shape, or size of a close object It is intended to
- the image pickup apparatus when the irradiation light is irradiated to the close object by the image pickup panel or the display image pickup panel, the reflected light by the irradiation light and the ambient light Charged charge is accumulated in each imaging pixel in accordance with the total amount of light.
- discharge charges are released from the respective imaging pixels in accordance with the light amount of the environmental light.
- An imaging signal is obtained from each imaging pixel by these. Then, using captured images based on imaging signals obtained from each imaging pixel, object information including at least one of the position, the shape, and the size of the close object is obtained.
- the image pickup signal obtained from each image pickup pixel the component due to the ambient light is subtracted, so that it is possible to obtain the object information of the close object without being affected by the ambient light.
- the number of frame memories and the like necessary for generating a pickup image from the imaging signal is smaller than that in the prior art. It will end.
- the imaging device the display imaging device, the electronic apparatus, or the method for detecting an object of the present invention
- the imaging panel or the display imaging panel irradiates the near object with the irradiation light
- the reflected light by the irradiation light The charging charge is accumulated in each imaging pixel according to the total light amount with the ambient light, and the discharge charge is released from each imaging pixel according to the light amount of the environmental light when the irradiation light is not irradiated. Since an imaging signal is acquired from each imaging pixel, it is possible to acquire object information of a close object without being affected by ambient light, and a frame memory necessary for generating an imaging image from an imaging signal. Etc. will be less than before. Therefore, it becomes possible to detect an object stably irrespective of use conditions, restraining a manufacturing cost.
- FIG. 7 is a timing waveform chart illustrating an example of an imaging operation in the display and imaging device according to the embodiment.
- FIG. 8 is a timing waveform chart for explaining the details of the imaging operation shown in FIG. 7;
- FIG. 9 is a circuit diagram for illustrating charge operation and discharge operation in the imaging operation shown in FIGS. 7 and 8;
- FIG. 9 is a timing waveform chart for explaining the details of the SW control signal and the backlight state shown in FIG. 8;
- It is a characteristic view for explaining imaging operation concerning a comparative example.
- It is a schematic diagram for demonstrating the imaging operation shown in FIG.
- It is another characteristic view for demonstrating the imaging operation which concerns on a comparative example.
- It is another characteristic view for demonstrating the imaging operation which concerns on a comparative example.
- FIG. 14 is a circuit diagram for describing a charge operation and a discharge operation at the time of an imaging operation according to Modification 1;
- FIG. 13 is a timing waveform diagram for describing a SW control signal and a backlight state in an imaging operation according to Modification Example 1.
- It is a circuit diagram showing the example of composition of the other light sensing portion concerning modification 1 of the present invention.
- FIG. 14 is a timing waveform chart for describing an imaging operation according to Modification 2;
- FIG. 13 is a circuit diagram for describing a charge operation and a discharge operation at the time of an imaging operation according to Modification 2;
- FIG. 16 is a timing waveform diagram for describing a SW control signal and a backlight state in an imaging operation according to Modification Example 2.
- It is a circuit diagram showing the example of composition of the light sensing portion in the display area (sensor area) concerning modification 3 of the present invention. It is a figure for demonstrating the on-operation area
- FIG. 14 is a timing waveform chart for describing an imaging operation according to Modification 3;
- FIG. 18 is a timing waveform chart for describing in detail an on operation and an off operation in the photoelectric conversion element according to the modification example 3;
- FIG. 16 is a circuit diagram for describing charge operation and discharge operation in an imaging operation according to Modification 3; It is a circuit diagram showing the example of composition of the light sensing portion in the display area (sensor area) concerning modification 4 of the present invention.
- FIG. 18 is a diagram for describing an on-operation area and an off-operation area in the photoelectric conversion element according to Modification Example 4.
- FIG. 18 is a diagram for describing an on-operation area and an off-operation area in the photoelectric conversion element according to Modification Example 4.
- FIG. 18 is a timing waveform chart for describing an imaging operation according to Modification 4;
- FIG. 18 is a timing waveform diagram for describing in detail an on operation and an off operation in the photoelectric conversion element according to the modification 4;
- FIG. 17 is a circuit diagram for describing charge operation and discharge operation in an imaging operation according to Modification Example 4; It is a circuit diagram showing the example of composition of the light sensing portion in the display area (sensor area) concerning modification 5 of the present invention. It is a figure for demonstrating the on-operation area
- FIG. 21 is a timing waveform chart for describing an imaging operation according to a fifth modification.
- FIG. 18 is a timing waveform chart for describing in detail an on operation and an off operation in the photoelectric conversion element according to the modification example 5;
- FIG. 17 is a circuit diagram for describing charge operation and discharge operation in an imaging operation according to Modification Example 5; It is a circuit diagram showing the example of composition of the light sensing portion in the display area (sensor area) concerning modification 6 of the present invention.
- FIG. 1 It is a top view showing the example of pixel arrangement in the display area (sensor area) shown in FIG. It is a figure for demonstrating an example of the application using the imaging result (result of a fingertip extraction process) by this invention. It is a figure for demonstrating an example of the application using the imaging result (result of a fingertip extraction process) by this invention. It is a figure for demonstrating an example of the application using the imaging result (result of a fingertip extraction process) by this invention. It is a figure for demonstrating an example of the application using the imaging result (result of a fingertip extraction process) by this invention. It is a perspective view showing the appearance of application example 1 of the display imaging device of the present invention.
- FIG. 21 is a perspective view illustrating an appearance of application example 3; 21 is a perspective view illustrating an appearance of application example 4.
- FIG. (A) is a front view of the open state of application example 5, (B) is a side view thereof, (C) is a front view of a closed state, (D) is a left side view, (E) is a right side view, (F) is a top view, (G) is a bottom view.
- It is a characteristic view showing the other example of the imaging operation (difference image fingertip extraction process) by the conventional display imaging device.
- FIG. 1 shows the entire configuration of a display-and-image-pickup apparatus according to an embodiment of the present invention.
- the display and imaging apparatus includes an I / O display panel 20, a backlight 15, a display drive circuit 12, a light reception drive circuit 13, an image processing unit 14, and an application program execution unit 11.
- the I / O display panel 20 is a liquid crystal panel (LCD (Liquid Crystal Display)) in which a plurality of pixels are arranged in a matrix over the entire surface, and predetermined figures and characters based on display data while performing line sequential operation.
- LCD Liquid Crystal Display
- the backlight 15 is a light source for display and detection of the I / O display panel 20 in which, for example, a plurality of light emitting diodes are arranged, and the predetermined timing synchronized with the operation timing of the I / O display 20 as described later. At the timing, the on / off operation is performed at high speed.
- the display drive circuit 12 drives the I / O display panel 20 so that an image based on display data is displayed on the I / O display panel 20 (to perform a display operation) (line sequential display operation). Drive) circuit.
- the light reception drive circuit 13 drives the I / O display panel 20 so that a light reception signal (imaging signal) can be obtained from each pixel of the I / O display panel 20 (for imaging an object) (line sequential) Drive the imaging operation).
- the light reception signal from each pixel is accumulated, for example, in frame units in the frame memory 13A, and is output to the image processing unit 14 as a captured image.
- the image processing unit 14 performs predetermined image processing (arithmetic processing) based on the captured image output from the light receiving drive circuit 13, and performs object information (position coordinate data, object related to an object in contact with or in proximity to the I / O display 20). Data related to the shape and size of the The details of the processing to be detected will be described later.
- the application program execution unit 11 executes processing according to predetermined application software based on the detection result by the image processing unit 14, and, for example, includes position coordinates of a detected object in display data, and the I / O What is displayed on the display panel 20 may, for example, be mentioned.
- the display data generated by the application program execution unit 11 is supplied to the display drive circuit 12.
- the I / O display panel 20 has a display area (sensor area) 21, a display H driver 22, a display V driver 23, a sensor readout H driver 25, and a sensor V driver 24. There is.
- the display area (sensor area) 21 modulates the light from the backlight 15 and includes irradiation light (display light and irradiation light for detection by, for example, an infrared light source or the like (not shown).
- irradiation light display light and irradiation light for detection by, for example, an infrared light source or the like (not shown).
- Liquid crystal element which is a light emitting element (display element) and a light receiving element (image pickup element) described later are arranged in a matrix, respectively. There is.
- the display H driver 22 sequentially drives the liquid crystal elements of the respective pixels in the display area 21 together with the display V driver 23 on the basis of the display drive signal and control clock supplied from the display drive circuit 12. It is.
- the sensor readout H driver 25 line-sequentially drives the light receiving elements of the respective pixels in the sensor area 21 together with the sensor V driver 24 to acquire a light reception signal.
- the sensor readout H driver 25 and the sensor V driver 24 will be described later in detail, but when irradiation light from the I / O display panel 20 is irradiated to the close object, reflection by the irradiation light is performed. While charge charges are accumulated in each pixel according to the combined light amount of light and ambient light, and when the irradiation light is not irradiated from the I / O display panel 20 panel, according to the light amount of ambient light An image pickup signal (light reception signal) is acquired from each pixel by performing an image pickup drive so that a discharge charge is released.
- the pixel 31 is configured of a display pixel (display unit) 31 RGB, an imaging pixel (light receiving unit) 33, and a wiring unit 32 in which a wiring for the imaging pixel 33 is formed. It is done.
- the display pixel 31RGB is composed of a display pixel 31R for red (R), a display pixel 31G for green (G), and a display pixel 31B for blue (B).
- the display pixels 31 RGB, the imaging pixels 33, and the wiring sections 32 are arranged in a matrix on the display area 21 (sensor area). Further, the imaging pixel 33 and the wiring portion 32 for imaging and driving the imaging pixel 33 are separately arranged at a constant cycle.
- the display pixel 31RGB With such an arrangement, it becomes extremely difficult for the display pixel 31RGB to recognize the sensor rear composed of the imaging pixel 32 and the wiring section 33, and reduction of the aperture ratio in the display pixel 31RGB can be minimized.
- the wiring portion 32 is arranged in an area not contributing to the opening of the display pixel 31RGB (for example, an area shielded by black matrix, a reflection area, etc.), the light receiving circuit is arranged without degrading the display quality. It becomes possible. Although details will be described later, for example, as illustrated in FIG. 4, reset signal lines Reset_1 to Reset_n and read signal lines Read_1 to Read_n are connected to each imaging pixel (light receiving unit) 33 along the horizontal line direction. It is supposed to be
- each light receiving unit 33 includes a photoelectric conversion element PD1 as a light receiving element that generates an electric charge according to the amount of received light, a capacitor C1 as a capacitance element, and four switching elements SW11 to It is configured to include the SW 14 and three transistors Tr1 to Tr3.
- the photoelectric conversion element PD1 generates a charge and a discharge for the capacitor C1 in common, and is formed of, for example, a photodiode or a phototransistor.
- the photoelectric conversion element PD1 is a PIN photodiode having an anode (formed in a p-type semiconductor region), a cathode (formed in an n-type semiconductor region), and a gate (formed in an intrinsic semiconductor region). The same applies to a photoelectric conversion element according to a modification, which is configured and will be described later.
- Each of the transistors Tr1 to Tr3 is formed of, for example, a thin film transistor (TFT) or the like.
- TFT thin film transistor
- one end of the switching element SW1 is connected to the power supply VDD, and the other end is connected to one end of the switching element SW13 and the cathode and gate of the photoelectric conversion element PD1.
- one end of the switching element SW12 is connected to the ground VSS, and the other end is connected to one end of the switching element SW14 and the anode of the photoelectric conversion element PD1.
- the other ends of the switching elements SW13 and SW14 are connected to one end of the capacitor C1, the drain of the transistor Tr1 and the gate of the transistor Tr2 through the connection point P1.
- the other end of the capacitor C1 is connected to the ground VSS.
- the gate of the transistor Tr1 is connected to the reset signal line Reset, and the source is connected to the reset power supply Vrst.
- the source of the transistor Tr2 is connected to the power supply VDD, and the drain is connected to the drain of the transistor Tr3.
- the gate of the transistor Tr3 is connected to the read signal line Read, and the source is connected to the read line 41.
- the switching elements SW11 and SW14 that function as charging switching elements constitute a charging circuit for accumulating (charging) the charge in the capacitor C1.
- the switching elements SW12 and SW13 which function as switching elements for discharging constitute a discharging circuit for discharging the discharge charge from the capacitor C1.
- the switching elements SW11 to SW14 in the charging circuit and the discharging circuit are respectively imaged and driven so that the light is discharged from the capacitor C1 via the switching elements SW12 and SW13 for discharging, and imaging is performed from each light receiving portion (imaging pixel) 33 A signal can be obtained.
- the voltage of the reset power supply Vrst described above (the reset voltage for releasing all the charge accumulated in the capacitor C1 in each light receiving unit 33) can be set to any voltage between the power supply voltage VDD and the ground voltage VSS. It has become.
- a drive signal for display is generated in the display drive circuit 12 based on the display data supplied from the application program execution unit 11, and this drive signal causes a line to the I / O display 20.
- Display drive is sequentially performed to display an image. Further, at this time, the backlight 15 is also driven by the display drive circuit 12, and the on / off operation synchronized with the I / O display 20 is performed.
- line-sequential light reception driving by the light reception drive circuit 13 causes each imaging pixel 33 in this I / O display panel 20 to An object is imaged, and an imaging signal from each imaging pixel 32 is supplied to the light receiving drive circuit 13.
- the light reception drive circuit 13 accumulates the image pickup signal of the image pickup pixel 33 for one frame, and outputs the image pickup signal to the image processing unit 14 as a pickup image.
- the image processing unit 14 performs predetermined image processing (arithmetic processing) to be described below based on the captured image to obtain object information (position coordinate data, etc.) regarding an object in contact with or in proximity to the I / O display 20. Data on the shape and size of the object can be obtained.
- FIG. 7 Details of the imaging operation of the present embodiment and extraction processing of an object (proximity object such as a finger tip or the like) touching or approaching I / O display 20 by image processing unit 14
- an object proximity object such as a finger tip or the like
- FIG. 7 is a timing waveform diagram showing an example of the imaging operation of the present embodiment, and (A) to (D) are reset signal voltages V (Reset_1), V (Reset_2), V (Reset_3), V (Reset_n), (E) to (H) are read signal voltages V (Read_1), V (Read_2), V (Read_3), V (Read_n), (I) is a switching element
- the control signal (SW control signal Sa) of SW11 and SW14 and (J) represent the control signal (SW control signal Sb) of the switching elements SW12 and SW13, respectively.
- 8 is a timing waveform diagram showing the details of the imaging operation shown in FIG.
- V (Read_n) become H (high) state by the line sequential operation.
- a period from when the reset signal voltage V (Reset) is in the H state to when the read signal voltage V (Read) is in the H state will be described later.
- the ON state and the OFF state of the backlight 15 are alternately switched in synchronization with the line sequential operation of the reset signal voltage V (Reset) and the read signal voltage V (Read).
- the SW control signals Sa and Sb respectively become H (high) states in synchronization with the switching operation of the light 15 in the ON / OFF state, and the switching SW 11 to SW 14 are respectively turned on. Also, the SW control signal Sa and the SW control signal Sb are alternately in the H state. Specifically, when the backlight 15 is in the ON state, the SW control signal Sa is in the H state and the SW control signal Sb is in the L (low) state, and when the backlight 15 is in the OFF state, the SW control signal Sa is While being in the L state, the SW control signal Sb is in the H state.
- the backlight 15 is turned on, the SW control signal Sa is in the H state, and the SW control signal Sb is in the L state, thereby accumulating charge of the capacitor C1.
- (Charging operation) is performed in synchronization with a horizontal period (display operation switching ON / OFF of the backlight 15) in display driving. Specifically, for example, as shown in FIG. 9A, the charging switching elements SW11 and SW14 are turned on, and the discharging switching elements SW12 and SW13 are turned off.
- the charge is charged to the capacitor C1 through the path of the charging current I11 Is accumulated, and the accumulated potential VP1 rises.
- the backlight 15 is turned off, the SW control signal Sa is turned to the L state, and the SW control signal Sb is turned to the H state, thereby discharging the discharge charge from the capacitor C1.
- discharge operation is performed in synchronization with a horizontal period (display operation switching ON / OFF of the backlight 15) in display driving.
- the switching elements SW11 and SW14 for charging are turned off, and the switching elements SW12 and SW13 for discharging are turned on.
- the discharge charge is discharged from the capacitor C1 through the path of the discharge current I12 shown in the figure, and the accumulated potential VP1 drops.
- the charge charge accumulation operation and the discharge charge emission operation are switched a plurality of times until timing t14 (during the exposure period), and thereafter, the charge accumulated in the capacitor C1 is read out as an imaging signal Be Specifically, when the read signal voltage V (Read) changes to the H state at timing t14, the transistor Tr3 is turned on, and the voltage of the storage potential VP1 is read from the read line 41 in the period from timing t15 to t16. . In this manner, since the imaging signal is read out after the charge charge accumulation operation and the discharge charge emission operation are switched a plurality of times, the exposure period becomes longer, as shown in FIG. The signal component (accumulated potential VP1) of the imaging signal is increased.
- the imaging signal obtained here is an analog value
- a / D (analog / digital) conversion is performed in the light receiving drive circuit 13. After that, the reset signal voltage V (Reset) becomes H state at timing t16, and thereafter, the same operation as timings t10 to t16 is repeated.
- the period in which the SW control signal Sa is in the H state and the period in which the SW control signal Sb is in the H state do not overlap each other in more detail, for example, as shown in FIG. A lap period is set), whereby generation of a leak current in the light receiving unit 33 is avoided.
- the image processing unit 14 calculation processing is performed to determine the center of gravity of the captured image for one frame generated in the light reception drive circuit 13, and the contact (proximity) center is specified. Then, the detection result of the proximity object is output from the image processing unit 14 to the application program execution unit 11, and the fingertip extraction process by the image processing unit 14 is completed.
- the image processing unit 14 obtains object information including at least one of the position, the shape, or the size of the close object.
- the imaging signal obtained from each imaging pixel 33 the component due to the ambient light L0 is subtracted, so that it is possible to obtain the object information of the close object without being affected by the ambient light L0.
- a frame memory necessary for generating a pickup image from the imaging signal in the light reception drive circuit 13 13A requires a frame memory for two images of the prior art (for example, as shown in FIG. 58, an image when the backlight is off (image A) and an image when the backlight is on (image B) Will be less than).
- object information such as position, shape, or size of each of the plurality of contacts or approaching objects simultaneously disposed on the display area 21 of the I / O display panel 20 is similarly obtained. Is available for acquisition.
- the display area 21 of the I / O display panel 20 If a close object is moving above, the following problems occur. That is, under actual use conditions, as shown in FIG. 12 for example, there is a time difference between the image A 101 obtained in the back light off state and the image B 101 obtained in the back light on state . Therefore, for example, when the close object is moving at high speed on the I / O display panel 20 as shown in FIG. 11, the light reception output signal Voff (A101) and the image B101 in the image A101 are generated due to this time difference.
- At least the proximity object is moving on the I / O display panel 20, for example, charge charges are accumulated in each imaging pixel 33 and the discharge is performed.
- an imaging signal is acquired from each imaging pixel 33, and object information of a close object is acquired using a captured image based on such an imaging signal. It has become. Therefore, for example, by setting the switching cycle of the ON / OFF state of the backlight 15 extremely short, it is assumed that the close object is moving on the display area 21 of the I / O display panel 20 during such switching cycle. Even in such a case, the movement distance is extremely short, so the generation of false signals is minimized (or avoided).
- the following problems may occur. That is, in each imaging pixel in the actual I / O display panel 20, the storage capacity for light is limited.
- the case where the proximity object is the finger f or the like
- the shadowed portion of the finger f does not exceed the storage capacity, so that it is possible to detect a close object.
- FIGS. 13A and 13B when the light amount of the external light L0 is large and the charge flows in more than the full storage capacity, the case where the proximity object is the finger f or the like As indicated by the light reception output signals Von 301 and Voff 301 in the drawing, the shadowed portion of the finger f does not exceed the storage capacity, so that it is possible to detect a close object.
- the close object when the close object is a very thin object such as the pen ob1, or when the close object is separated from the I / O display panel 20, etc.
- the shadow portion of the pen ob 1 may reach a high level beyond the storage capacity.
- the difference signal between the light reception output signals Von 401 and Voff 401 exceeds the storage capacity of the imaging pixel, and it becomes impossible to detect a close object using the difference result.
- a finite storage capacity is a component of the external light L0 under any external light condition. It can be avoided that the battery is used up (capacitance accumulated by one charge operation and discharge operation is determined by the switching cycle of the ON / OFF state of the backlight 15).
- each imaging pixel 33 when the I / O display panel 20 irradiates the near object with the irradiation light from the backlight 15, the reflected light Lon by the irradiation light and the ambient light (external light ) Makes each imaging pixel 33 accumulate charge charge according to the combined light amount with L0, and discharge the discharge charge from each imaging pixel 33 according to the light amount of the environmental light L0 when the irradiation light is not irradiated.
- the image pickup signal is obtained from each image pickup pixel 33, in the image pickup signal obtained from each image pickup pixel 33, the component due to the environmental light L0 is subtracted, so that the influence of the environmental light L0 is not received.
- the light receiving drive circuit 13 needs to generate a pickup image from the pickup signal.
- the number of frame memories 13A can be smaller than in the prior art (the frame memory for difference detection is not required). Therefore, it becomes possible to detect an object stably irrespective of use conditions, restraining a manufacturing cost.
- the photoelectric conversion element PD1 in each imaging pixel (light receiving portion) 33, the photoelectric conversion element PD1 generates charge charge and discharge charge in common, and the charge charge generated in the photoelectric conversion element PD1 is used for charging. Because the discharge charge generated in the photoelectric conversion element PD1 is released from the capacitor C1 via the switching elements SW12 and SW13 for discharge, while being stored in the capacitor C1 via the switching elements SW11 and SW14. It is possible to obtain an effect like
- the exposure period can be extended.
- the detection sensitivity can be improved by increasing the signal component (accumulated potential VP1) of the above, and the exposure time can be freely set, so that the S / N ratio can be increased.
- the imaging drive according to the present embodiment it is possible to remove the dark current component due to the temperature change.
- shadow detection can be performed using the external light L0.
- FIG. 15 illustrates a circuit configuration of a light receiving unit (light receiving unit 33A) according to the first modification.
- the light receiving unit (imaging pixel) 33A includes two photoelectric conversion elements PD21 and PD22 and two switching elements SW21 and SW22.
- the cathode and the gate of the photoelectric conversion element PD21 are connected to the power supply VDD, and the anode is connected to one end of the switching element SW21.
- the anode of the photoelectric conversion element PD22 is connected to the ground VSS, and the cathode and the gate are connected to one end of the switching element SW22.
- the other ends of the switching elements SW21 and SW22 are connected to the connection point P1.
- an imaging pixel 33A and a wiring section for imaging driving the imaging pixel 33A. 32 are arranged separately from each other.
- the other circuit configuration is the same as that of the light receiving unit 33, so the description will be omitted.
- the photoelectric conversion element PD21 functions as a photoelectric conversion element for generating charge charge
- the photoelectric conversion element PD22 functions as a photoelectric conversion element for generating discharge charge.
- the switching element SW21 functions as a charging switching element
- the switching element SW22 functions as a discharging switching element.
- the charge generated by the photoelectric conversion element PD21 (based on the combined light amount of the reflected light Lon and the external light L0) passes through the charging switching element SW21.
- the discharge charge (based on the light quantity of the external light L0) generated in the photoelectric conversion element PD22 is stored in C1 via the switching element SW22 for discharging the capacitor C1.
- the imaging drive is performed so as to be emitted from the
- the wiring capacitance in the signal lines of the SW control signals Sa and Sb is The power consumption can also be reduced.
- the SW control signal Sa is in the H state, for example, as shown in FIG.
- the period may overlap with the period in which the SW control signal Sb is in the H state (a non-overlap period is not set).
- switching elements SW23 and SW24 for removing charges accumulated in parasitic capacitances when switching elements SW21 and SW22 are in the off state for example, as in light receiving unit (imaging pixel) 33B shown in FIG. Is preferably provided.
- the switching element SW23 is provided between the anode of the photoelectric conversion element PD21 and one end of the switching element SW21, and the ground VSS.
- the switching element SW24 is provided between the cathode of the photoelectric conversion element PD22 and one end of the switching element SW22, and the power supply VDD.
- FIG. 20 illustrates a circuit configuration of a light receiving unit (light receiving unit 33C) according to the second modification.
- the light receiving unit (imaging pixel) 33C is configured to include two photoelectric conversion elements PD31 and PD32, and four switching elements SW31 to SW34. Specifically, one ends of the switching elements SW31 and SW33 are connected to the power supply VDD, the other end of the switching element SW31 is connected to the cathode and gate of the photoelectric conversion element PD31, and the other end of the switching element SW33 is one end of the capacitor C1. And one end of the switching element SW34.
- the anode of the photoelectric conversion element PD31 is connected to the other end of the capacitor C1, the cathode and gate of the photoelectric conversion element PD32, and the connection point P1.
- the anode of the photoelectric conversion element PD32 is connected to one end of the switching element SW32. Further, the other ends of the switching elements SW32 and SW34 are connected to the ground VSS.
- the other circuit configuration is the same as that of the light receiving units 33 and 33A, and thus the description thereof is omitted.
- the timing in the figure is determined by the SW control signals S1 to S4 for the switching elements SW31 to SW34, the read signal voltage V (Read) and the reset signal voltage V (Reset).
- the imaging operation is performed as in t20 to t26.
- the charge generated by the photoelectric conversion element PD31 (based on the combined light amount of the reflected light Lon and the external light L0) is the switching elements SW31 and SW34 for charging.
- FIG. 22A the charge generated by the photoelectric conversion element PD31 (based on the combined light amount of the reflected light Lon and the external light L0) is the switching elements SW31 and SW34 for charging.
- the discharge charge (based on the light quantity of the external light L0) generated by the photoelectric conversion element PD32 is discharged to the switching element SW32,
- the imaging drive is performed so as to be released from the capacitor C1 via the SW33.
- one end side of the capacitor C1 is alternately connected to the power supply VDD or the ground VSS via the switching element SW33 or the switching element SW34 at the time of accumulation of charge and discharge of discharge. ing.
- the reset voltage applied to both ends of the photoelectric conversion elements PD31 and PD32 can be approximately (VDD-VSS) without arbitrarily inputting V.sub.D, and the number of reset signal lines can be reduced.
- the SW control signal Sa is in the H state, for example, as illustrated in FIG. It is preferable that the period and the period in which the SW control signal Sb is in the H state do not overlap each other (the non-overlap period is set).
- FIG. 24 illustrates a circuit configuration of a light receiving unit (light receiving unit 33D) according to the third modification.
- the light receiving unit (imaging pixel) 33D is configured to include two photoelectric conversion elements PD41 (first photoelectric conversion elements) and PD42 (second photoelectric conversion elements).
- the photoelectric conversion element PD41 has a cathode connected to the power supply VDD, an anode connected to the cathode of the photoelectric conversion element PD42 and the connection point P1, and a gate connected to the signal line ⁇ 1.
- the photoelectric conversion element PD42 has an anode connected to the ground VSS and a gate connected to the signal line ⁇ 2.
- the light receiving portion 33D is not provided with switching elements, while the gate potentials of the photoelectric conversion elements PD41 and PD42 are controlled by the signal lines ⁇ 1 and ⁇ 2. It is supposed to be The other circuit configuration is the same as that of the light receiving unit 33 and the like, and thus the description thereof is omitted.
- the photoelectric conversion element itself is switched between the on operation area and the off operation area instead of the switching operation of the current path using the switching element described above. , Switching operation of the current path.
- the photoelectric conversion elements PD41 and PD42 are changed by changing the potential relation between the cathode and the gate in the photoelectric conversion element PD41 and the potential relation between the anode and the gate in the photoelectric conversion element PD42. Are set individually to the on / off state.
- the anode potential of the photoelectric conversion elements PD41 and PD42 is Vp
- the cathode potential is Vn
- the gate potential is Vg
- the photocurrent flowing from the cathode to the anode is Inp.
- the IV characteristics of the conversion elements PD41 and PD42 are represented, for example, as shown in FIG. 25 (B). That is, when the gate potential (gate voltage) Vg is changed, a voltage region of Vp ⁇ Vg ⁇ Vn is an on operation region ⁇ in which the photoelectric conversion elements PD41 and PD42 are turned on.
- the voltage region of Vn ⁇ Vg becomes the off operation region ⁇ 1 in which the photoelectric conversion elements PD41 and PD42 are in the off state
- the voltage region of Vg ⁇ Vp also makes the photoelectric conversion elements PD41 and PD42 in the off state It becomes the region ⁇ 2.
- the photoelectric conversion elements PD41 and PD42 are changed by changing the potential relation between the cathode and the gate in the photoelectric conversion element PD41 and the potential relation between the anode and the gate in the photoelectric conversion element PD42.
- the photoelectric conversion elements PD41 and PD42 are individually set to the on / off state by actively utilizing such a change in photoelectric conversion efficiency. There is.
- the cathode potential in the photoelectric conversion element PD41 is fixed to the power supply VDD, and the anode potential in the photoelectric conversion element PD42 is fixed to the ground VSS.
- the gate voltages Vg in these photoelectric conversion elements PD41 and PD42 are individually changed by the signal lines ⁇ 1 and ⁇ 2.
- gate voltage Vg of photoelectric conversion element PD41 is alternately changed between Vg11 and Vg12 by potential V ( ⁇ 1) of signal line ⁇ 1 (see arrow P41 in FIG. 26C).
- Vg12 ⁇ Vg11 the cathode potential in the photoelectric conversion element PD41 is fixed to the power supply VDD
- the anode potential in the photoelectric conversion element PD42 is fixed to the ground VSS.
- the gate voltages Vg in these photoelectric conversion elements PD41 and PD42 are individually changed by the signal lines ⁇ 1 and ⁇ 2.
- gate voltage Vg of photoelectric conversion element PD41 is alternately changed between Vg11 and Vg12 by potential V (
- the gate voltage Vg of the photoelectric conversion element PD42 is alternately changed between Vg21 and Vg22 by the potential V ( ⁇ 2) of the signal line ⁇ 2 (see arrow P42 in FIG. 26C; Vg22 ⁇ Vg21).
- the on / off state of the photoelectric conversion element PD41 can be arbitrarily set by the potential V ( ⁇ 1) of the signal line ⁇ 1.
- the on / off state of the photoelectric conversion element PD42 can be arbitrarily set by the potential V ( ⁇ 2) of the signal line ⁇ 2 .
- the reset signal voltage V (Reset) and the read signal voltage V (Read) The imaging operation is performed at the timing shown in the figure by the potentials V ( ⁇ 1) and V ( ⁇ 2) of the signal lines ⁇ 1 and ⁇ 2 and the reset power supply Vrst.
- ⁇ 10 and ⁇ 20 respectively indicate transition regions from the on operation region to the off operation region in the photoelectric conversion elements PD41 and PD42.
- the charge generated by the photoelectric conversion element PD41 (based on the combined light amount of the reflected light Lon and the external light L0) is turned on by the photoelectric conversion element PD41.
- the current is accumulated in the capacitor C1 by the current path I41 in the figure.
- the discharge charge (based on the light amount of the external light L0) generated in the photoelectric conversion element PD42 turns on the photoelectric conversion element PD42 and the photoelectric conversion element PD41. Is released from the capacitor C1 by the current path I42 in the figure.
- the current flowing even when the photoelectric conversion elements PD41 and PD42 are in the off state is not 0, exactly, the current of the photoelectric conversion element in the on state and the photoelectric conversion element in the off state The difference causes charging or discharging.
- the imaging operation as described above makes it possible to stably detect an object regardless of the use condition while suppressing the manufacturing cost.
- FIG. 30 illustrates a circuit configuration of a light receiving unit (light receiving unit 33E) according to the fourth modification.
- the light receiving unit (imaging pixel) 33E is configured to include two photoelectric conversion elements PD51 (first photoelectric conversion elements) and PD 52 (second photoelectric conversion elements).
- the cathode is connected to the signal line ⁇ 1
- the anode is connected to the cathode of the photoelectric conversion element PD52 and the connection point P1
- the gate is connected to the power supply VDD.
- the photoelectric conversion element PD52 has an anode connected to the signal line ⁇ 2 and a gate connected to the ground VSS.
- the gate potentials of the photoelectric conversion elements PD51 and PD52 are fixed by the power supply VDD or the ground VSS, while the cathode potential of the photoelectric conversion element PD51
- the anode potential of the conversion element PD52 is controlled by the signal lines ⁇ 1 and ⁇ 2.
- the other circuit configuration is the same as that of the light receiving unit 33D, so the description will be omitted.
- the gate potential of the photoelectric conversion element PD51 is fixed to the power supply VDD, and the gate potential of the photoelectric conversion element PD52 is fixed to the ground VSS.
- the cathode voltage Vn in the photoelectric conversion element PD51 is changed by the signal line ⁇ 1.
- cathode voltage Vn of photoelectric conversion element PD51 is alternately changed between Vn1 and Vn2 by potential V ( ⁇ 1) of signal line ⁇ 1 (see arrow P51 in FIG. 31C). Vn2 ⁇ Vn1).
- anode voltage Vp of photoelectric conversion element PD52 is alternately changed between Vp1 and Vp2 by potential V ( ⁇ 2) of signal line ⁇ 2 (see arrow P52 in FIG. 32C).
- Vp2 ⁇ Vp1).
- the reset signal voltage V (Reset) and the read signal voltage V (Read) The imaging operation is performed at the timing shown in the figure by the potentials V ( ⁇ 1) and V ( ⁇ 2) of the signal lines ⁇ 1 and ⁇ 2 and the reset power supply Vrst.
- ⁇ 120 and ⁇ 210 respectively indicate transition regions from the on operation region to the off operation region in the photoelectric conversion elements PD51 and PD52.
- the charge generated by the photoelectric conversion element PD51 (based on the total light amount of the reflected light Lon and the external light L0) is turned on by the photoelectric conversion element PD51.
- the current is accumulated in the capacitor C 1 by the current path I 51 in the figure.
- the discharge charge (based on the light amount of the external light L0) generated in the photoelectric conversion element PD52 turns on the photoelectric conversion element PD52 and the photoelectric conversion element PD51. Is released from the capacitor C1 by the current path I52 in the figure.
- the current flowing even when the photoelectric conversion elements PD51 and PD52 are in the off state is not 0, exactly, the current of the photoelectric conversion element in the on state and the photoelectric conversion element in the off state The difference causes charging or discharging.
- the imaging operation as described above makes it possible to stably detect an object regardless of the use condition while suppressing the manufacturing cost.
- FIG. 36 illustrates a circuit configuration of a light receiving unit (light receiving unit 33F) according to the fifth modification.
- the light receiving unit (imaging pixel) 33F includes two photoelectric conversion elements PD61 (first photoelectric conversion element) and PD62 (second photoelectric conversion element), and five transistors Tr1 to Tr3 and Tr61 (N channel transistors), It comprises Tr62 (P-channel transistor) and four capacitive elements C61A, C61B, C62A, C62B.
- the cathode is connected to the signal line ⁇ 1 and the source of the transistor Tr61
- the anode is connected to the cathode of the photoelectric conversion element PD62 and the connection point P1
- the gate is the drain and capacitance element of the transistor Tr61. It is connected to one end of C61A and C61B.
- the gate of the transistor Tr61 is connected to the other end of the capacitive element C61A and the reset line Reset1 (corresponding to the reset line Reset described above).
- the anode is connected to the signal line ⁇ 2 and the source of the transistor Tr62, and the gate is connected to the drain of the transistor Tr62 and one end of the capacitive elements C62A and C62B.
- the gate of the transistor Tr62 is connected to the other end of the capacitive element C61A and the inverted signal of the reset line Reset2.
- the other ends of the capacitive elements C61B and C62B are connected to the power supply VDD.
- the gate potentials of the photoelectric conversion elements PD61 and PD62 are not fixed, and after the precharging period to the reading period (during charge charge accumulation) And at the time of discharge discharge), it is in a high impedance state.
- the other circuit configuration is the same as that of the light receiving unit 33E, so the description will be omitted.
- the cathode potential Vn in the photoelectric conversion element PD61 is changed by the signal line ⁇ 1.
- the gate voltage Vg1 in the photoelectric conversion element PD61 is changed by turning on and off between the signal line ⁇ 1 and the gate electrode via the switching element SW61 formed by the transistor Tr61.
- the cathode potential Vn in the photoelectric conversion element PD61 is set between Vn1 and Vn2 by the potential V ( ⁇ 1) of the signal line ⁇ 1. Alternately (see arrow P611 in FIG. 38A; Vn2 ⁇ Vn1).
- the switching element SW61 is turned on to reset the gate voltage Vg1 of the photoelectric conversion element PD61 to Vn1, and then the switching element SW61 is turned off.
- Vn Vn2
- Vg Vg12
- the on / off state of the photoelectric conversion element PD61 can be arbitrarily set by the operation of the switching element SW61.
- the anode potential Vp in the photoelectric conversion element PD62 is changed by the signal line ⁇ 2.
- the gate voltage Vg2 in the photoelectric conversion element PD62 is changed by turning on and off between the signal line ⁇ 2 and the gate electrode via the switching element SW62 formed of the transistor Tr62.
- the anode potential Vp in the photoelectric conversion element PD62 is set between Vp1 and Vp2 by the potential V (.phi.2) of the signal line .phi.2. Alternately (see arrow P621 in FIG. 40A; Vp2 ⁇ Vp1).
- the gate voltage Vg2 of the photoelectric conversion element PD62 is reset to Vp2, and then the switching element SW62 is turned off.
- Vp Vp1
- the gate voltage Vg of the photoelectric conversion element PD62 decreases due to the coupling capacitance between the gate and the anode.
- the on / off state of the photoelectric conversion element PD62 can be arbitrarily set by the operation of the switching element SW62.
- the charge generated by the photoelectric conversion element PD61 (based on the combined light amount of the reflected light Lon and the external light L0) is turned on by the photoelectric conversion element PD61.
- the current is accumulated in the capacitor C 1 by the current path I 61 in the drawing.
- the discharge charge (based on the light quantity of the external light L0) generated in the photoelectric conversion element PD62 turns on the photoelectric conversion element PD62 and turns on the photoelectric conversion element PD61. Is released from the capacitor C1 by the current path I62 in the figure.
- the current flowing even when the photoelectric conversion elements PD61 and PD62 are in the off state is not 0, exactly, the current of the photoelectric conversion element in the on state and the photoelectric conversion element in the off state The difference causes charging or discharging.
- the imaging operation as described above makes it possible to stably detect an object regardless of the use condition while suppressing the manufacturing cost.
- FIG. 44 illustrates a circuit configuration of light receiving units (light receiving units 33A1 and 33A2) in a display area (display area (sensor area) 21C) according to the sixth modification.
- an imaging pixel (light receiving portion) 33A1 including a light shielding portion 341, 342 for blocking the reflected light Lon and the environmental light L0 from the photoelectric conversion elements PD21 and PD22.
- the image pickup pixel (light receiving unit) 33A2 does not include such a light shielding unit.
- the circuit configuration other than the light shielding portion in the light receiving portions 33A1 and 33A2 is the same as that of the light receiving portion 33A described in the first modification, and thus the description thereof is omitted.
- imaging pixels 33A1 and 33A2 in the display area (sensor area) 21C, for example, as shown in FIG. 45, the imaging pixels 33A1 and 33A2 and the imaging pixels 33A1 and 33A2 are driven for imaging
- the wiring portions 32 for this purpose are disposed separately from each other.
- Object information is to be acquired.
- an imaging pixel is configured by an imaging pixel (light receiving unit) 33A1 including light shielding units 341 and 342, and an imaging pixel (light receiving unit) 33A2 not including such a light shielding unit Since object information is acquired based on the difference signal between the imaging signal obtained by the pixel 33A1 and the imaging signal obtained by the imaging pixel 33A2, the influence of the voltage inclination or the like generated in the I / O display panel 20, the display It becomes possible to remove the noise component etc. which are mixed from driving of the system, and it becomes possible to improve the S / N ratio.
- the two types of light receiving sections 33A1 and 33A2 are described based on the circuit configuration of the light receiving section 33A according to the first modification, but such two types of light receiving sections are It may be based on the circuit configuration of the light receiving unit 33 of the above-described embodiment and the light receiving units 33B to 33F according to the above second to fifth modifications.
- FIG. 46A is an example in which the surface of the I / O display panel 20 is touched with a fingertip 61, and the locus of the touched portion is displayed on the screen as a drawing line 611.
- FIG. 46 (B) is for gesture recognition using a hand shape. Specifically, the shape of the hand 62 touching (or in proximity to) the I / O display panel 20 is recognized, the recognized hand shape is displayed as an image, and some processing is performed by moving the display object 621. It is intended to
- the hand 63A in the closed state is changed to the hand 63B in the open state, and the I / O display panel 20 recognizes the contact or proximity of the hand in each state.
- processing based on the image recognition For example, an instruction such as zooming in can be given.
- an instruction such as zooming in can be given.
- the operation of switching the command on the computer device by connecting the I / O display panel 20 to the personal computer device can be more naturally formed by these image recognitions. You can enter
- a plurality of I / O display panels 20 are prepared, and by connecting the plurality of I / O display panels by any transmission means, an image in which contact or proximity is detected. May be transmitted and displayed on the other party's I / O display panel to communicate between users who operate both display panels. That is, as shown in FIG. 48, two I / O display panels are prepared, and the handprinted hand-printed hand-printed one hand of the panel is sent to the other party to display the handprint 642 on the other panel. Then, processing such as sending the locus 641 displayed by touching the other panel with the hand 64 to the other panel for display can be performed.
- the state of drawing is transmitted as a moving image, and there is the possibility of a new communication tool by sending handwritten characters and figures to the other party.
- application of the I / O display panel 20 to a display panel of a mobile phone terminal is assumed.
- a brush 66 is used to touch the surface of the I / O display panel 20 to write characters, and the portion touched by the brush 66 is displayed on the I / O display panel 20.
- the image 661 By displaying as the image 661, handwriting input by a brush can be performed. In this case, it is possible to recognize and realize even a fine touch of the brush.
- a special pen tilt has been realized by electric field detection, but in this example, it is more realistic by detecting the contact surface of the real brush itself. Information can be input with
- the display-and-image-pickup apparatus includes a video signal input from the outside or a video signal generated internally such as a television apparatus, a digital camera, a laptop computer, a portable terminal such as a mobile phone or a video camera
- the present invention can be applied to electronic devices of all fields displayed as an image or a video.
- FIG. 50 shows the appearance of a television set to which the display-and-image-pickup device of the above-described embodiment and the like is applied.
- the television apparatus includes, for example, a video display screen unit 510 including a front panel 511 and a filter glass 512.
- the video display screen unit 510 is configured by the display and imaging device according to the above-described embodiment and the like. There is.
- FIG. 51 illustrates an appearance of a digital camera to which the display-and-image-pickup device of the above-described embodiment and the like is applied.
- the digital camera has, for example, a light emitting unit 521 for flash, a display unit 522, a menu switch 523 and a shutter button 524.
- the display unit 522 is configured by the display and imaging device according to the above-described embodiment and the like. ing.
- FIG. 52 illustrates the appearance of a notebook personal computer to which the display-and-image-pickup device of the above-described embodiment and the like is applied.
- the notebook personal computer includes, for example, a main body 531, a keyboard 532 for input operation of characters and the like, and a display unit 533 for displaying an image.
- the display unit 533 is a display according to the above-described embodiment and the like. It comprises an imaging device.
- FIG. 53 illustrates an appearance of a video camera to which the display-and-image-pickup device of the above-described embodiment and the like is applied.
- This video camera has, for example, a main body 541, a lens 542 for photographing an object provided on the front side of the main body 541, a start / stop switch 543 at the time of photographing, and a display 544.
- Reference numeral 544 is configured by the display and imaging device according to the above-described embodiment and the like.
- FIG. 54 illustrates an appearance of a mobile phone to which the display-and-image-pickup device of the above-described embodiment and the like is applied.
- This mobile phone is, for example, one in which an upper housing 710 and a lower housing 720 are connected by a connecting portion (hinge portion) 730, and has a display 740, a sub display 750, a picture light 760 and a camera 770. There is.
- the display 740 or the sub display 750 is configured by the display and imaging device according to the above-described embodiment and the like.
- the display backlight may also serve as the detection illumination, or A light may be provided.
- the light for example, infrared light
- the reset operation or the read operation for the light receiving portion (imaging pixel) for one line in one ON period or OFF period of the backlight 15 In the above case (when the high frequency backlight blinks) is performed, the case is not limited. That is, for example, during one ON period or OFF period of the backlight 15, the reset operation or the readout operation is performed on the light receiving units (imaging pixels) for a plurality of lines (a low frequency backlight blink operation is performed) You may do so.
- the I / O display panel 20 has been described as a liquid crystal element and separately providing a light receiving element as a display element.
- the I / O display panel may be configured by a light emitting / receiving element (display image pickup element) capable of performing time division of the light receiving operation and the light receiving operation. Note that the period in which the display light is not emitted in this case is a period in which the light emitting operation by the display imaging device is not performed.
- the present invention provides the display elements Alternatively, the present invention can be applied to an imaging apparatus having an imaging panel including a plurality of imaging elements.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Computer Hardware Design (AREA)
- Signal Processing (AREA)
- Multimedia (AREA)
- Control Of Indicators Other Than Cathode Ray Tubes (AREA)
- Liquid Crystal (AREA)
- Position Input By Displaying (AREA)
- Studio Devices (AREA)
- Liquid Crystal Display Device Control (AREA)
Abstract
Description
図15は、変形例1に係る受光部(受光部33A)の回路構成を表したものである。この受光部(撮像画素)33Aは、上記実施の形態で説明した受光部33とは異なり、2つの光電変換素子PD21,PD22と、2つのスイッチング素子SW21,SW22とを含んで構成されている。具体的には、光電変換素子PD21のカソードおよびゲートが電源VDDに接続され、アノードがスイッチング素子SW21の一端に接続されている。また、光電変換素子PD22のアノードが接地VSSに接続され、カソードおよびゲートがスイッチング素子SW22の一端に接続されている。また、スイッチング素子SW21,SW22の他端同士は、接続点P1に接続されている。また、この受光部33Aは、表示エリア(センサエリア)21Aにおいて、上記実施の形態と同様に例えば図16に示したように、撮像画素33Aと、この撮像画素33Aを撮像駆動するための配線部32とが、互いに分離配置されている。なお、他の回路構成は受光部33と同様であるため、説明を省略する。
図20は、変形例2に係る受光部(受光部33C)の回路構成を表したものである。この受光部(撮像画素)33Cは、2つの光電変換素子PD31,PD32と、4つのスイッチング素子SW31~SW34とを含んで構成されている。具体的には、スイッチング素子SW31,SW33の一端同士が電源VDDに接続され、スイッチング素子SW31の他端が光電変換素子PD31のカソードおよびゲートに接続され、スイッチング素子SW33の他端がコンデンサC1の一端およびスイッチング素子SW34の一端に接続されている。また、光電変換素子PD31のアノードが、コンデンサC1の他端、光電変換素子PD32のカソードおよびゲート、ならびに接続点P1に接続されている。また、光電変換素子PD32のアノードがスイッチング素子SW32の一端に接続されている。また、スイッチング素子SW32,SW34の他端同士は、接地VSSに接続されている。なお、他の回路構成は受光部33,33Aと同様であるため、説明を省略する。
図24は、変形例3に係る受光部(受光部33D)の回路構成を表したものである。この受光部(撮像画素)33Dは、2つの光電変換素子PD41(第1の光電変換素子),PD42(第2の光電変換素子)を含んで構成されている。具体的には、光電変換素子PD41は、カソードが電源VDDに接続され、アノードが光電変換素子PD42のカソードおよび接続点P1に接続され、ゲートが信号線φ1に接続されている。また、光電変換素子PD42は、アノードが接地VSSに接続され、ゲートが信号線φ2に接続されている。すなわち、この受光部33Dには、これまで説明した受光部33,33A~Cとは異なり、スイッチング素子が設けられていない一方、光電変換素子PD41,PD42のゲート電位が信号線φ1,φ2によって制御されるようになっている。なお、他の回路構成は受光部33等と同様であるため、説明を省略する。
図30は、変形例4に係る受光部(受光部33E)の回路構成を表したものである。この受光部(撮像画素)33Eは、2つの光電変換素子PD51(第1の光電変換素子),PD52(第2の光電変換素子)を含んで構成されている。具体的には、光電変換素子PD51は、カソードが信号線φ1に接続され、アノードが光電変換素子PD52のカソードおよび接続点P1に接続され、ゲートが電源VDDに接続されている。また、光電変換素子PD52は、アノードが信号線φ2に接続され、ゲートが接地VSSに接続されている。すなわち、この受光部33Eでは、上記変形例3の受光部33Dとは異なり、光電変換素子PD51,PD52のゲート電位が電源VDDまたは接地VSSによって固定される一方、光電変換素子PD51のカソード電位および光電変換素子PD52のアノード電位が信号線φ1,φ2によって制御されるようになっている。なお、他の回路構成は受光部33Dと同様であるため、説明を省略する。
図36は、変形例5に係る受光部(受光部33F)の回路構成を表したものである。この受光部(撮像画素)33Fは、2つの光電変換素子PD61(第1の光電変換素子),PD62(第2の光電変換素子)と、5つのトランジスタTr1~Tr3,Tr61(Nチャネルトランジスタ),Tr62(Pチャネルトランジスタ)と、4つの容量素子C61A,C61B,C62A,C62Bとを含んで構成されている。具体的には、光電変換素子PD61は、カソードが信号線φ1およびトランジスタTr61のソースに接続され、アノードが光電変換素子PD62のカソードおよび接続点P1に接続され、ゲートがトランジスタTr61のドレインおよび容量素子C61A,C61Bの一端に接続されている。また、トランジスタTr61のゲートは容量素子C61Aの他端およびリセット線Reset1(これまで説明したリセット線Resetに対応)に接続されている。一方、光電変換素子PD62は、アノードが信号線φ2およびトランジスタTr62のソースに接続され、ゲートがトランジスタTr62のドレインおよび容量素子C62A,C62Bの一端に接続されている。また、トランジスタTr62のゲートは容量素子C61Aの他端およびリセット線Reset2の反転信号に接続されている。なお、容量素子C61B,C62Bの他端同士は電源VDDに接続されている。すなわち、この受光部33Fでは、上記変形例4の受光部33Eとは異なり、光電変換素子PD61,PD62のゲート電位が固定されず、プリチャージ期間後から読み出し期間までの間(充電電荷の蓄積時および放電電荷の放出時において)、ハイインピーダンス状態となっている。なお、他の回路構成は受光部33Eと同様であるため、説明を省略する。
図44は、変形例6に係る表示エリア(表示エリア(センサエリア)21C)内の受光部(受光部33A1,33A2)の回路構成を表したものである。この表示エリア21Cでは、撮像画素(受光部)が、反射光Lonおよび環境光L0を光電変換素子PD21,PD22に対して遮断するための遮光部341,342を含む撮像画素(受光部)33A1と、そのような遮光部を含まない撮像画素(受光部)33A2とにより構成されている。なお、受光部33A1,33A2における遮光部以外の回路構成は、上記変形例1で説明した受光部33Aと同様であるので、説明を省略する。
次に、図46~図49を参照して、これまで説明した指先抽出処理によって検出された物体の位置情報等を利用した、アプリケーションプログラム実行部11によるアプリケーションプログラム実行例について、いくつか説明する。
次に、図50~図54を参照して、上記実施の形態および変形例で説明した表示撮像装置の適用例について説明する。上記実施の形態等の表示撮像装置は、テレビジョン装置,デジタルカメラ,ノート型パーソナルコンピュータ、携帯電話等の携帯端末装置あるいはビデオカメラなど、外部から入力された映像信号あるいは内部で生成した映像信号を、画像あるいは映像として表示するあらゆる分野の電子機器に適用することが可能である。
図50は、上記実施の形態等の表示撮像装置が適用されるテレビジョン装置の外観を表したものである。このテレビジョン装置は、例えば、フロントパネル511およびフィルターガラス512を含む映像表示画面部510を有しており、この映像表示画面部510は、上記実施の形態等に係る表示撮像装置により構成されている。
図51は、上記実施の形態等の表示撮像装置が適用されるデジタルカメラの外観を表したものである。このデジタルカメラは、例えば、フラッシュ用の発光部521、表示部522、メニュースイッチ523およびシャッターボタン524を有しており、その表示部522は、上記実施の形態等に係る表示撮像装置により構成されている。
図52は、上記実施の形態等の表示撮像装置が適用されるノート型パーソナルコンピュータの外観を表したものである。このノート型パーソナルコンピュータは、例えば、本体531,文字等の入力操作のためのキーボード532および画像を表示する表示部533を有しており、その表示部533は、上記実施の形態等に係る表示撮像装置により構成されている。
図53は、上記実施の形態等の表示撮像装置が適用されるビデオカメラの外観を表したものである。このビデオカメラは、例えば、本体部541,この本体部541の前方側面に設けられた被写体撮影用のレンズ542,撮影時のスタート/ストップスイッチ543および表示部544を有しており、その表示部544は、上記実施の形態等に係る表示撮像装置により構成されている。
図54は、上記実施の形態等の表示撮像装置が適用される携帯電話機の外観を表したものである。この携帯電話機は、例えば、上側筐体710と下側筐体720とを連結部(ヒンジ部)730で連結したものであり、ディスプレイ740,サブディスプレイ750,ピクチャーライト760およびカメラ770を有している。そのディスプレイ740またはサブディスプレイ750は、上記実施の形態等に係る表示撮像装置により構成されている。
Claims (28)
- 複数の撮像画素により構成され、近接物体に対する照射光源を有する撮像パネルと、
各撮像画素に対して撮像駆動を行うことにより、各撮像画素から撮像信号を取得する撮像駆動手段と、
各撮像画素から得られる撮像信号に基づく撮像画像を用いて、前記近接物体の位置、形状または大きさの少なくとも1つを含む物体情報を取得する画像処理手段と
を備え、
前記撮像駆動手段は、前記照射光源から照射光が照射されているときに、この照射光による反射光と環境光との合算光量に応じて各撮像画素に充電電荷が蓄積されると共に、前記照射光源から照射光が照射されていないときに、環境光の光量に応じて各撮像画素から放電電荷が放出されるように撮像駆動を行うことにより、各撮像画素から撮像信号を取得する
撮像装置。 - 各撮像画素が、
受光光量に応じた電荷を発生させる光電変換素子と、
容量素子と、
前記充電電荷を前記容量素子へ蓄積させるための充電回路と、
前記放電電荷を前記容量素子から放出させるための放電回路と
を含んで構成され、
前記撮像駆動手段は、前記充電回路および前記放電回路に対して撮像駆動を行うことにより、各撮像画素から撮像信号を取得する
請求項1に記載の撮像装置。 - 前記光電変換素子が、前記充電電荷および前記放電電荷をそれぞれ共通して発生させるものであり、
前記充電回路が充電用スイッチング素子を含むと共に、前記放電回路が放電用スイッチング素子を含み、
前記撮像駆動手段は、前記光電変換素子で発生した充電電荷が前記充電用スイッチング素子を介して前記容量素子へ蓄積されると共に、前記光電変換素子で発生した放電電荷が前記放電用スイッチング素子を介して前記容量素子から放出されるように、前記充電用スイッチング素子および前記放電用スイッチング素子に対して撮像駆動を行う
請求項2に記載の撮像装置。 - 前記撮像駆動手段は、前記充電用スイッチング素子のオン状態期間と前記放電用スイッチング素子のオン状態期間とが互いに重ならないように、撮像駆動を行う
請求項3に記載の撮像装置。 - 前記光電変換素子が、前記充電電荷を発生させる第1の光電変換素子と、前記放電電荷を発生させる第2の光電変換素子とにより構成され、
前記充電回路が充電用スイッチング素子を含むと共に、前記放電回路が放電用スイッチング素子を含み、
前記撮像駆動手段は、前記第1の光電変換素子で発生した充電電荷が前記充電用スイッチング素子を介して前記容量素子へ蓄積されると共に、前記第2の光電変換素子で発生した放電電荷が前記放電用スイッチング素子を介して前記容量素子から放出されるように、前記充電用スイッチング素子および前記放電用スイッチング素子に対して撮像駆動を行う
請求項2に記載の撮像装置。 - 前記撮像駆動手段は、前記容量素子の一端側が、前記充電電荷の蓄積時と前記放電電荷の放出時とにおいて、電源または接地に交互に接続されるように、前記充電用スイッチング素子および前記放電用スイッチング素子に対して撮像駆動を行う
請求項5に記載の撮像装置。 - 前記撮像駆動手段は、前記充電用スイッチング素子のオン状態期間と前記放電用スイッチング素子のオン状態期間とが互いに重ならないように、撮像駆動を行う
請求項6に記載の撮像装置。 - 前記光電変換素子が、前記充電電荷を発生させる第1の光電変換素子と、前記放電電荷を発生させる第2の光電変換素子とにより構成され、
前記第1および第2の光電変換素子がそれぞれ、アノード電極、カソード電極およびゲート電極を有するPIN型のフォトダイオードにより構成され、
前記撮像駆動手段は、
前記第1の光電変換素子におけるカソード電極とゲート電極との間の電位関係と、前記第2の光電変換素子におけるアノード電極とゲート電極との間の電位関係とをそれぞれ変化させることにより、前記第1および第2の光電変換素子を個別にオン・オフ状態に設定すると共に、
前記第1の光電変換素子で発生した充電電荷が、この第1の光電変換素子がオン状態となると共に前記第2の光電変換素子がオフ状態になることにより前記容量素子へ蓄積される一方、前記第2の光電変換素子で発生した放電電荷が、この第2の光電変換素子がオン状態となると共に前記第1の光電変換素子がオフ状態になることにより前記容量素子から放出されるように、前記第1および第2の光電変換素子に対して撮像駆動を行う
請求項2に記載の撮像装置。 - 前記撮像駆動手段は、
前記第1の光電変換素子におけるカソード電極の電位および前記第2の光電変換素子におけるアノード電極の電位をそれぞれ個別に固定すると共に、
前記第1および第2の光電変換素子におけるゲート電極の電位をそれぞれ個別に変化させることにより、前記第1および第2の光電変換素子を個別にオン・オフ状態に設定する
請求項8に記載の撮像装置。 - 前記撮像駆動手段は、前記第1の光電変換素子におけるカソード電極の電位および前記第2の光電変換素子におけるアノード電極の電位をそれぞれ個別に変化させることにより、前記第1および第2の光電変換素子を個別にオン・オフ状態に設定する
請求項8に記載の撮像装置。 - 前記撮像駆動手段は、前記第1および第2の光電変換素子におけるゲート電極の電位をそれぞれ個別に固定する
請求項10に記載の撮像装置。 - 前記撮像駆動手段は、前記第1および第2の光電変換素子におけるゲート電極の電位をそれぞれ、前記充電電荷の蓄積時および前記放電電荷の放出時においてハイインピーダンス状態に設定する
請求項10に記載の撮像装置。 - 前記撮像画素が、
前記反射光および前記環境光を遮断する遮光部を含む第1の撮像画素と、
前記遮光部を含まない第2の撮像画素と
により構成され、
前記画像処理手段は、前記第1の撮像画素により得られる撮像信号と前記第2の撮像画素により得られる撮像信号との差分信号に基づいて、前記物体情報を取得する
請求項1ないし請求項12のいずれか1項に記載の撮像装置。 - 前記画像処理手段は、前記充電電荷の蓄積動作と前記放電電荷の放出動作とが複数回切り換えられた後に得られた撮像信号に基づいて、前記物体情報を取得する
請求項1ないし請求項12のいずれか1項に記載の撮像装置。 - 各撮像画素に蓄積された電荷を全て放出させるための初期化電圧が、電源電圧と接地電圧との間の任意の電圧である
請求項1ないし請求項12のいずれか1項に記載の撮像装置。 - 前記撮像パネルにおいて各撮像画素がマトリクス状に配置され、
前記撮像駆動手段は、各撮像画素に対して線順次撮像駆動を行う
請求項1ないし請求項12のいずれか1項に記載の撮像装置。 - 前記撮像パネルにおいて、前記撮像画素と、この撮像画素を撮像駆動するための配線とが、互いに分離配置されている
請求項1ないし請求項12のいずれか1項に記載の撮像装置。 - 前記撮像駆動手段は、前記近接物体が前記撮像パネル上で動いているときに、前記充電電荷が蓄積されると共に前記放電電荷が放出されるように撮像駆動を行うことにより、各撮像画素から撮像信号を取得する
請求項1ないし請求項12のいずれか1項に記載の撮像装置。 - 前記画像処理手段は、前記撮像パネル上に同時に位置する複数の近接物体について、前記物体情報をそれぞれ取得する
請求項1ないし請求項12のいずれか1項に記載の撮像装置。 - 複数の表示画素および複数の撮像画素により構成された表示撮像パネルと、
各表示画素に対して表示駆動を行うことにより、画像表示を行う表示駆動手段と、
各撮像画素に対して撮像駆動を行うことにより、各撮像画素から撮像信号を取得する撮像駆動手段と、
各撮像画素から得られる撮像信号に基づく撮像画像を用いて、近接物体の位置、形状または大きさの少なくとも1つを含む物体情報を取得する画像処理手段と
を備え、
前記撮像駆動手段は、前記近接物体に対して前記表示撮像パネルから照射光が照射されているときに、この照射光による反射光と環境光との合算光量に応じて各撮像画素に充電電荷が蓄積されると共に、前記表示撮像パネルから照射光が照射されていないときに、環境光の光量に応じて各撮像画素から放電電荷が放出されるように撮像駆動を行うことにより、各撮像画素から撮像信号を取得する
表示撮像装置。 - 前記充電電荷の蓄積動作および前記放電電荷の放出動作がそれぞれ、前記表示駆動の際の水平期間と同期している
請求項20に記載の表示撮像装置。 - 前記画像処理手段は、前記充電電荷の蓄積動作と前記放電電荷の放出動作とが複数回切り換えられた後に得られた撮像信号に基づいて、前記物体情報を取得する
請求項21に記載の表示撮像装置。 - 前記表示駆動手段は、前記画像処理手段により得られた物体情報に基づいて表示駆動を行うことにより、この物体情報を前記表示撮像パネルに表示させる
請求項20に記載の表示撮像装置。 - 前記表示画素が、液晶素子を含んで構成されている
請求項20に記載の表示撮像装置。 - 前記表示画素および前記撮像画素が、発光動作および受光動作を時分割に行うことが可能な複数の表示撮像素子を含んで構成されている
請求項20に記載の表示撮像装置。 - 前記表示撮像素子が、有機EL素子である
請求項25に記載の表示撮像装置。 - 画像表示機能および撮像機能を有する表示撮像装置を備え、
前記表示撮像装置は、
複数の表示画素および複数の撮像画素により構成された表示撮像パネルと、
各表示画素に対して表示駆動を行うことにより、画像表示を行う表示駆動手段と、
各撮像画素に対して撮像駆動を行うことにより、各撮像画素から撮像信号を取得する撮像駆動手段と、
各撮像画素から得られる撮像信号に基づく撮像画像を用いて、近接物体の位置、形状または大きさの少なくとも1つを含む物体情報を取得する画像処理手段と
を有し、
前記撮像駆動手段は、前記近接物体に対して前記表示撮像パネルから照射光が照射されているときに、この照射光による反射光と環境光との合算光量に応じて各撮像画素に充電電荷が蓄積されると共に、前記表示撮像パネルから照射光が照射されていないときに、環境光の光量に応じて各撮像画素から放電電荷が放出されるように撮像駆動を行うことにより、各撮像画素から撮像信号を取得する
電子機器。 - 複数の撮像画素により構成された撮像パネルによって近接物体に対して照射光を照射させているときに、この照射光による反射光と環境光との合算光量に応じて各撮像画素に充電電荷を蓄積させると共に、前記照射光を照射させていないときに、環境光の光量に応じて各撮像画素から放電電荷を放出させるようにして各撮像画素に対して撮像駆動を行うことによって、各撮像画素から撮像信号を取得し、
各撮像画素から得られる撮像信号に基づく撮像画像を用いて、前記近接物体の位置、形状または大きさの少なくとも1つを含む物体情報を取得する
物体の検出方法。
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020097024306A KR101547703B1 (ko) | 2008-03-24 | 2009-03-18 | 촬상 장치, 표시 촬상 장치, 전자기기 및 물체의 검출 방법 |
US12/451,485 US9063613B2 (en) | 2008-03-24 | 2009-03-18 | Image pickup device, display-and-image-pickup device, electronic apparatus and method of detecting an object |
CN2009800004743A CN101689083B (zh) | 2008-03-24 | 2009-03-18 | 成像设备、显示与成像设备、检测对象的电子设备和方法 |
JP2010505579A JP5291090B2 (ja) | 2008-03-24 | 2009-03-18 | 撮像装置、表示撮像装置、および電子機器 |
EP09724453.7A EP2151740B1 (en) | 2008-03-24 | 2009-03-18 | Displaying and imaging apparatus, electronic apparatus and method for detecting an object |
US14/705,664 US9405406B2 (en) | 2008-03-24 | 2015-05-06 | Image pickup device, display-and-image-pickup device, electronic apparatus and method of detecting an object |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2008076281 | 2008-03-24 | ||
JP2008-076281 | 2008-03-24 |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/451,485 A-371-Of-International US9063613B2 (en) | 2008-03-24 | 2009-03-18 | Image pickup device, display-and-image-pickup device, electronic apparatus and method of detecting an object |
US14/705,664 Continuation US9405406B2 (en) | 2008-03-24 | 2015-05-06 | Image pickup device, display-and-image-pickup device, electronic apparatus and method of detecting an object |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2009119417A1 true WO2009119417A1 (ja) | 2009-10-01 |
Family
ID=41113614
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2009/055328 WO2009119417A1 (ja) | 2008-03-24 | 2009-03-18 | 撮像装置、表示撮像装置、電子機器および物体の検出方法 |
Country Status (7)
Country | Link |
---|---|
US (2) | US9063613B2 (ja) |
EP (1) | EP2151740B1 (ja) |
JP (1) | JP5291090B2 (ja) |
KR (1) | KR101547703B1 (ja) |
CN (1) | CN101689083B (ja) |
TW (1) | TWI423086B (ja) |
WO (1) | WO2009119417A1 (ja) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102063224A (zh) * | 2009-11-13 | 2011-05-18 | 索尼公司 | 传感器设备、传感器元件驱动方法,显示设备及电子装置 |
JP2011113384A (ja) * | 2009-11-27 | 2011-06-09 | Sony Corp | センサ装置、センサ素子の駆動方法、入力機能付き表示装置および電子機器 |
JP2011135561A (ja) * | 2009-11-27 | 2011-07-07 | Sony Corp | センサ装置、センサ素子の駆動方法、入力機能付き表示装置および電子機器、ならびに放射線撮像装置 |
JP2012008541A (ja) * | 2010-05-28 | 2012-01-12 | Semiconductor Energy Lab Co Ltd | 光検出装置 |
WO2012014864A1 (ja) * | 2010-07-27 | 2012-02-02 | シャープ株式会社 | 表示装置 |
JP2013009388A (ja) * | 2010-03-08 | 2013-01-10 | Semiconductor Energy Lab Co Ltd | 半導体装置、表示装置 |
Families Citing this family (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5291090B2 (ja) * | 2008-03-24 | 2013-09-18 | 株式会社ジャパンディスプレイウェスト | 撮像装置、表示撮像装置、および電子機器 |
KR101319346B1 (ko) * | 2009-09-15 | 2013-10-16 | 엘지디스플레이 주식회사 | 광센싱 방식의 터치 패널 내장형 액정 표시 장치 및 이의 구동 방법 |
RU2012117771A (ru) | 2009-09-30 | 2013-11-10 | Шарп Кабусики Кайся | Устройство отображения |
JP5349607B2 (ja) * | 2009-09-30 | 2013-11-20 | シャープ株式会社 | 表示装置 |
WO2011145676A1 (ja) * | 2010-05-20 | 2011-11-24 | シャープ株式会社 | タッチセンサ付き表示装置 |
JP5516145B2 (ja) * | 2010-06-30 | 2014-06-11 | セイコーエプソン株式会社 | 光学式検出装置、表示装置及び電子機器 |
US8605059B2 (en) * | 2010-07-02 | 2013-12-10 | Semiconductor Energy Laboratory Co., Ltd. | Input/output device and driving method thereof |
US20130113768A1 (en) * | 2010-07-27 | 2013-05-09 | Sharp Kabushiki Kaisha | Display device and drive method for same |
TWI416387B (zh) * | 2010-08-24 | 2013-11-21 | Au Optronics Corp | 觸控面板 |
US20150022754A1 (en) * | 2013-07-19 | 2015-01-22 | Google Inc. | Configurations for tileable display apparatus with multiple pixel arrays |
JP6247530B2 (ja) * | 2013-12-27 | 2017-12-13 | キヤノン株式会社 | 撮像装置 |
US9778713B2 (en) * | 2015-01-05 | 2017-10-03 | Synaptics Incorporated | Modulating a reference voltage to preform capacitive sensing |
TWI660165B (zh) * | 2018-05-14 | 2019-05-21 | 和碩聯合科技股份有限公司 | 檢測系統及檢測方法 |
CN114187870B (zh) * | 2020-09-14 | 2023-05-09 | 京东方科技集团股份有限公司 | 光电检测电路及其驱动方法、显示装置及其制作方法 |
TWI753634B (zh) * | 2020-10-30 | 2022-01-21 | 胡厚飛 | 毛筆分類之測定方法 |
CN112530375B (zh) * | 2020-12-02 | 2022-01-11 | 福州京东方光电科技有限公司 | 一种显示装置的驱动方法、显示装置及存储介质 |
JP2022190312A (ja) * | 2021-06-14 | 2022-12-26 | セイコーエプソン株式会社 | 表示装置及び表示装置の制御方法 |
CN115394244B (zh) * | 2022-10-28 | 2023-01-24 | 惠科股份有限公司 | 显示驱动电路、显示驱动方法和显示装置 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004127272A (ja) | 2002-09-10 | 2004-04-22 | Sony Corp | 情報処理装置および方法、記録媒体、並びにプログラム |
JP2006276223A (ja) | 2005-03-28 | 2006-10-12 | Sony Corp | 表示装置及び表示方法 |
JP2007011228A (ja) * | 2005-07-04 | 2007-01-18 | Toshiba Matsushita Display Technology Co Ltd | 平面表示装置 |
JP2007065508A (ja) * | 2005-09-01 | 2007-03-15 | Sony Corp | 表示装置及び表示方法 |
JP2007206192A (ja) * | 2006-01-31 | 2007-08-16 | Toshiba Matsushita Display Technology Co Ltd | 表示装置 |
Family Cites Families (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3275579B2 (ja) * | 1994-10-24 | 2002-04-15 | 日産自動車株式会社 | イメージセンサ |
JP3521187B2 (ja) * | 1996-10-18 | 2004-04-19 | 株式会社東芝 | 固体撮像装置 |
JP3471201B2 (ja) * | 1997-09-05 | 2003-12-02 | 株式会社東芝 | 撮像モジュール、および撮像装置 |
US6798446B2 (en) * | 2001-07-09 | 2004-09-28 | Logitech Europe S.A. | Method and system for custom closed-loop calibration of a digital camera |
JP3800052B2 (ja) * | 2001-08-20 | 2006-07-19 | ソニー株式会社 | 静止画像撮像装置及び方法 |
US7053967B2 (en) | 2002-05-23 | 2006-05-30 | Planar Systems, Inc. | Light sensitive display |
US7233354B2 (en) * | 2002-10-11 | 2007-06-19 | Hewlett-Packard Development Company, L.P. | Digital camera that adjusts resolution for low light conditions |
JP4323772B2 (ja) * | 2002-10-31 | 2009-09-02 | キヤノン株式会社 | 固体撮像装置、カメラ及びカメラ制御システム |
JP4489608B2 (ja) * | 2004-03-31 | 2010-06-23 | 富士フイルム株式会社 | ディジタル・スチル・カメラ,画像再生装置および顔画像表示装置ならびにそれらの制御方法 |
JP2005303081A (ja) * | 2004-04-13 | 2005-10-27 | Matsushita Electric Ind Co Ltd | 光センサーおよび固体撮像装置 |
US7593593B2 (en) | 2004-06-16 | 2009-09-22 | Microsoft Corporation | Method and system for reducing effects of undesired signals in an infrared imaging system |
JP4072732B2 (ja) * | 2004-10-29 | 2008-04-09 | ソニー株式会社 | 入出力装置および方法、記録媒体、並びにプログラム |
JP4618026B2 (ja) * | 2005-07-12 | 2011-01-26 | ソニー株式会社 | 表示装置及び表示および受光を行う表示装置を用いた検出方法 |
TW200720829A (en) * | 2005-11-25 | 2007-06-01 | Altek Corp | Photographic apparatus of sensible environment parameters |
TWI280409B (en) * | 2006-04-14 | 2007-05-01 | Asustek Comp Inc | Reflective photo device, an electronic apparatus with a built-in camera using the device for providing colorimeter and ambient light sensor functions and its method |
US8610670B2 (en) * | 2006-08-31 | 2013-12-17 | Japan Display West Inc. | Imaging and display apparatus, information input apparatus, object detection medium, and object detection method |
US8674949B2 (en) * | 2007-02-20 | 2014-03-18 | Japan Displays Inc. | Liquid crystal display apparatus |
JP5291090B2 (ja) * | 2008-03-24 | 2013-09-18 | 株式会社ジャパンディスプレイウェスト | 撮像装置、表示撮像装置、および電子機器 |
WO2012014864A1 (ja) * | 2010-07-27 | 2012-02-02 | シャープ株式会社 | 表示装置 |
-
2009
- 2009-03-18 JP JP2010505579A patent/JP5291090B2/ja not_active Expired - Fee Related
- 2009-03-18 WO PCT/JP2009/055328 patent/WO2009119417A1/ja active Application Filing
- 2009-03-18 KR KR1020097024306A patent/KR101547703B1/ko active IP Right Grant
- 2009-03-18 CN CN2009800004743A patent/CN101689083B/zh not_active Expired - Fee Related
- 2009-03-18 EP EP09724453.7A patent/EP2151740B1/en active Active
- 2009-03-18 US US12/451,485 patent/US9063613B2/en active Active
- 2009-03-24 TW TW098109596A patent/TWI423086B/zh not_active IP Right Cessation
-
2015
- 2015-05-06 US US14/705,664 patent/US9405406B2/en not_active Expired - Fee Related
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004127272A (ja) | 2002-09-10 | 2004-04-22 | Sony Corp | 情報処理装置および方法、記録媒体、並びにプログラム |
JP2006276223A (ja) | 2005-03-28 | 2006-10-12 | Sony Corp | 表示装置及び表示方法 |
JP2007011228A (ja) * | 2005-07-04 | 2007-01-18 | Toshiba Matsushita Display Technology Co Ltd | 平面表示装置 |
JP2007065508A (ja) * | 2005-09-01 | 2007-03-15 | Sony Corp | 表示装置及び表示方法 |
JP2007206192A (ja) * | 2006-01-31 | 2007-08-16 | Toshiba Matsushita Display Technology Co Ltd | 表示装置 |
Non-Patent Citations (1)
Title |
---|
See also references of EP2151740A4 |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8593442B2 (en) | 2009-11-13 | 2013-11-26 | Japan Display West Inc. | Sensor device, method of driving sensor element, display device with input function and electronic apparatus |
CN102063224A (zh) * | 2009-11-13 | 2011-05-18 | 索尼公司 | 传感器设备、传感器元件驱动方法,显示设备及电子装置 |
JP2011113384A (ja) * | 2009-11-27 | 2011-06-09 | Sony Corp | センサ装置、センサ素子の駆動方法、入力機能付き表示装置および電子機器 |
JP2011135561A (ja) * | 2009-11-27 | 2011-07-07 | Sony Corp | センサ装置、センサ素子の駆動方法、入力機能付き表示装置および電子機器、ならびに放射線撮像装置 |
US9250743B2 (en) | 2009-11-27 | 2016-02-02 | Japan Display Inc. | Sensor device, method of driving sensor element, display device with input function, electronic unit and radiation image pickup device |
US8665243B2 (en) | 2009-11-27 | 2014-03-04 | Japan Display West Inc. | Sensor device, method of driving sensor element, display device with input function and electronic unit |
US11710751B2 (en) | 2010-03-08 | 2023-07-25 | Semiconductor Energy Laboratory Co., Ltd. | Semiconductor device and manufacturing method thereof |
JP2013009388A (ja) * | 2010-03-08 | 2013-01-10 | Semiconductor Energy Lab Co Ltd | 半導体装置、表示装置 |
US9153619B2 (en) | 2010-03-08 | 2015-10-06 | Semiconductor Energy Laboratory Co., Ltd. | Semiconductor device and manufacturing method thereof |
US10535691B2 (en) | 2010-03-08 | 2020-01-14 | Semiconductor Energy Laboratory Co., Ltd. | Semiconductor device and manufacturing method thereof |
US11139327B2 (en) | 2010-03-08 | 2021-10-05 | Semiconductor Energy Laboratory Co., Ltd. | Semiconductor device and manufacturing method thereof |
US9257567B2 (en) | 2010-03-08 | 2016-02-09 | Semiconductor Energy Laboratory Co., Ltd. | Semiconductor device and manufacturing method thereof |
US9515107B2 (en) | 2010-03-08 | 2016-12-06 | Semiconductor Energy Laboratory Co., Ltd. | Semiconductor device and manufacturing method thereof |
JP2016192222A (ja) * | 2010-05-28 | 2016-11-10 | 株式会社半導体エネルギー研究所 | 光検出装置、タッチパネル |
US9846515B2 (en) | 2010-05-28 | 2017-12-19 | Semiconductor Energy Laboratory Co., Ltd. | Photodetector and display device with light guide configured to face photodetector circuit and reflect light from a source |
JP2015181022A (ja) * | 2010-05-28 | 2015-10-15 | 株式会社半導体エネルギー研究所 | 光検出装置、及び、タッチパネル |
JP2012008541A (ja) * | 2010-05-28 | 2012-01-12 | Semiconductor Energy Lab Co Ltd | 光検出装置 |
WO2012014864A1 (ja) * | 2010-07-27 | 2012-02-02 | シャープ株式会社 | 表示装置 |
Also Published As
Publication number | Publication date |
---|---|
EP2151740A4 (en) | 2015-05-06 |
TW201007527A (en) | 2010-02-16 |
KR20110001853A (ko) | 2011-01-06 |
CN101689083A (zh) | 2010-03-31 |
US20100128004A1 (en) | 2010-05-27 |
CN101689083B (zh) | 2013-01-02 |
EP2151740A1 (en) | 2010-02-10 |
EP2151740B1 (en) | 2017-05-10 |
US20150301689A1 (en) | 2015-10-22 |
JPWO2009119417A1 (ja) | 2011-07-21 |
JP5291090B2 (ja) | 2013-09-18 |
US9063613B2 (en) | 2015-06-23 |
TWI423086B (zh) | 2014-01-11 |
US9405406B2 (en) | 2016-08-02 |
KR101547703B1 (ko) | 2015-08-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2009119417A1 (ja) | 撮像装置、表示撮像装置、電子機器および物体の検出方法 | |
TWI439904B (zh) | 感測器裝置、驅動感測器元件之方法、具有輸入功能之顯示裝置及電子單元 | |
JP5481127B2 (ja) | センサ素子およびその駆動方法、センサ装置、ならびに入力機能付き表示装置および電子機器 | |
JP5510998B2 (ja) | センサ装置、センサ素子の駆動方法、入力機能付き表示装置および電子機器 | |
US8188987B2 (en) | Display and imaging apparatus and object detecting method | |
KR101534588B1 (ko) | 표시 촬상장치, 및 물체의 검출 방법 | |
JP5246795B2 (ja) | センサ装置、センサ素子の駆動方法、入力機能付き表示装置および電子機器 | |
US8514201B2 (en) | Image pickup device, display-and-image pickup device, and electronic device | |
JP5093523B2 (ja) | 撮像装置、表示撮像装置および撮像処理装置 | |
JP4270247B2 (ja) | 表示撮像装置、物体検出プログラムおよび物体の検出方法 | |
JP2009069159A (ja) | 表示撮像装置、物体検出プログラムおよび物体の検出方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 200980000474.3 Country of ref document: CN |
|
WWE | Wipo information: entry into national phase |
Ref document number: 12451485 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2010505579 Country of ref document: JP |
|
ENP | Entry into the national phase |
Ref document number: 20097024306 Country of ref document: KR Kind code of ref document: A |
|
REEP | Request for entry into the european phase |
Ref document number: 2009724453 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2009724453 Country of ref document: EP |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 09724453 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |