US20060146175A1 - Image capture module and image capture apparatus for inputting shape of object on three dimensional space - Google Patents

Image capture module and image capture apparatus for inputting shape of object on three dimensional space Download PDF

Info

Publication number
US20060146175A1
US20060146175A1 US11/366,494 US36649406A US2006146175A1 US 20060146175 A1 US20060146175 A1 US 20060146175A1 US 36649406 A US36649406 A US 36649406A US 2006146175 A1 US2006146175 A1 US 2006146175A1
Authority
US
United States
Prior art keywords
light
image
image capture
object
light sources
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/366,494
Inventor
Shunichi Numazaki
Takahiro Murata
Miwako Doi
Akira Morishita
Naoko Umeki
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Toshiba Corp
Original Assignee
Toshiba Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to JP24069297A priority Critical patent/JP3471201B2/en
Priority to JP9-240692 priority
Priority to US09/146,957 priority patent/US6628335B1/en
Priority to US10/612,957 priority patent/US7042490B2/en
Application filed by Toshiba Corp filed Critical Toshiba Corp
Priority to US11/366,494 priority patent/US20060146175A1/en
Publication of US20060146175A1 publication Critical patent/US20060146175A1/en
Application status is Abandoned legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment ; Cameras comprising an electronic image sensor, e.g. digital cameras, video cameras, TV cameras, video cameras, camcorders, webcams, camera modules for embedding in other devices, e.g. mobile phones, computers or vehicles
    • H04N5/225Television cameras ; Cameras comprising an electronic image sensor, e.g. digital cameras, video cameras, camcorders, webcams, camera modules specially adapted for being embedded in other devices, e.g. mobile phones, computers or vehicles
    • H04N5/2256Television cameras ; Cameras comprising an electronic image sensor, e.g. digital cameras, video cameras, camcorders, webcams, camera modules specially adapted for being embedded in other devices, e.g. mobile phones, computers or vehicles provided with illuminating means
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/0304Detection arrangements using opto-electronic means
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/042Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/30Transforming light or analogous information into electric information
    • H04N5/335Transforming light or analogous information into electric information using solid-state image sensors [SSIS]
    • H04N5/369SSIS architecture; Circuitry associated therewith
    • H04N5/374Addressed sensors, e.g. MOS or CMOS sensors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N2013/0074Stereoscopic image analysis
    • H04N2013/0081Depth or disparity estimation from stereoscopic image signals

Abstract

An image capture module and an image capture apparatus for inputting shape of an object in three dimensional space, including a light emitting section which emits light to the object for reflection by the object and imaging of reflected light on a photo-receiving plane of a reflected light image detecting section via an image capture optical system. The reflected light image detecting section detects a reflected light image indicative of a distribution of intensity of the reflected light. The light may be irradiated evenly to the object and the shape of the object may be detected correctly even if the object has irregularity by disposing the light emitting section closely to the image capture optical system. The object may be illuminated evenly by the light from a light emitting section having plural light emitting elements by disposing the light emitting elements symmetrically about the image capture optical system. The object may be illuminated more evenly by the light from the light emitting elements by using reflecting elements which surround respective light emitting elements and focus light toward the object.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a Divisional of co-pending U.S. patent application Ser. No. 10/612,957, filed Jul. 7, 2003, which is a Divisional of U.S. application Ser. No. 09/146,957, filed Sep. 4, 1998, now U.S. Pat. No. 6,628,335, and claims priority to Japanese Patent Application No. 9-240692, filed Sep. 5, 1997. The entire contents of these applications are incorporated herein by reference in their entirety.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention relates to an image capture module and an image capture apparatus including light emitting sections to input a shape of an object on a three-dimensional space.
  • 2. Discussion of the background
  • Although a mouse is usually used as an input device of a computer, the mouse is a two-dimensional pointing device for moving a cursor, for selecting a menu and the like. The mouse operates on two-dimensional information so that it is difficult to make a selection on an object having a depth such as an object within a three-dimensional space. It has been also difficult to add mouse driven natural motion to characters in creating animation.
  • In order to solve the above-mentioned problems, the present inventors have invented a device that allows a shape of an object, motion, distance information and the like to be inputted in non-contact mode of operation. This device obtains an image of reflected light by operating in synchronism with light emitting sections an image capture sensor having two capacitors in a photo-receiving cell that corresponds to one pixel. Electric charge generated by a photoelectric converting section of the photo-receiving cell by receiving the light is accumulated selectively in either one of the two capacitors. Then, electric charge generated by receiving light while the light emitting section emits light is accumulated in the first capacitor and electric charge generated by receiving light while the light emitting section does not emit light is accumulated in the second capacitor. Image data of only reflected light which is a difference between the quantities of electric charge of the two capacitors may be obtained by providing a mechanism for outputting the difference between the two quantities of electric charge in reading the charge from the cell.
  • SUMMARY OF THE INVENTION
  • The present invention provides an image capture module for inputting a shape of an object on a three-dimensional space, including a light emitting section configured to irradiate light to the object to be imaged; and image capture section, disposed in the vicinity of the light emitting section, configured to detect the light emitted by the emitting device and reflected by the object.
  • In the image capture module, the light emitting section is disposed in the vicinity of and around the image capture section so as to be geometrically symmetrical. The light emitting section is disposed at positions where the light irradiated from the light emitting section does not directly enter the image capture section.
  • The image capture module further includes a reflecting device, mounted around the light emitting section, and configured to irradiate light emitted from the light emitting section in an object image capture range. The light emitting section has a directivity that illuminates a range wider than the image capture range and illuminates more brightly the peripheral part of the image capture range than the center of the image capture range.
  • The image capture module further includes a filtering device, mounted in front of the light emitting section or the image capture section, for protecting the light emitting section or the image capture section. The filtering device selectivity passes a wavelength of light to be inputted.
  • The image capture module further includes a casing for concealing and storing the whole of the light emitting section and the image capture section. The casing has a connector for exchanging signals with an internal circuit of the image capture section.
  • The image capture module further includes a rotary shaft, mounted to the image capture module, for rotatably moving the image capture module. The connector is disposed near the rotary shaft.
  • The present invention is also directed to an image capture section configured to input to a shape of an object on a three-dimensional space, including an image capture module, containing a light emitting section configured to irradiate light to an object to be imaged; and an image capture section, disposed in the vicinity of the light emitting section, configured to detect the light emitted from the light emitting section and reflected by the object, and a mechanism configured to move the image capture module within predetermined range of an angle of elevation.
  • The image capture apparatus further includes handles, attached to the image capture module, configured to move the image capture module, and a casing configured to conceal and store an external circuit of the image capture module.
  • BRIEF DESCRIPTION OF DRAWINGS
  • A more complete appreciation of the invention and many of the attendant advantages thereof will be readily obtained as the same becomes better understood by reference to the following detailed description when considered in connection with the accompanying drawings, wherein:
  • FIG. 1 is a block diagram showing a structure of an inputting device having a light emitting section and image capture sensor according to the present invention;
  • FIG. 2 is a schematic block diagram showing a reflected light image detecting section of the present invention;
  • FIG. 3 is a diagram illustrating an effect of the relationship between positions of the light emitting sections and the image capture optical system of the invention;
  • FIGS. 4 a and 4 b are plan and sectional views showing a structure of the image capture module of the present invention;
  • FIGS. 5 a and 5 b are diagrams illustrating changes in directivity characteristics of the light emitting section obtained by using a reflecting plate according to the present invention;
  • FIG. 6 is an illustration of light reflected by the reflecting plate of the present invention;
  • FIG. 7 is an illustration of light reflected by the reflecting plate of the present invention;
  • FIGS. 8 a-8 h are illustrations of various arrangements of the image capture optical system and the light emitting sections of the present invention;
  • FIG. 9 is a sectional view showing a concealed image capture module of the present invention;
  • FIG. 10 is an illustration of the image capture apparatus of the present invention in which the image capture module is made movable;
  • FIG. 11 is an illustration of the image capture apparatus of the present invention having handles for moving the image capture module of the present invention; and
  • FIG. 12 is an illustration of the image capture apparatus of the present invention having a handle for moving the image capture module of the present invention.
  • DESCRIPTION OF PREFERRED EMBODIMENTS
  • The fundamental principle and the basic structure of an image capture section according to the present invention is first explained. The image capture section has a light emitting section configured to emit invisible light to an object and an image capture section configured to detect the light reflected from the object. The image capture device detects only the light emitted by the light emitting section and reflected by the object and to the end separates reflected light from external light such as room light or sun light by operating the light emitting section and the image capture section in synchronism.
  • Referring now to the drawings, wherein like reference numerals designate identical or corresponding parts throughout the several views, FIG. 1 shows the basic structure of the inventive image capture section and FIG. 2 shows a more detailed structure of a reflected light image detecting section 4 in the image capture section. The light emitted from the light emitting section 1 is reflected by the object 2, e.g., a hand, and the reflected light of the object 2 is imaged on a photo-receiving plane of the reflected light image detecting section 4 via an image-capture optical system 3, e.g., a lens. The reflected light image detecting section 4 detects a reflected light image indicative of a distribution of intensity of the reflected light. How to take out only the reflected light by the reflected light image detecting section 4 separately from the external light is hereinafter described.
  • The reflected light image detecting section 4 continuously outputs a quantity of reflected light of each pixel of the reflected light image. The output from the reflected light image detecting section 4 is amplified by an amplifier 5, is converted to digital data by an A/D converter section 6 and is then stored in a memory 7. The data stored in the memory 7 is read at predetermined timing. The read data is sent to a computer via an I/F circuit 8 to be processed. A timing controller section 9 controls operation of the image detecting section 4, A/D converter section 6, memory 7 and I/F circuit 8.
  • Next, the structure of the reflected light image detecting section 4 will be explained in detail by using FIG. 2. In the reflected light image detecting section 4, photo-receiving cells 15, each corresponds to one pixel of an image, are arrayed two-dimensionally. Although it is similar to other image-capture elements such as a CCD image sensor in that respect, it is different from the CCD in that it includes a mechanism configured to separate reflected light from the external light and take out only reflected light within the structure of the cell. The cell 15 includes a photo-electric converter section 10, an accumulating direction control section 11, a first charge accumulating section 12, a second charge accumulating section 13 and a selective output section 14.
  • The photo-electric converter section 10 converts incident light into electric charge. The first and second charge accumulating sections 12 and 13 accumulate the electric charge generated by the photo-electric converter section 10. The accumulating direction control section 11 controls to which of the first and second charge accumulating sections 12 and 13 the electric charge generated by the photo-electric converter section 10 should be accumulated. The selective output section 14 selects either the first charge accumulating section 12 or the second charge accumulating section 13 and reads its electric charge to the outside of the cell.
  • The following operation is carried out in order to obtain one reflected light image. At first, the light emitting section 1 emits light in pulses. During the emission, electric charge generated by the photo-electric converter section 10 is accumulated in the first charge accumulating section 12. Then, while the light emitting section 1 is not emit light, electric charge generated by the photo-electric converter section 10 is accumulated in the second charge accumulating section 13.
  • That is, the electric charge obtained by receiving the reflected light and the external light such as room light and sun light is accumulated in the first charge accumulating section 12 and the electric charge obtained by receiving only the external light is accumulated in the second charge accumulating section 13. When the electric charge is to be read from the cell, the two quantities of accumulated electric charge are read sequentially to take out a difference there between by a differential circuit 17. Only the reflected light may be taken out by subtracting the quantity of accumulated electric charge of the second charge accumulating section 13 is from the quantity of accumulated electric charge of the first charge accumulating section 12.
  • The relationship between positions of the light emitting section and the image capture section is important for the image capture section of the present invention. When there are objects at the same distance, it is desirable to be able to illuminate the surfaces of the objects equally. In other words, it is then desirable that the intensity of illumination be equal.
  • Further, the photo-receiving cells 15 may be a CCD image sensor. In this case, the CCD image sensor is controlled to output alternately electric charge generated by receiving light while the light emitting section emits and does not emit.
  • First Embodiment
  • In FIG. 3, reflected light from an object 25 is detected by an image capture element (not shown) via the image capture optical system 3. Assume that there are two light emitting sections at positions A1 and A2. An LED is used as a light emitting element in each light emitting section.
  • When the object 25 has a concave shape in which the center thereof is dented, there is a portion 26 at the dented part where the light emitting from the light emitting section A1 does not reach. Similarly, there is a portion 27 where the light emitting from the light emitting section A2 does not reach. The illuminance of these portions is low because only the light of one light emitting section strikes as compared to the other part where the light from both of the light emitting sections strike. Accordingly, a part in the reflected light image corresponding to that portion becomes dark.
  • Further, light from neither of the light emitting sections A1 and A2 reaches a portion 28 and no reflected light returns to photo-receiving cells from the portion 28 at all. Therefore, the reflected light image appears as if it has a hole at the center.
  • When the light emitting sections are located at positions B1 and B2, not at the positions A1 and A2, light reaches also to the concave portion at the center of the object 25, so that such a problem do not arise.
  • The same applies also to a case when an object has a convex shape whose center protrudes. Because light of only one light source reaches the outer slant portions, the illuminance there drops with the result that such portions are darkened relative to the other portions. To avoid the occurrence of darkened portions, it is important the all of points on an object to be picked up be coupled in a straight line to the light source and not be obstructed by the other part of the same object.
  • To that end, it is preferable to dispose light emitting sections as close as possible to the image capture optical system. This is important for the device emitting light from the light source, detecting its reflected light as an image and processing each pixel value of the image in connection with distance as described in the present embodiment.
  • FIGS. 4 a and 4 b show the image capture module constructed by disposing the light emitting sections close to the image capture optical system, wherein FIG. 4 a is a plan view in which the image capture module is seen from the top (from the object side) and FIG. 4 b is a sectional view when the image capture module is seen just from the side.
  • The image capture module shown in FIG. 4 a includes the image capture optical system 3 at the center and eight light emitting sections on a concentric circle around that. Here, the light emitting sections are composed of LEDs 30.
  • In FIG. 4 b, the image capture optical system 3 is located at the center and is surrounded by the LEDs 30. The image capture element is located below the image capture optical system 3. Covering the image capture module is part of a frame or a casing 32 for fixing the relationship of positions of the LEDs 30 and the image capture optical system 3.
  • In the first embodiment, an angle of view of the image capture optical system 3 is set large for the reason explained hereinafter. Because the image capture section forms an image by detecting light emitted from the light emitting sections and reflected by the object, the further the distance to the object, the smaller the quantity of reflected light to be received becomes. The distance to the object is desirable to be short for higher S/N. However, the shorter the distance, the narrower a range where an image can be detected becomes.
  • Suppose that a gesture inputting apparatus is to be constructed by utilizing the inventive device to input and analyze a shape of hand, for example. In such a case, the image-capture range must be sufficient to include the whole hand and to allow a certain degree of the hand movement. In particular, the image capture range must be at least 30 cm square. A diagonal distance of the image capture range of 30 cm square is about 42 cm. When it is realized by using an optical system having about 50 degrees of angle of view, the distance to the object is around 45 cm. When it is realized by an optical system whose angle of view is about 70 degrees, the distance to the object is around 30 cm.
  • Because the quantity of received light is inversely proportional to a square of the distance to the object when the quantity of emitted light of the light emitting sections is equal as described above, the quantity of reflected light is reduced by half or less when the distance to the object increases 45 cm from 35 cm. Accordingly, the angle of view of the image capture optical system is desirable set to be large.
  • Because this image capture section is arranged so as to form an image by receiving the light emitted by the light emitting section and reflected from the object, the light emitted by the light emitting section should not enter the photo-receiving element directly. Accordingly, the light emitting sections should be disposed at positions where the light thereof will not enter the image capture element directly.
  • As a method for realizing that disposition, the positional relationship between the light emitting sections 30 and the image capture optical system 3 should be defined so that the light will not enter the image capture optical system 3 directly. Otherwise a blocking member that blocks only the direct light without affecting the irradiation of light to the object or the imaging of the reflected light should be provided. The image capture module shown in FIGS. 4 a and 4 b combines both such structures.
  • Each LED of the light emitting section shown in FIG. 4 b is surrounded by a reflecting element in the form of a sloped recess 33 cut aslant around each LED so that the surface of the sloped recess reflects the light of each LED. Although the LED 30 has a wide directivity and illuminates a range exceeding far beyond the image capture range, the light illuminated beyond the image capture range will not reach the image capture element even if it is reflected by the object. The quantity of reflected light may be increased and the S/N may be approved by directing the light emitted toward the outside of the image capture range within the image capture range by the reflecting element.
  • Further, the use of the reflecting element allows the brightness of the peripheral part to be enhanced more than that of the center and to prevent the S/N at the peripheral part from dropping. This point will be explained further in detail in connection with FIG. 5.
  • The LED is constructed by concealing an LED chip, i.e., a light emitting body, by resin having an effect of lens. Although the LED chip itself emits light in omnidirectionally, various directive characteristics may be given to the LED by changing the lens effect of the concealing resin.
  • FIG. 5 a shows a light emitting directivity characteristic of an LED suitable for the present embodiment. This LED has a wide directivity and is capable of illuminating a wide range. As it is apparent from FIG. 5 a, the intensity of emission is gradually reduced as the direction of emission deviates from the front direction and is reduced almost to a half in the direction deviated by 50 degrees from the front direction.
  • Then, the reflecting element 33 is mounted in relation to the LED as shown in FIG. 6. The LED 30 is located at the center of and surrounded by the reflecting element 33. Light emission from the LED is reflected and the directivity thereof is changed by the reflecting element 33. FIG. 5 b shows a directive characteristic of the emission which has been changed by utilizing such reflecting element. Here, the reflecting element is disposed so that light emitted at an angle of 40 degrees or more from the front direction is reflected. Thus, the intensity of emission in the peripheral direction becomes greater than that of the front direction.
  • However, because the LED light emission may not be always directed to the center of the object space. For example, there may be a case when the slope and the shape of the reflecting element are better not to be symmetrical. While the symmetrical slope may be created just by creating a conical hole from the top, an asymmetrical slope may be created by creating a conical hole from the diagonal direction deviated from the orientation of the LED.
  • For instance, it is desirable to set the angle of the slope on the lens side at 38 degrees and that of the opposite slope at about 35 degrees (both are angles from the direction of the optical axis) when a distance from the center of the lens to the LED is 20 mm and in order to illuminate a space of 20 cm square distance from the lens by 20 cm.
  • Further, the reflection characteristic of the reflecting element 33 is not limited in specular reflection. Usually, the accuracy of the position of the LED chip is not high and it is not reasonable to mount the LED high-accurately. Then, the reflecting element which has reflection characteristic of specular reflection makes loss of the illuminance for the object space by the error of the position of the LED.
  • In this case, the surface of the reflecting element may be diffuse reflection surface. In this way, the reflecting light reflected by the reflecting element diffuses any direction. So, the reflecting element of diffuse reflection surface is resolved the loss of the illuminance for the object space by the error of the position of the LED.
  • The shape of the reflecting element is arranged by a directivity characteristic of an LED, a characteristic of the surface of the reflecting element and the desirable characteristic of the emission. One of the desirable shape of the reflecting element is a shape 63 as, shown in FIG. 7. It is supposed that LED exists a starting point in this graph of FIG. 7. A direction of lengthwise axis the front direction of the LED. This graph shows a right side of a sectional view. The shape of this reflecting element has a characteristic mentioned below.
  • A light 64 of the 50 degrees direction from the center is reflected by the vertically reflecting element. And, the light 64 is reflected to the direction 50 degrees. A light 65 of the direction more than 50 degrees from the LED is reflected to the vertical direction according to the degrees increased. A light 66 of the 90 degrees direction from the center is reflected to the vertical direction by the reflecting element of 45 degrees. And, the light 66 is reflected to the vertical direction. The light of the direction which is from 50 degrees to 90 degrees is reflected uniformly to the opposite side which is from 50 degrees to 0 degree.
  • Because of the LED is a small surface light source which is not a point light source, the reflecting element which has specular reflection of this shape may be loss of the illuminance for the object space by the error of the position and the shape of the LED. So, the surface of the reflecting element has the characteristic of the diffuse reflection. In the case of the specular reflection, the illuminance is bright for the reflected light reflects to the vertical direction from the 90 degrees direction. So, it is desirable that the reflecting element is designed for the reflected light reflects to the direction shifted from 90 degrees.
  • Next, difference possible arrangements of the plurality of LEDs will be explained in relation to FIGS. 8 a-8 h which show examples of arrangements corresponding to different numbers of LEDs. Generally, the LEDs are arranged symmetrically about the center of the lens, except in the case 40 where only a single LED is provided, as shown in FIG. 8 a.
  • It is desirable to be able to illuminate the space to be detected by uniform light as much as possible. When all of the LEDs are disposed on one side of the lens, the side of the space to be detected as far from the LEDs is darkened as compared to the opposite side.
  • Although FIG. 8 a-8 f show exemplary arrangements 40 through 45 in which 2, 3, 4, 6 and 8 LEDs are used, the number of LEDs of the invention is not limited to those numbers. Although the LEDs cannot be disposed symmetrically except when there is only one LED, it is effective to dispose them closely to the lens.
  • Also, the LEDs may be disposed at the bilaterally symmetrical positions in the examples 46 and 47 shown in FIGS. 8 g and 8 h, they need not be disposed evenly around the image capture optical system. This arrangement may be adopted when the image capture range is rectangular and is lengthy in the horizontal direction. An advantage of this arrangement is that because the LEDs are mounted at the two places to the right and left, the LEDs may be mounted per group on a small substrate, thus allowing the substrate to be miniaturized and facilitating the packaging.
  • Second Embodiment
  • This image capture module allows a device to be manipulated in a non-contact manner when it is not desirable to manipulate the device directly by hand because the hands are dirty for example. However, there has been a possibility that this image capture module may malfunction when exposed in an environment where it is liable to get wet by water or to become dirty by mud for example.
  • Further, the quality of the detected image deteriorates when foreign substances such as mud and dust are lodged between the image capture optical system and the image capture sensor. Therefore, it is very important for the device that might be used in such an adverse environment to construct the image capture module so that it will not affected by such substances.
  • FIG. 9 shows an image capture module having a concealed structure to solve such problems. In FIG. 9, the LEDs 30, i.e., the light emitting section, the image capture optical system 3, the image capture sensor 31 and some circuit parts are stored in a concealed casing 52.
  • The image capture optical system 3, the LEDs 30 and the image capture element 31 mounted on a substrate 51 as well as their positional relationship are fixed by the casing 52. A peripheral circuit is mounted on the substrate 51 as necessary. Cables for emitting the LEDs are led out of the substrate.
  • Preferably, the upper part thereof is covered by an optical filter 50 which selectively transmits light of wavelength to be received to form an image. Because near infrared rays are used in the present embodiment, a near infrared ray transmission filter is used here. This filter will not transmit visible light, so that the user is unable to see the inside beyond the filter.
  • The image capture optical system, the substrate on which the image capture element is mounted, and the LEDs are concealed by the casings 52 and 53 as well as the optical filter 50.
  • It is advantageous to use the optical filter also from the following viewpoint. When the near infrared LED is used as the light emitting section, the user cannot see the light. However, in one embodiment a small fluorescent lamp may be used instead of the near infrared LED in order to assure a much larger quantity of light. However, because the user can see the light when the fluorescent lamp is used, a filter for cutting the visible light must be used to prevent the glare.
  • Meanwhile, an optical filter for filtering out light other than that of predetermined wavelength in order to reduce the amount of external light such as room light and sun light may be used together in the image capture optical system. It is, however, cumbersome to mount the similar optical filters in the light emitting section and the image capture optical system.
  • Then, it is preferable to dispose the light emitting section and the image capture optical system as close as possible and to cover the front face thereof by the optical filter filtering out light other than that of the predetermined wavelength. The cost may be also lowered by disposing the image capture optical system and the light emitting section closely.
  • The image capture module includes the image capture sensor, forming an electrical circuit together with several accompanying electronic parts. Therefore, the internal circuit must be connected with the external circuit when the image capture module is concealed. Although it is possible to connect the cables directly with the external circuit, it is preferable to provide a connector 54 for connecting to the outside in order to facilitate the assembly of the image capture module to other equipments.
  • The image capture section of the present invention may be used as an input device of various equipments. A designer of an equipment with which the inventive image capture section of the present invention is to be assembled can use the device of the present invention as an input device of arbitrary information by arranging the equipment so as to detect the shape of a hand, for example, by the image capture module, to include an image processor to convert the image into data which can be used as controllable information and then to take out only the data to the outside of the connector 54. The equipment can be used in an environmentally hostile application by utilizing a water-proof or anti-dust connector.
  • Preferably, the image capture section is installed so as to be able to change the image capture direction. When it is used as a device for inputting the shape and the motion of moving hand, for example, the position where the hand can be readily moved differs depending on person. In such a case, it is preferable to arrange the device so that the user can control the image capture direction.
  • The simplest structure which allows the image capture direction to be controlled is to arrange the image capture module so that it is rotatably mounted on a rotary shaft extending in the direction vertical to the optical axis of the image capture optical system in the present embodiment. In that case, when the connector is mounted at a position distant from the rotary shaft, the position of the connector may largely changes due to the rotation. Then, the cable connecting from the connector to the other circuit must be made longer. In order to avoid this problem, the connector is positioned as close to the rotary shaft as possible. Thereby, the cable connecting the image capture module with the external circuit may be shortened. It also avoids a long cable from becoming entangled with other parts when the image capture module is moved. In FIG. 9, the image capture module is constructed so that the rotary shaft 55 is close to the connector 54.
  • Third Embodiment
  • FIG. 10 shows an image capture section arranged such that an angle of deviation of the image capture module can be controlled in a range from 30 degrees to 90 degrees. Because the image capture section allows the image capture direction to be controlled, each user can control its direction.
  • This device is arranged so that it can be placed on a desk and be manually manipulated by user. Normally, it can be used readily with the user making such manipulation by resting his/her elbow upon the desk. Then, the position of hand may differ depending on the length from the elbow to the tip of the hand. The above-mentioned mechanism can accommodate differences in location of the hand due to the noted user manipulation.
  • FIGS. 11 and 12 show image capture sections having attached handles 60 and 61 for changing the angle of the image capture module. When one touches the image capture optical system or the optical filter in changing the direction of the image capture section, the contacted part may be stained, thus deteriorating the quality of image detected. Then, the handles are attached so that the user by using the handles need not erroneously touch the image capture optical system or the optical filter.
  • The handle 60 is attached around the top of the LEDs of the image capture module in FIG. 11. The handle is attached at the position when it will not block the path of the light emitted by the light emitting section, reflected by the object and reaching to the image capture element. The handles are attached on the right and left sides to allow users of both right-hander and left-hander to manipulate the module readily.
  • FIG. 12 shows a casing when the handle 61 is attached on the side of the main body. The image capture module moves in linkage by sliding the handle along a groove 62. Although the handle is attached only at the right side in the figure, the handle may be attached on the both sides as a matter of course.
  • Obviously, numerous modifications and variations of the present invention are possible in light of the above teachings. It is therefore to be understood that within the scope of the appended claims, the invention may be practiced than as specifically described herein.

Claims (17)

1. An imaging apparatus comprising:
a casing including an optical filter;
a plurality of light sources, in the casing, each configured to emit an invisible light to be reflected by an object;
a two dimensional image sensor, in the casing, surrounded by the light sources and configured to receive the reflected invisible lights from the object through the optical filter; and
an image processing unit that obtains three dimensional image information from a plurality of images,
wherein the optical filter selectively transmits the invisible lights and blocks a visible light.
2. The imaging apparatus according to claim 1, wherein the light sources are infrared LEDs.
3. The imaging apparatus according to claim 1, wherein the two dimensional image sensor is a CCD image sensor.
4. The imaging apparatus according to claim 1, wherein a number of the light sources is more than or equal to 3.
5. The imaging apparatus according to claim 1, wherein the light sources are arranged symmetrically about a center of the two dimensional image sensor.
6. The imaging apparatus according to claim 1, further comprising:
a control mechanism configured to control the two dimensional image sensor to generate a first image when the light sources are emitting the invisible lights and a second image when the light sources are not emitting the invisible lights.
7. The imaging apparatus according to claim 6, wherein the three dimensional image information of the external object is obtained from the first image and the second image.
8. A portable imaging apparatus configured to be placed on a desk, comprising:
a casing including an optical filter at a top surface of the casing;
at least two light sources, in the casing, each configured to upwardly emit an invisible light, through the optical filter, to be reflected by an external object;
an image sensor, disposed in the casing, surrounded by the light sources, and configured to downwardly receive the reflected invisible lights from the external object through the optical filter; and
an image processing unit that obtains three dimensional image information form a plurality of images,
wherein the optical filter is configured to selectively transmit light having a predetermined wavelength corresponding to the light sources in the casing.
9. The image apparatus according to claim 8, wherein the light sources are infrared LEDs.
10. The imaging apparatus according to claim 8, wherein the image sensor is a CCD image sensor.
11. The imaging apparatus according to claim 10, further comprising:
a control mechanism configured to control the CCD image sensor to generate a first image when the light sources are emitting the invisible lights and a second image when the light sources are not emitting the invisible lights.
12. The imaging apparatus according to claim 11, wherein three dimensional image information of the external object is obtained from the first image and the second image.
13. The imaging apparatus according to claim 8, wherein the light sources are arranged symmetrically around the imaging sensor.
14. The imaging apparatus according to claim 13, further comprising:
a control mechanism configured to control the image sensor to generate a first image when the light sources are emitting the invisible lights and a second image when the light sources are not emitting the invisible lights.
15. The imaging apparatus according to claim 14, wherein three dimensional image information of the external object is obtained from the first image and the second image.
16. The imaging apparatus according to claim 8, wherein a number of the light sources is more than or equal to 3.
17. The imaging apparatus according to claim 8, wherein the light sources are arranged symmetrically about a center of the image sensor.
US11/366,494 1997-09-05 2006-03-03 Image capture module and image capture apparatus for inputting shape of object on three dimensional space Abandoned US20060146175A1 (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
JP24069297A JP3471201B2 (en) 1997-09-05 1997-09-05 Imaging module, and an imaging device
JP9-240692 1997-09-05
US09/146,957 US6628335B1 (en) 1997-09-05 1998-09-04 Image capture module and image capture apparatus for inputting shape of object on three dimensional space
US10/612,957 US7042490B2 (en) 1997-09-05 2003-07-07 Image capture module and image capture apparatus for inputting shape of object on three dimensional space
US11/366,494 US20060146175A1 (en) 1997-09-05 2006-03-03 Image capture module and image capture apparatus for inputting shape of object on three dimensional space

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US11/366,494 US20060146175A1 (en) 1997-09-05 2006-03-03 Image capture module and image capture apparatus for inputting shape of object on three dimensional space

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US10/612,957 Division US7042490B2 (en) 1997-09-05 2003-07-07 Image capture module and image capture apparatus for inputting shape of object on three dimensional space

Publications (1)

Publication Number Publication Date
US20060146175A1 true US20060146175A1 (en) 2006-07-06

Family

ID=17063292

Family Applications (3)

Application Number Title Priority Date Filing Date
US09/146,957 Expired - Fee Related US6628335B1 (en) 1997-09-05 1998-09-04 Image capture module and image capture apparatus for inputting shape of object on three dimensional space
US10/612,957 Expired - Fee Related US7042490B2 (en) 1997-09-05 2003-07-07 Image capture module and image capture apparatus for inputting shape of object on three dimensional space
US11/366,494 Abandoned US20060146175A1 (en) 1997-09-05 2006-03-03 Image capture module and image capture apparatus for inputting shape of object on three dimensional space

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US09/146,957 Expired - Fee Related US6628335B1 (en) 1997-09-05 1998-09-04 Image capture module and image capture apparatus for inputting shape of object on three dimensional space
US10/612,957 Expired - Fee Related US7042490B2 (en) 1997-09-05 2003-07-07 Image capture module and image capture apparatus for inputting shape of object on three dimensional space

Country Status (2)

Country Link
US (3) US6628335B1 (en)
JP (1) JP3471201B2 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2151740A4 (en) * 2008-03-24 2015-05-06 Japan Display West Inc Imaging apparatus, displaying/imaging apparatus, electronic apparatus and method for detecting object
US9113094B2 (en) 2011-08-04 2015-08-18 Panasonic Intellectual Property Management Co., Ltd. Camera apparatus

Families Citing this family (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4157223B2 (en) * 1999-04-13 2008-10-01 Hoya株式会社 3-dimensional image input device
JP4303354B2 (en) * 1999-04-13 2009-07-29 Hoya株式会社 3-dimensional image input device
US6822687B1 (en) * 1999-07-08 2004-11-23 Pentax Corporation Three-dimensional image capturing device and its laser emitting device
US20030076431A1 (en) * 2001-10-24 2003-04-24 Krymski Alexander I. Image sensor with pixels having multiple capacitive storage elements
JP2003259182A (en) * 2002-03-01 2003-09-12 Nidec Copal Corp Mobile information terminal device
ES2291670T3 (en) * 2002-10-03 2008-03-01 Etview Ltd. endotracheal tube image sensor.
JP4590851B2 (en) * 2003-10-15 2010-12-01 カシオ計算機株式会社 Noncontact control apparatus and program
DE10360761A1 (en) * 2003-12-23 2005-07-28 Airbus Deutschland Gmbh Lighting device for a surveillance camera
WO2005071945A1 (en) * 2004-01-22 2005-08-04 Ssd Company Limited An image pickup apparatus and an object
JP4566929B2 (en) * 2006-03-03 2010-10-20 富士通フロンテック株式会社 Imaging device
JP4804962B2 (en) * 2006-03-03 2011-11-02 富士通フロンテック株式会社 Imaging device
US8170322B2 (en) * 2006-03-22 2012-05-01 Jadak Llc Optical imaging system and method using a reflective background
JP4797752B2 (en) * 2006-03-31 2011-10-19 株式会社デンソー Moving object manipulation object extracting apparatus
TWI319490B (en) * 2006-08-02 2010-01-11 Compal Communications Inc Lens atapted to image capturing apparatus
JP4820255B2 (en) * 2006-09-27 2011-11-24 クラリオン株式会社 Vehicle equipment
JP2009252203A (en) * 2008-04-11 2009-10-29 Japan Aviation Electronics Industry Ltd Slit for optical touch panel, and optical touch panel
KR20100119936A (en) * 2009-05-04 2010-11-12 삼성전자주식회사 Apparaus for input in portable termianl
JP5401645B2 (en) * 2009-07-07 2014-01-29 学校法人立命館 Human interface device
JP2011199798A (en) * 2010-03-24 2011-10-06 Sony Corp Physical information obtaining apparatus, solid-state imaging apparatus, and physical information obtaining method
EP2754384B1 (en) 2013-01-10 2018-07-11 Ambu A/S Endobronchial tube with integrated image sensor and cleaning nozzle arrangement
EP2611356B1 (en) 2011-07-11 2014-12-24 Etview Ltd. Endobronchial tube
US20130128104A1 (en) * 2011-11-22 2013-05-23 Cognex Corporation Camera system with exchangeable illumination assembly
CN103676408A (en) * 2012-09-10 2014-03-26 富泰华工业(深圳)有限公司 Electronic device
DE102013105105B3 (en) * 2013-05-17 2014-11-06 Sick Ag 3D camera with at least one lighting device
US9658061B2 (en) 2013-12-31 2017-05-23 Faro Technologies, Inc. Line scanner that uses a color image sensor to improve dynamic range
US9531967B2 (en) * 2013-12-31 2016-12-27 Faro Technologies, Inc. Dynamic range of a line scanner having a photosensitive array that provides variable exposure
KR20160024307A (en) * 2014-08-25 2016-03-04 삼성전자주식회사 Apparatus and method for recognizing movement of a subject
KR20170131040A (en) * 2016-05-20 2017-11-29 한화테크윈 주식회사 Camera having infrared reflector

Citations (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3375752A (en) * 1965-02-19 1968-04-02 Itek Corp Copyboard lighting system
US3829662A (en) * 1969-04-17 1974-08-13 Canon Kk Recording medium having concealed information as input for electronic computer
US4681432A (en) * 1983-04-01 1987-07-21 Canon Kabushiki Kaisha Photo-electric converting device
US4739160A (en) * 1985-05-17 1988-04-19 Alps Electric Co., Ltd. Optical coordinate data entry device with integral lens and spacer
US4769535A (en) * 1986-01-07 1988-09-06 Alps Electric Co., Ltd. Dustproof structure for optical coordinate input apparatus
US4873651A (en) * 1987-04-21 1989-10-10 Case Western Reserve University Method and apparatus for reconstructing three-dimensional surfaces from two-dimensional images
US5023709A (en) * 1989-11-06 1991-06-11 Aoi Studio Kabushiki Kaisha Automatic follow-up lighting system
US5457492A (en) * 1993-07-15 1995-10-10 Citizen Watch Co., Ltd. Light source device for an image processor
US5781229A (en) * 1997-02-18 1998-07-14 Mcdonnell Douglas Corporation Multi-viewer three dimensional (3-D) virtual display system and operating method therefor
US5801763A (en) * 1995-07-06 1998-09-01 Mitsubishi Denki Kabushiki Kaisha Face image taking device
US5801373A (en) * 1993-01-01 1998-09-01 Canon Kabushiki Kaisha Solid-state image pickup device having a plurality of photoelectric conversion elements on a common substrate
US5810717A (en) * 1995-09-22 1998-09-22 Mitsubishi Cable Industries, Ltd. Bending mechanism and stereoscope using same
US5977938A (en) * 1996-05-21 1999-11-02 Fujitsu Limited Apparatus and method for inputting and outputting by using aerial image
US6002423A (en) * 1996-01-16 1999-12-14 Visidyne, Inc. Three-dimensional imaging system
US6091453A (en) * 1996-07-29 2000-07-18 Coan; Steven Hand held remote camera
US6102556A (en) * 1998-04-21 2000-08-15 Panavision, Inc. Detachable light filter holder
US6150930A (en) * 1992-08-14 2000-11-21 Texas Instruments Incorporated Video equipment and method to assist motor vehicle operators
US6554452B1 (en) * 1997-08-19 2003-04-29 Ppt Vision, Inc. Machine-vision ring-reflector illumination system and method
US6558606B1 (en) * 2000-01-28 2003-05-06 3D Systems, Inc. Stereolithographic process of making a three-dimensional object
US6580449B1 (en) * 2000-07-18 2003-06-17 Dhv International, Inc. Borehole inspection instrument having a low voltage, low power fiber optic light-head

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH036531B2 (en) 1985-08-31 1991-01-30 Nippon Denki Hoomu Erekutoronikusu Kk
JPH0570075A (en) 1991-09-11 1993-03-23 Hitachi Elevator Eng Co Ltd Automatic operation control method for wheelchair escalator
JP3419050B2 (en) 1993-11-19 2003-06-23 株式会社日立製作所 Input device
JPH08340479A (en) 1995-06-14 1996-12-24 Canon Inc Image pickup device
JPH09128141A (en) 1995-11-07 1997-05-16 Sony Corp Controller and control method
JP3257585B2 (en) 1996-03-29 2002-02-18 株式会社ビジュアルサイエンス研究所 Imaging apparatus using a spatial mouse
JP3321053B2 (en) 1996-10-18 2002-09-03 株式会社東芝 Information input apparatus and an information input method and correction data generator

Patent Citations (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3375752A (en) * 1965-02-19 1968-04-02 Itek Corp Copyboard lighting system
US3829662A (en) * 1969-04-17 1974-08-13 Canon Kk Recording medium having concealed information as input for electronic computer
US4681432A (en) * 1983-04-01 1987-07-21 Canon Kabushiki Kaisha Photo-electric converting device
US4739160A (en) * 1985-05-17 1988-04-19 Alps Electric Co., Ltd. Optical coordinate data entry device with integral lens and spacer
US4769535A (en) * 1986-01-07 1988-09-06 Alps Electric Co., Ltd. Dustproof structure for optical coordinate input apparatus
US4873651A (en) * 1987-04-21 1989-10-10 Case Western Reserve University Method and apparatus for reconstructing three-dimensional surfaces from two-dimensional images
US5023709A (en) * 1989-11-06 1991-06-11 Aoi Studio Kabushiki Kaisha Automatic follow-up lighting system
US6150930A (en) * 1992-08-14 2000-11-21 Texas Instruments Incorporated Video equipment and method to assist motor vehicle operators
US5801373A (en) * 1993-01-01 1998-09-01 Canon Kabushiki Kaisha Solid-state image pickup device having a plurality of photoelectric conversion elements on a common substrate
US5457492A (en) * 1993-07-15 1995-10-10 Citizen Watch Co., Ltd. Light source device for an image processor
US5801763A (en) * 1995-07-06 1998-09-01 Mitsubishi Denki Kabushiki Kaisha Face image taking device
US5810717A (en) * 1995-09-22 1998-09-22 Mitsubishi Cable Industries, Ltd. Bending mechanism and stereoscope using same
US6002423A (en) * 1996-01-16 1999-12-14 Visidyne, Inc. Three-dimensional imaging system
US5977938A (en) * 1996-05-21 1999-11-02 Fujitsu Limited Apparatus and method for inputting and outputting by using aerial image
US6091453A (en) * 1996-07-29 2000-07-18 Coan; Steven Hand held remote camera
US5781229A (en) * 1997-02-18 1998-07-14 Mcdonnell Douglas Corporation Multi-viewer three dimensional (3-D) virtual display system and operating method therefor
US6554452B1 (en) * 1997-08-19 2003-04-29 Ppt Vision, Inc. Machine-vision ring-reflector illumination system and method
US6102556A (en) * 1998-04-21 2000-08-15 Panavision, Inc. Detachable light filter holder
US6558606B1 (en) * 2000-01-28 2003-05-06 3D Systems, Inc. Stereolithographic process of making a three-dimensional object
US6580449B1 (en) * 2000-07-18 2003-06-17 Dhv International, Inc. Borehole inspection instrument having a low voltage, low power fiber optic light-head

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2151740A4 (en) * 2008-03-24 2015-05-06 Japan Display West Inc Imaging apparatus, displaying/imaging apparatus, electronic apparatus and method for detecting object
US9063613B2 (en) 2008-03-24 2015-06-23 Japan Display Inc. Image pickup device, display-and-image-pickup device, electronic apparatus and method of detecting an object
US9405406B2 (en) 2008-03-24 2016-08-02 Japan Display Inc. Image pickup device, display-and-image-pickup device, electronic apparatus and method of detecting an object
US9113094B2 (en) 2011-08-04 2015-08-18 Panasonic Intellectual Property Management Co., Ltd. Camera apparatus
US10070027B2 (en) 2011-08-04 2018-09-04 Panasonic Intellectual Property Management Co., Ltd. Camera apparatus

Also Published As

Publication number Publication date
KR19990029605A (en) 1999-04-26
US7042490B2 (en) 2006-05-09
US20040095465A1 (en) 2004-05-20
JP3471201B2 (en) 2003-12-02
JPH1185386A (en) 1999-03-30
US6628335B1 (en) 2003-09-30

Similar Documents

Publication Publication Date Title
US8743345B2 (en) Three-dimensional imaging and display system
US7442914B2 (en) System and method of determining a position of a radiation emitting element
US6744537B1 (en) Image reader
EP2927839B1 (en) Hand-mounted indicia-reading device with finger motion triggering
ES2207754T3 (en) Method and system for imaging an object or configuration.
US5686942A (en) Remote computer input system which detects point source on operator
US7177451B2 (en) Fingerprint input apparatus
US7321359B2 (en) Method and device for optical navigation
JP4748257B2 (en) Biometric authentication device
US7705835B2 (en) Photonic touch screen apparatus and method of use
KR100429302B1 (en) High contrast, low distortion optical acquisition system for image capturing
US7158659B2 (en) System and method for multiplexing illumination in combined finger recognition and finger navigation module
US8600196B2 (en) Optical scanners, such as hand-held optical scanners
US20020117549A1 (en) Barcode-readable computer mouse
US6300940B1 (en) Input device for a computer and the like and input processing method
US20040046737A1 (en) Information input apparatus, information input method, and recording medium
US6002124A (en) Portable image scanner with optical position sensors
US6587099B2 (en) Coordinate input/detection device detecting installation position of light-receiving device used for detecting coordinates
US20070018970A1 (en) Optical slider for input devices
US7164782B2 (en) System and method for time-space multiplexing in finger-imaging applications
EP1752864A2 (en) Methods and systems for detecting selections on a touch screen display
US6618038B1 (en) Pointing device having rotational sensing mechanisms
US7737959B2 (en) Position detection system using laser speckle
US5187748A (en) Optical apparatus for fingerprint identification system
US6100538A (en) Optical digitizer and display means for providing display of indicated position

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION