WO2004072843A1 - Traitement du signal d'un ecran tactile - Google Patents

Traitement du signal d'un ecran tactile Download PDF

Info

Publication number
WO2004072843A1
WO2004072843A1 PCT/NZ2004/000029 NZ2004000029W WO2004072843A1 WO 2004072843 A1 WO2004072843 A1 WO 2004072843A1 NZ 2004000029 W NZ2004000029 W NZ 2004000029W WO 2004072843 A1 WO2004072843 A1 WO 2004072843A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
light
screen
outputs
sections
Prior art date
Application number
PCT/NZ2004/000029
Other languages
English (en)
Inventor
John David Newton
Original Assignee
Next Holdings Limited
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Next Holdings Limited filed Critical Next Holdings Limited
Priority to EP04711522A priority Critical patent/EP1599789A4/fr
Priority to CA2515955A priority patent/CA2515955C/fr
Priority to JP2006502767A priority patent/JP4668897B2/ja
Priority to AU2004211738A priority patent/AU2004211738B2/en
Publication of WO2004072843A1 publication Critical patent/WO2004072843A1/fr
Priority to US11/033,183 priority patent/US7629967B2/en
Priority to US12/569,166 priority patent/US8456447B2/en
Priority to US12/578,165 priority patent/US8466885B2/en
Priority to US12/580,409 priority patent/US8289299B2/en
Priority to US12/582,092 priority patent/US20100103143A1/en
Priority to US12/709,803 priority patent/US8508508B2/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/042Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means
    • G06F3/0421Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means by interrupting or reflecting a light beam, e.g. optical touch-screen
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form

Definitions

  • the present invention relates to a touch sensitive screen and in particular to optically detecting the presence of an object by using signal processing.
  • Touch screens of the prior art can take on five main forms. These five forms of touch screen input device include resistive, capacitive, surface acoustic wave (SAW), infrared (IR), and optical. Each of these types of touch screen has its own features, advantages and disadvantages. Resistive is the most common type of touch screen technology. It is a low-cost solution found in many touch screen applications, including hand-held computers, PDA's, consumer electronics, and point-of-sale-applications.
  • a resistive touch screen uses a controller and a specifically coated glass overlay on the display face to produce the touch connection. The primary types of resistive overlays are 4-wire, 5-wire, and 8 wires.
  • the 5-wire and 8 -wire technologies are more expensive to manufacture and calibrate, while 4- wire provides lower image clarity.
  • Two options are generally given: polished or antiglare. Polished offers clarity of image, but generally introduces glare. Anti-glare will minimize glare, but will also further diffuse the light thereby reducing the clarity.
  • One benefit of using a resistive display is that it can be accessed with a finger (gloved o not), pen, stylus, or a hard object.
  • resistive displays are less effective in public environments due to the degradation in image clarity caused by the layers of resistive film, and its susceptibility to scratching.
  • Capacitive touch screens are all glass and designed for use in ATM's and similar kiosk type applications. A small current of electricity runs across the screen with circuits located at the comers of the screen to measure the capacitance of a person touching the overlay. Touching the screen interrupts the current and activates the software operating the kiosk. Because the glass and bezel that mounts it to the monitor can be sealed, the touch screen is both durable and resistant to water, dirt and dust. This makes it commonly used in harsher environments like gaming, vending retail displays, public kiosks and industrial applications. However, the capacitive touch screen is only activated by the touch of a human finger and a gloved finger, pen, stylus or hard object will not work. Hence, it is inappropriate for use in many applications, including medical and food preparation.
  • SAW Surface acoustic wave
  • a SAW touch screen uses a glass display overlay. Sound waves are transmitted across the surface of the display. Each wave is spread across the screen by bouncing off reflector arrays along the edges of the overlay. Two receivers detect the waves. When the user touches the glass surface, the user's finger absorbs some of the energy of the acoustic wave and the controller circuitry measures the touch location.
  • SAW touch screen technology is used in ATM's, Amusements Parks, Banking and Financial Applications and kiosks. The technology is not able to be gasket sealed, and hence is not suitable to many industrial or commercial applications. Compared to resistive and capacitive technologies, it provides superior image clarity, resolution, and higher light transmission.
  • Infrared technology relies on the interruption of an infrared light grid in front of the display screen.
  • the touch frame or opto-matrix frame contains a row of infrared LEDs and photo transistors; each mounted on two opposite sides to create a grid of invisible infrared light.
  • the frame assembly is comprised of printed wiring boards on which the opto-electronics are mounted and is concealed behind an infrared-transparent bezel.
  • the bezel shields the opto-electronics from the operating environment while allowing the infrared beams to pass through.
  • the infrared controller sequentially pulses the LEDs to create a grid of infrared light beams.
  • Infrared touch screens are often used in manufacturing and medical applications because they can be completely sealed and operated using any number of hard or soft objects. The major issue with infrared is the "seating" of the touch frame is slightly above the screen. Consequently, it is susceptible to "early activation" before the finger or stylus has actually touched the screen. The cost to manufacture the infrared bezel is also quite high.
  • Optical imaging for touch screens uses a combination of line-scan cameras, digital signal processing, front or back illumination and algorithms to determine a point of touch.
  • the imaging lenses image the user's finger, stylus or object by scanning along the surface of the display.
  • This type of touch screen is susceptible to false readings due to moving shadows and bright lights and also requires that the screen be touched before a reading is taken. Attempts have been made to overcome these disadvantages. Touch screens using optical imaging technology are disclosed in the following publications.
  • a touch screen using digital ambient light sampling is disclosed in US4943806, in particular this patent discloses a touch input device that continuously samples and stores ambient light readings and compares these with previously taken readings. This is done to minimise the effect of bright light and shadows.
  • a touch screen for use with a computer system is disclosed in US5914709.
  • a user input device sensitive to touch is disclosed that uses threshold adjustment processing. A light intensity value is read and an "ON" threshold is established, this threshold measurement and adjustment is frequently and periodically performed.
  • This US Patent Number 5317140 patent discloses a method for optically determining the position and direction of an object on a touch screen display.
  • a diffuser is positioned over the light sources to produce an average light intensity over the touch screen.
  • US Patent Number 5698845 discloses a touch screen display that uses an optical detection apparatus to modulate the ON/OFF frequency of light emitters at a frequency of twice the commercial AC line source. The receiver determines the presence of light and compares this to the actual signal transmitted.
  • US Patent Number 4782328 discloses a touch screen that uses a photosensor unit positioned at a predetermined height above the touch screen, and when a pointer nears the touch screen, rays of its reflected or shadowed ambient light allow it to be sensed.
  • US Patent Number 4868551 discloses a touch screen that can detect a pointer near the surface of the display by detecting light reflected by the pointer (reflected or diffusive).
  • the invention may broadly be said to consist in a touch display comprising: a screen for a user to touch and view an image on or through; light sources at one or more edges of said screen, said light sources directing light across the surface of said screen; at least two cameras having outputs, each said camera located at the periphery of said screen to image the space in front of said screen, said output including a scanned image; means for processing said outputs to detect the level of light, said light including: direct light from said light sources, and/or reflected light from said light sources; a processor receiving the processed outputs of said cameras, said processor employing triangulation techniques and said processed outputs to determine whether the processed outputs indicate the presence of an object proximate to said screen and if so the location of said object.
  • said processed output indicates the relative bearing of a presumed object location relative to said camera.
  • said processed output indicates the relative bearing of a presumed object location relative to the centre of the lens of said camera.
  • said processor determines location of said object as a planar screen coordinate.
  • said light sources are behind said screen arranged to project light through said screen and said display includes at each edge having a light source, light deflectors in front of said screen, directing light emitted from said light sources across the surface of said screen.
  • said cameras are line scan cameras, said camera output including information on line scanned and said processor using said information in determining location of said object.
  • said touch display including: means for modulating said light from said light sources to provide a frequency band within the imageable range of said cameras; means for excluding image data outside said frequency band.
  • said means for processing said outputs includes said means for excluding image data outside said frequency band and said means for excluding image data outside said frequency includes filtering.
  • said filtering includes applying a filter selected from the group consisting of: a comb filter; a high pass filter; a notch filter; and a band pass filter.
  • said touch display including means for controlling said light sources; and means for taking and processing an image taken in a non lighted ambient light state and in a lighted state; wherein said means for processing said outputs subtracts the ambient state from the lighted state before detecting the level of light.
  • said said light sources are LEDs and said touch display includes means for controlling the operation of sections of said light source independent of other sections of said light source.
  • means for controlling the operation of sections of said light source includes means for independently controlling the effective intensity of said light source.
  • said means for controlling sections of said light source comprises wiring said sections in antiphase and driving using a bridge drive.
  • means for controlling sections of said light source comprises using a diagonal bridge drive.
  • said means for controlling sections of said light source comprises using a shift register for each section to be controlled.
  • said means for taking and processing images includes controlling sections of said light sources and each said camera and said means for processing said outputs includes processing information on whether a said section is lighted or not.
  • the invention may broadly be said to consist in a touch display comprising: a screen for a user to touch and view an image on or through; light sources at one or more edges edge of said screen, said light sources directing light across the surface of said screen; at least two cameras having outputs located at the periphery of said screen, said cameras located so as not to receive direct light from said light sources, each said camera imaging the space in front of said screen, said output including a scanned image; means for processing said outputs to detect level of reflected light; and a processor receiving the processed outputs of said cameras, said processor employing triangulation techniques and said processed outputs to determine whether the processed outputs indicate the presence of an object proximate to said screen and if so the location of said object.
  • Preferably said processed output indicates the relative bearing of a presumed object location relative to said camera.
  • said processed output indicates the relative bearing of a presumed object location relative to the centre of the lens of said camera.
  • said processor determines location of said object as a planar screen coordinate.
  • said touch display including: means for modulating said light from said light sources to provide a frequency band within the imageable range of said cameras; means for excluding image data outside said frequency band.
  • said means for processing said outputs includes said means for excluding image data outside said frequency band and said means for excluding image data outside said frequency includes filtering.
  • filtering includes applying a filter selected from the group consisting of: a comb filter; a high pass filter; a notch filter; and a band pass filter.
  • said touch display including: means for controlling said light sources; and means for taking and processing an image taken in a non lighted ambient light state and in a lighted state; wherein said means for processing said outputs subtracts the ambient state from the lighted state before detecting the level of light.
  • said light sources are LEDs and said touch display includes means for controlling the operation of sections of said light source independent of other sections of said light source.
  • means for controlling the operation of sections of said light source includes means for independently controlling the effective intensity of said light source.
  • the means for controlling sections of said light source comprises wiring said sections in antiphase and driving using a bridge drive.
  • the means for controlling sections of said light source comprises using a diagonal bridge drive.
  • the means for controlling sections of said light source comprises using a shift register for each section to be controlled.
  • said means for taking and processing images includes controlling sections of said light sources and each said camera and said means for processing said outputs includes processing information on whether a said section is lighted or not.
  • some sections are lighted and others are not when an image is taken.
  • said screen is reflective, said camera further images said screen, and said means for processing outputs detects the level of light from the mirror image.
  • said processed out put indicates the relative bearing of a presumed object relative to said camera and the distance of said object from said screen.
  • the invention may broadly be said to consist in a method of receiving user inputs in reference to an image including the steps of: providing a screen for a user to touch and view an image on or through; providing light sources at one or more edges of said screen, said light sources directing light across the surface of said screen; providing at least two cameras having outputs, each said camera located at the periphery of said screen to image the space in front of said screen, said output including a scanned image; processing said outputs to detect the level of light, said light including: direct light from said light sources, and/or reflected light from said light sources; processing the processed outputs of said cameras, using triangulation techniques to obtain the location of said object.
  • Preferably said processed output indicates the relative bearing of a presumed object location relative to a said camera.
  • said processed output indicates the relative bearing of a presumed object location relative to the centre of the lens of said camera.
  • said location of is a planar screen co-ordinate.
  • said light sources are behind said screen and arranged to project light through said screen and said display includes at each edge having a light source, light deflectors in front of said screen, directing light emitted from said light sources across the surface of said screen.
  • said cameras are line scan cameras, said camera output including information on line scanned and said processor using said information in determining location of said object.
  • said method including the steps of: modulating said light from said light sources to provide a frequency band within the imageable range of said cameras ; excluding image data outside said frequency band.
  • the step of processing said outputs includes the steps of excluding image data outside said frequency band and said step of excluding image data outside said frequency includes filtering.
  • filtering includes the step of applying a filter selected from the group consisting of: a comb filter; a high pass filter; a notch filter; and a band pass filter.
  • said method including the steps of: controlling said light sources; and taking and processing an image taken in a non lighted ambient light state and in a lighted state; wherein said step of processing said outputs subtracts the ambient state from the lighted state before detecting the level of light.
  • said light sources are LEDs and said touch display includes means for controlling the operation of sections of said light source independent of other sections of said light source.
  • the step of controlling the operation of sections of said light source includes independently controlling the effective intensity of said light source.
  • the step of controlling sections of said light source comprises wiring said sections in antiphase and driving using a bridge drive.
  • the step of controlling sections of said light source comprises using a diagonal bridge drive.
  • the step of controlling sections of said light source comprises using a shift register for each section to be controlled.
  • the step of taking and processing images includes controlling sections of said light sources and each said camera and said step of processing said outputs includes processing information on whether a said section is lighted or not.
  • the invention may broadly be said to consist in a method of receiving user inputs in reference to an image including the steps of: providing a screen for a user to touch and view an image on or through; providing light sources at one or more edges edge of said screen, said light sources directing light across the surface of said screen; providing at least two cameras having outputs located at the periphery of said screen, said cameras located so as not to receive direct light from said light sources, each said camera imaging the space in front of said screen, said output including a scanned image; processing said outputs to detect level of reflected light; and processing the processed outputs of said cameras, employing triangulation techniques and said processed outputs to determine whether the processed outputs indicate the presence of an object proximate to said screen and if so the location of said object.
  • said processed output indicates the relative bearing of a presumed object location relative to said camera.
  • said processed output indicates the relative bearing of a presumed object location relative to the centre of the lens of said camera.
  • said processor determines location of said object as a planar screen coordinate.
  • said method including: means for modulating said light from said light sources to provide a frequency band within the imageable range of said cameras; means for excluding image data outside said frequency band.
  • said means for processing said outputs includes said means for excluding image data outside said frequency band and said means for excluding image data outside said frequency includes filtering.
  • Preferably filtering includes applying a filter selected from the group consisting of: a comb filter; a high pass filter; a notch filter; and a band pass filter.
  • a filter selected from the group consisting of: a comb filter; a high pass filter; a notch filter; and a band pass filter.
  • said method including means for controlling said light sources; and means for taking and processing an image taken in a non lighted ambient light state and in a lighted state; wherein said means for processing said outputs subtracts the ambient state from the lighted state before detecting the level of light.
  • said light sources are LEDs and said touch display includes means for controlling the operation of sections of said light source independent of other sections of said light source.
  • the means for controlling the operation of sections of said light source includes means for independently controlling the effective intensity of said light source.
  • the means for controlling sections of said light source comprises wiring said sections in antiphase and driving using a bridge drive.
  • the means for controlling sections of said light source comprises using a diagonal bridge drive.
  • the means for controlling sections of said light source comprises using a shift register for each section to be controlled.
  • said means for taking and processing images includes controlling sections of said light sources and each said camera and said means for processing said outputs includes processing information on whether a said section is lighted or not.
  • said screen is reflective
  • said camera further images said screen
  • said means for processing outputs detects the level of light from the mirror image.
  • said processed out put indicates the relative bearing of a presumed object relative to said camera and the distance of said object from said screen.
  • the invention may broadly be said to consist in a method of receiving user inputs in reference to an image: providing at least one light sources on or adjacent the periphery of said image, said light sources directing light across said image; detecting at at least two locations on or adjacent the periphery of said image, the level of light and providing said level as an output; processing said outputs using triangulation techniques to determine whether said outputs indicate the presence of an object proximate to said image and if so the location of said object.
  • said locations are substantially non-opposite so that when an object is present said output is substantially indicative of light reflected from said object.
  • the invention may broadly be said to consist in a user input device for locating an object with reference to an image comprising: at least one light source at or proximate to the periphery of said image, said light source directing light across said image; at one detector having an output, said detector located or in proximity to said image to image the space in front of said screen, said output indicative of a level of light; a processor receiving said outputs and using triangulation techniques and said outputs determining the presence of said object and if so the location of said object.
  • Figure 1 is a diagrammatic illustration of a front view of the preferred embodiment of the touch screen of the present invention
  • Figure la is an illustration of a cross sectional view through X-X of Figure 1
  • Figure lb is an illustration of front illumination of the preferred embodiment of the touch screen of the present invention
  • Figure 2 is an illustration of the mirroring effect in the preferred embodiment of the touch screen of the present invention
  • Figure 2a is a block diagram of the filter implementation of the preferred embodiment of the touch screen of the present invention.
  • Figure 2b is a diagrammatic illustration of the pixels seen by an area camera and transmitted to the processing module in the preferred embodiment of the present invention
  • FIG. 3 is a block diagram of the system of the preferred embodiment of the touch screen of the present invention.
  • Figure 4 is a side view of the determination of the position of an object using the mirrored signal in the preferred embodiment of the touch screen of the present invention
  • Figure 4a is top view of the determination of the position of an object using the mirrored signal in the preferred embodiment of the touch screen of the present invention
  • Figure 5 is an illustration of the calibration in the preferred embodiment of the touch screen of the present invention
  • Figure 6 is a graph representing in the frequency domain the output from the imager in the processing module in the preferred embodiment of the touch screen of the present invention
  • Figure 6a is a graph representing in the frequency domain the filters responses on the signal from the imager in the preferred embodiment of the touch screen of the present invention
  • Figure 6b is a graph representing in the frequency domain the separation of the object from the background after two types of filtering in the preferred embodiment of the touch screen of the present invention
  • Figure 7 is an illustration of a front view of the alternate embodiment of the touch screen of the present invention.
  • Figure 7a is an illustration of a cross sectional view through X-X of the alternate embodiment of the touch screen of the present invention.
  • Figure 7b is an illustration of rear illumination of the alternate embodiment of the touch screen of the present invention.
  • Figure 7c is an illustration of rear illumination controlling the sense height of the alternate embodiment of the present invention
  • Figure 7d is a diagrammatic illustration of the pixels seen by a line scan camera and transmitted to the processing module in the alternate embodiment of the present invention
  • Figure 8 is a graph representing simple separation of an object from the background in the alternate embodiment of the present invention
  • Figure 9 shows various driving arrangements for sectional backlights of the present invention
  • Figure 9a shows a two section backlight driven by two wires of the present invention
  • Figure 9b shows a twelve section backlight driven by 4 wires of the present invention.
  • Figure 9c shows a piece of distributed shift register backlight of the present invention.
  • the present invention relates to improvements in signal processing in the field of optical imaging touch screens.
  • the optical touch screen uses front illumination and is comprised of a screen, a series of light sources, and at least two area scan cameras located in the same plane and at the periphery of the screen.
  • the optical touch screen uses backlight illumination; the screen is surrounded by an array of light sources located behind the touch panel which are redirected across the surface of the touch panel. At least two line scan cameras are used in the same plane as the touch screen panel.
  • FIG. 3 A block diagram of a general touch screen system 1 is shown in Figure 3.
  • the processing module 10 performs many types of calculations including filtering, data sampling, and triangulation and controls the modulation of the illumination source 4.
  • the touch screen system 1 is comprised of a monitor 2, a touch screen panel 3, at least two lights 4, a processing module (not shown) and at least two area scan cameras 6.
  • the monitor 2 which displays information to the user, is positioned behind the touch screen panel 3.
  • Below the touch screen panel 3 and the monitor 2 are the area scan cameras 6 and light sources 4.
  • the light sources 4 are preferably Light Emitting Diodes (LED) but may be another type of light source, for example, a fluorescent tube. LEDs are ideally used as they may be modulated as required, they do not have an inherent switching frequency.
  • the cameras 6 and LEDs 4 are in the same plane as the touch panel 3.
  • the viewing field 6a of the area scan camera 6 and the radiation path 4a of the LEDs 4 are in the same plane and parallel to the touch panel 3.
  • an object 7, shown as a finger enters into the radiation path 4a, it is illuminated.
  • the mirrored signal occurs when the object 7 nears the touch panel 3.
  • the touch panel 3 is preferably made from glass which has reflective properties.
  • the finger 7 is positioned at a distance 8 above the touch panel 3 and is mirrored 7a in the touch panel 3.
  • the camera 6 (only shown as the camera lens) images both the finger 7 and the reflected image 7a.
  • the image of finger 7 is reflected 7a in panel 3; this can be seen through the field lines 6b, 6c and virtual field line 6d. This allows the camera 6 to image the reflected 7a image of the finger 7.
  • the data produced from the camera 6 corresponds to the position of the field lines 6e, 6b as they enter the camera 6. This data is then fed into a processing module 10 for analysis.
  • a section of the processing module 10 is shown in Figure 2a.
  • Within the processing module 10 is a series of scanning imagers 13 and digital filters 11 and comparators 12 implemented in software.
  • There are a set number of pixels on the touch panel for example 30,000 pixels. These may be divided up into 100 columns of 300 pixels. The number of pixels may be more or less than the numbers used, the numbers are used for example only.
  • a representation of this is shown in Figure 2a as one column is serviced by one image scanner 13 and three sets 14a, 14b, 14c of digital filters 11 and comparators 12, this allows information from three pixels to be read.
  • FIG. 2b A more illustrated example of this matrix is shown in Figure 2b.
  • Eight pixels 3a-3h are connected, in groups of columns, to an image scanner 13 that is subsequently connected to a filter 11 and a comparator 12 (as part of the processing module 10).
  • the numbers used in Figure 2b are used for illustration only; an accurate number of pixels could be greater or less in number.
  • the pixels shown in this diagram may not form this shape in the panel 3, their shape will be dictated by the position and type of camera 6 used.
  • finger 7 and mirrored finger 7a activates at least two pixels; two pixels are used for simplicity. This is shown by the field lines 6e and 6b entering the processing module 10.
  • the comparator 12 compares the output from the filter 11 to a predetermined threshold value. If there is a finger 7 detected at the pixel in question, the output will be high, otherwise it will be low.
  • the mirrored signal also provides information about the position of the finger 7 in relation to the cameras 6. It can determine the height 8 of the finger 7 above the panel 3 and its angular position. The information gathered from the mirrored signal is enough to determine where the finger 7 is in relation to the panel 3 without the finger 7 having to touch the panel 3.
  • Figures 4 and 4a show the positional information that is able to be obtained from the processing of the mirrored signal.
  • the positional information is given in polar coordinates.
  • the positional information relates to the height of the finger 7, and the position of the finger 7 over the panel 3.
  • the height that the finger 7 is above the panel 3 can be seen in the distance between the outputs 12a-12e.
  • the finger 7 is a height 8 above the panel 3 and the outputs 12b and 12e are producing a high signal.
  • the other outputs 12a, 12d are producing a low signal. It has been found that the distance 9 between the high outputs 12b, 12e is twice as great as the actual height 8 of the finger above the panel 3.
  • the processing module 10 modulates and collimates the LEDs 4 and sets a sampling rate.
  • the LEDs 4 are modulated, in the simplest embodiment the LEDs 4 are switched on and off at a predetermined frequency. Other types of modulation are possible, for example modulation with a sine wave. Modulating the LEDs 4 at a high frequency results in a frequency reading (when the finger 7 is sensed) that is significantly greater than any other frequencies produced by changing lights and shadows.
  • the modulation frequency is greater than 500Hz but no more than 10kHz.
  • the cameras 6 continuously generate an output, which due to data and time constraints is periodically sampled by the processing module 10.
  • the sampling rate is at least two times the modulation frequency; this is used to avoid aliasing.
  • the modulation of the LEDs and the sampling frequency does not need to be synchronised.
  • no signal is transmitted to the area camera so there are no other peaks in the output.
  • a signal 24 corresponding to the LED modulated frequency, for example 500Hz.
  • the lower unwanted frequencies 22, 23 can be removed by various forms of filters. Types of filters can include comb, high pass, notch, and band pass filters.
  • FIG 6a the output from the image scanner is shown with a couple of different filter responses26, 27 being applied to the signal 20.
  • a 500Hz comb filter 26 may be implemented (if using a 500Hz modulation frequency). This will remove only the lowest frequencies.
  • a more advanced implementation would involve using a band pass 27 or notch filter. In this situation, all the data, except the region where the desired frequency is expected, is removed.
  • this is shown as a 500Hz narrow band filter 27 applied to the signal 20 with a modulation frequency of 500Hz.
  • These outputs 30, 31 from the filters 26, 27 are further shown in Figure 6b.
  • the top graph shows the output 30 if a comb filter 26 is used while the bottom graph shows the output 31 when a band filter 27 is used.
  • the band filter 27 removes all unwanted signals while leaving the area of interest. Once the signal has been filtered and the signal in the area of interest identified, the resulting signal is passed to the comparators to be converted into a digital signal and triangulation is performed to determine the actual position of the object. Triangulation is known in the prior art and disclosed in US5534917 and US4782328, and are herein incorporated by reference. Calibration
  • the preferred embodiment of the touch screen of the present invention uses very quick and easy calibration that allows the touch screen to be used in any situation and moved to new locations, for example the touch screen is manufactured as a lap top.
  • Calibration involves touching the panel 3 in three different locations 31a, 31b, 31c, as shown in Figure 5; this defines the touch plane of the touch panel 3.
  • These three touch points 31a, 31b, 31c provide enough information to the processing module (not shown) to calculate the position and size of the touch plane in relation to the touch panel 3.
  • Each touch point 31a, 31b, 31c uses both mirrored and direct signals, as previously described, to generate the required data.
  • These touch points 31a, 31b, 31c may vary around the panel 3, they need not be the actual locations shown.
  • Figure 7 shows the alternate embodiment of the touch screen of the present invention.
  • the monitor 40 is behind the touch panel 41 and around the sides and the lower edge of the panel 41 is an array of lights 42. These point outwards towards the user and are redirected across the panel 41 by a diffusing plate 43.
  • the array of lights 42 consists of numerous Light Emitting Diodes (LEDs).
  • the diffusing plates 43 are used redirect and diffuse the light emitted from the LEDs 42 across the panel 41.
  • At least two line-scan cameras 44 are placed in the upper two corners of the panel 3 and are able to image an object.
  • the cameras 44 can be alternately placed at any position around the periphery of the panel 41.
  • Around the periphery of the touch panel 41 is a bezel 45 or enclosure.
  • the bezel 45 acts as a frame that stops the light radiation from being transmitted to the external environment.
  • the bezel 45 reflects the light rays into the cameras 44 so a light signal is always read into the camera 44 when there is no object
  • the array of lights 42 may be replaced with cold cathode tubes.
  • a diffusing plate 43 is not necessary as the outer tube of the cathode tube diffuses the light.
  • the cold cathode tube runs along the entire length of one side of the panel 41. This provides a substantially even light intensity across the surface of the panel 41.
  • Cold cathode tubes are not preferably used as they are difficult and expensive to modify to suit the specific length of each side of the panel 41. Using LED's allows greater flexibility in the size and shape of the panel 41.
  • the diffusing plate 43 is used when the array of lights 42 consists of numerous LED's.
  • the plate 43 is used to diffuse the light emitted from an LED and redirect it across the surface of panel 41.
  • the light 47 from the LEDs 42 begins its path at right angles to the panel 41. Once it hits the diffusing plate 43, it is redirected parallel to the panel 41.
  • the light 47 travels slightly above the surface of the panel 41 so to illuminate the panel 41.
  • the light 47 is collimated and modulated by the processing module (not shown) as previously described. Referring to Figure 7a, increasing the width 46 of the bezel 45 can be increased or decreased. Increasing the width 46 of the bezel 45 increases the distance at which an object can be sensed. Similarly, the opposite applies to decreasing the width 10 of the bezel 45
  • the line scan cameras 44 consists of a CCD element, lens and driver control circuitry. When an image is seen by the cameras 44 a corresponding output signal is generated.
  • the line scan cameras will be continuously reading the modulated light transmitted from the LEDs. This will result in the modulated frequency being present in the output whenever there is no object to interrupt the light path. When an object interrupts the light path, the modulated frequency in the output will not be present. This indicates that an object is in near to or touching the touch panel.
  • the frequency present in the output signal is twice the height (twice the amplitude) than the frequency in the preferred embodiment. This is due to both signals (direct and mirrored) being present at once.
  • the output from the camera is sampled when the LEDs are modulating on and off. This provides a reading of ambient light plus backlight 50 and a reading of ambient light alone 51.
  • an object interrupts the light from the LEDs, there is a dip 52 in the output 50.
  • the ambient reading 51 is subtracted from the ambient and backlight reading 50.
  • Calibration of this alternate embodiment is performed in the same manner as previously described but the touch points 31a, 31b, 31c (referring to Figure 5) cannot be in the same line, they must be spread about the surface of the panel 3.
  • the backlight is broken up into a number of individual sections, 42a to 42f.
  • One section or a subset of sections is activated at any time.
  • Each of these sections is imaged by a subset of the pixels of the image sensors 44.
  • the backlight emitters are operated at higher current for shorter periods. As the average power of the emitter is limited, the peak brightness is increased. Increased peak brightness improves the ambient light performance.
  • the backlight switching may advantageously be arranged such that while one section is illuminated, the ambient light level of another section is being measured by the signal processor. By simultaneously measuring ambient and backlit sections, speed is improved over single backlight systems.
  • the backlight brightness is adaptively adjusted by controlling LED current or pulse duration, as each section is activated so as to use the minimum average power whilst maintaining a constant signal to noise plus ambient ratio for the pixels that view that section.
  • Control of the plurality of sections with a minimum number of control lines is achieved in one of several ways.
  • a first implementation of a two section backlight the two groups of diodes 44a,
  • 44b can be wired antiphase and driven with bridge drive.
  • diagonal bridge drive is used.
  • 4 wires are able to select 1 of 12 sections, 5 wires can drive 20 sections, and 6 wires drive 30 sections.
  • a shift register 60 is physically distributed around the backlight, and only two control lines are required.
  • X-Y multiplexing arrangements are well known in the art. For example an 8+4 wires are used to control a 4 digit display with 32 LED's. Fig9b shows a 4 wire diagonal multiplexing arrangement with 12 LEDs.
  • the control lines A,B,C,D are driven by tristate outputs such as are commonly found at the pins of microprocessors such as the Microchip PIC family. Each tristate output has two electronic switches which are commonly mosfets. Either or neither of the switches can be turned on. To operate led Lla, switches Al and BO only are enabled. To operate LIB, A0 and Bl are operated. To operate L2a, Al and DO are enabled, and so on.
  • This arrangement can be used with any number of control lines, but is particularly advantageous for the cases of 4,5,6 control lines, where 12,20,30 leds can be controlled whilst the printed circuit board tracking remains simple. Where higher control numbers are used it may be advantageous to use degenerate forms where some of the possible leds are omitted to ease the practical interconnection difficulties.
  • the diagonal multiplexing system has the following features:
  • the arrangement is represented by a ring of control lines with a pair of antiphase LED's arranged on each of the diagonals between the control lines.
  • Each LED can be uniquely selected, and certain combinations can also be selected.

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Length Measuring Devices By Optical Means (AREA)
  • Position Input By Displaying (AREA)

Abstract

Cette invention se rapporte à un écran tactile (1) qui utilise des sources lumineuses (4) sur un ou plusieurs bords de l'écran (1) pour diriger le rayonnement lumineux sur la surface de l'écran (1) et au moins deux caméras (6) comportant des sorties électroniques situées sur le pourtour de l'écran (1) pour recevoir le rayonnement lumineux provenant desdites sources lumineuses (4). Un processeur reçoit des sorties de ces caméras (6) et utilise des techniques de triangulation pour déterminer la position d'un objet proche de l'écran (1). Pour détecter la présence de l'objet, on détecte au niveau des caméras (6) la présence ou l'absence de rayonnement lumineux direct dû à l'objet, on utilise la surface de l'écran comme miroir et on détecte au niveau des caméras (6) la présence ou l'absence de rayonnement lumineux réfléchi dû à l'objet. Les sources lumineuses (4) peuvent être modulées pour produire une bande de fréquences dans la sortie des caméras (6).
PCT/NZ2004/000029 2003-02-14 2004-02-16 Traitement du signal d'un ecran tactile WO2004072843A1 (fr)

Priority Applications (10)

Application Number Priority Date Filing Date Title
EP04711522A EP1599789A4 (fr) 2003-02-14 2004-02-16 Traitement du signal d'un ecran tactile
CA2515955A CA2515955C (fr) 2003-02-14 2004-02-16 Traitement du signal d'un ecran tactile
JP2006502767A JP4668897B2 (ja) 2003-02-14 2004-02-16 タッチスクリーン信号処理
AU2004211738A AU2004211738B2 (en) 2003-02-14 2004-02-16 Touch screen signal processing
US11/033,183 US7629967B2 (en) 2003-02-14 2005-01-11 Touch screen signal processing
US12/569,166 US8456447B2 (en) 2003-02-14 2009-09-29 Touch screen signal processing
US12/578,165 US8466885B2 (en) 2003-02-14 2009-10-13 Touch screen signal processing
US12/580,409 US8289299B2 (en) 2003-02-14 2009-10-16 Touch screen signal processing
US12/582,092 US20100103143A1 (en) 2003-02-14 2009-10-20 Touch screen signal processing
US12/709,803 US8508508B2 (en) 2003-02-14 2010-02-22 Touch screen signal processing with single-point calibration

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
NZ52421103 2003-02-14
NZ524211 2003-02-14

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US11/033,183 Continuation US7629967B2 (en) 2003-02-14 2005-01-11 Touch screen signal processing

Publications (1)

Publication Number Publication Date
WO2004072843A1 true WO2004072843A1 (fr) 2004-08-26

Family

ID=32867173

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/NZ2004/000029 WO2004072843A1 (fr) 2003-02-14 2004-02-16 Traitement du signal d'un ecran tactile

Country Status (7)

Country Link
EP (1) EP1599789A4 (fr)
JP (1) JP4668897B2 (fr)
KR (1) KR101035253B1 (fr)
CN (1) CN100468303C (fr)
AU (1) AU2004211738B2 (fr)
CA (1) CA2515955C (fr)
WO (1) WO2004072843A1 (fr)

Cited By (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN100370403C (zh) * 2004-10-29 2008-02-20 索尼株式会社 输入/输出装置和方法
WO2008128096A3 (fr) * 2007-04-11 2009-04-16 Next Holdings Inc Système à écran tactile avec procédés de saisie par effleurement et clic
WO2009139971A3 (fr) * 2008-05-12 2010-01-07 Microsoft Corporation Détection tactile multipoint fondée sur la vision par ordinateur utilisant des lasers infrarouges
CN102202172A (zh) * 2010-03-23 2011-09-28 北京汇冠新技术股份有限公司 用于带摄像头触摸屏的摄像头同步方法和系统
WO2011117505A2 (fr) * 2010-03-25 2011-09-29 Intui Sense Dispositif interactif adaptable aux conditions de luminosite ambiante
US8055022B2 (en) 2000-07-05 2011-11-08 Smart Technologies Ulc Passive touch system and method of detecting user input
CN102281395A (zh) * 2010-06-08 2011-12-14 北京汇冠新技术股份有限公司 摄像头同步方法、控制板、触摸屏、触摸系统和显示器
US8089462B2 (en) 2004-01-02 2012-01-03 Smart Technologies Ulc Pointer tracking across multiple overlapping coordinate input sub-regions defining a generally contiguous input region
US8094137B2 (en) 2007-07-23 2012-01-10 Smart Technologies Ulc System and method of detecting contact on a display
US8120596B2 (en) 2004-05-21 2012-02-21 Smart Technologies Ulc Tiled touch system
WO2012070950A1 (fr) * 2010-11-22 2012-05-31 Epson Norway Research And Development As Système et procédé d'interaction à touchers multiples et d'éclairage à base de caméra
US8228304B2 (en) 2002-11-15 2012-07-24 Smart Technologies Ulc Size/scale orientation determination of a pointer in a camera-based touch system
US8274496B2 (en) 2004-04-29 2012-09-25 Smart Technologies Ulc Dual mode touch systems
US8289299B2 (en) 2003-02-14 2012-10-16 Next Holdings Limited Touch screen signal processing
US8325134B2 (en) 2003-09-16 2012-12-04 Smart Technologies Ulc Gesture recognition method and touch system incorporating the same
US8339378B2 (en) 2008-11-05 2012-12-25 Smart Technologies Ulc Interactive input system with multi-angle reflector
US8384693B2 (en) 2007-08-30 2013-02-26 Next Holdings Limited Low profile touch panel systems
US8405636B2 (en) 2008-01-07 2013-03-26 Next Holdings Limited Optical position sensing system and optical position sensor assembly
US8432377B2 (en) 2007-08-30 2013-04-30 Next Holdings Limited Optical touchscreen with improved illumination
US8456447B2 (en) 2003-02-14 2013-06-04 Next Holdings Limited Touch screen signal processing
US8456418B2 (en) 2003-10-09 2013-06-04 Smart Technologies Ulc Apparatus for determining the location of a pointer within a region of interest
US8456451B2 (en) 2003-03-11 2013-06-04 Smart Technologies Ulc System and method for differentiating between pointers used to contact touch surface
WO2013104061A1 (fr) * 2012-01-11 2013-07-18 Smart Technologies Ulc Étalonnage d'un rideau lumineux interactif
US8508508B2 (en) 2003-02-14 2013-08-13 Next Holdings Limited Touch screen signal processing with single-point calibration
US8847739B2 (en) 2008-08-04 2014-09-30 Microsoft Corporation Fusing RFID and vision for surface object tracking
US8902193B2 (en) 2008-05-09 2014-12-02 Smart Technologies Ulc Interactive input system and bezel therefor
US9168865B2 (en) 2011-10-11 2015-10-27 Hyundai Motor Company Organic electro-luminescence lighting device for vehicle
US9171454B2 (en) 2007-11-14 2015-10-27 Microsoft Technology Licensing, Llc Magic wand
US9442607B2 (en) 2006-12-04 2016-09-13 Smart Technologies Inc. Interactive input system and method
US9582116B2 (en) 2010-09-02 2017-02-28 Baanto International Ltd. Systems and methods for sensing and tracking radiation blocking objects on a surface
US9880668B2 (en) 2013-09-11 2018-01-30 Beijing Lenovo Software Ltd. Method for identifying input information, apparatus for identifying input information and electronic device
US10627973B2 (en) 2009-06-18 2020-04-21 Baanto International Ltd. Systems and sensors for sensing and tracking radiation blocking objects on a surface

Families Citing this family (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7538759B2 (en) 2004-05-07 2009-05-26 Next Holdings Limited Touch panel display system with illumination and detection provided from a single edge
KR100756026B1 (ko) * 2006-07-19 2007-09-07 주식회사 엠씨넥스 카메라를 이용한 조작 장치와 전자 기기
KR100894544B1 (ko) * 2006-11-30 2009-04-24 한국전자통신연구원 스페이스 터치스크린 장치 및 그 작동방법
WO2008087935A1 (fr) * 2007-01-17 2008-07-24 Kabushiki Kaisha Sega Doing Business As Sega Corporation Machine de jeu à véhicule stationnaire
KR100900321B1 (ko) * 2007-08-20 2009-06-02 윤주영 적외선카메라를 이용한 멀티 터치스크린
CN101216746B (zh) * 2007-12-26 2010-12-01 广东威创视讯科技股份有限公司 触摸屏装置及其定位方法
US8259208B2 (en) * 2008-04-15 2012-09-04 Sony Corporation Method and apparatus for performing touch-based adjustments within imaging devices
CN101576787B (zh) * 2008-05-07 2011-05-25 联想(北京)有限公司 一种电子设备、笔记本电脑及触摸控制实现方法
BRPI0822675B1 (pt) * 2008-09-15 2019-09-17 Hewlett-Packard Development Company, L.P. Sistema de display e método
KR100879328B1 (ko) 2008-10-21 2009-01-19 (주)컴버스테크 카메라를 이용한 핑거 뎁스 조절 장치 및 방법과 카메라를 이용한 핑거 뎁스 조절 장치를 갖는 터치 스크린
KR100901971B1 (ko) * 2008-11-07 2009-06-10 김용철 적외선 터치 스캐닝 모듈
JP2010277122A (ja) * 2009-05-26 2010-12-09 Xiroku:Kk 光学式位置検出装置
WO2010145002A1 (fr) * 2009-06-16 2010-12-23 Baanto International Ltd. Systèmes de détection bidimensionnelle de position et capteurs associés
CN201465076U (zh) * 2009-07-03 2010-05-12 北京汇冠新技术股份有限公司 一种用于触摸屏的光源及一种触摸屏
CN102129330A (zh) * 2009-07-07 2011-07-20 松翰科技股份有限公司 触控屏幕、触控模块及控制方法
JP5374266B2 (ja) * 2009-07-22 2013-12-25 株式会社シロク 光学式位置検出装置
JP5273676B2 (ja) * 2009-08-10 2013-08-28 ヌライテン トレーディング (シャンハイ)カンパニー 広域赤外線光源マルチタッチスクリーン
TWI420357B (zh) 2009-08-28 2013-12-21 Pixart Imaging Inc 觸控系統及其指示物座標偵測方法
JP2011090604A (ja) * 2009-10-26 2011-05-06 Seiko Epson Corp 光学式位置検出装置および位置検出機能付き表示装置
KR101715851B1 (ko) * 2009-12-15 2017-03-15 엘지디스플레이 주식회사 광학 센싱 유닛, 이를 이용한 표시 모듈 및 표시 장치
JP2011175543A (ja) * 2010-02-25 2011-09-08 Sanyo Electric Co Ltd 指示体検出装置及びタッチパネル
CN101847063B (zh) * 2010-03-03 2013-04-17 苏州佳世达电通有限公司 利用不同时形成的光域的物体侦测系统及方法
CN101923418B (zh) * 2010-03-07 2013-01-16 苏州佳世达电通有限公司 物体侦测系统及方法
US9557837B2 (en) 2010-06-15 2017-01-31 Pixart Imaging Inc. Touch input apparatus and operation method thereof
TWI416942B (zh) * 2010-08-27 2013-11-21 Disk King Technology Co Ltd High Panoramic Photographic Touch Device and Method with High Operation Speed
CN102692182A (zh) * 2011-03-23 2012-09-26 刘中华 一种用于屏幕触控输入装置的光学检测系统
US20120313895A1 (en) 2011-06-10 2012-12-13 Texas Instruments Incorporated Touch screen
TWI450159B (zh) * 2012-03-02 2014-08-21 Pixart Imaging Inc Optical touch device, passive touch system and its input detection method
CN103309517A (zh) * 2012-03-15 2013-09-18 原相科技股份有限公司 光学式输入装置及其输入检测方法
KR101386503B1 (ko) 2012-07-30 2014-04-17 서울시립대학교 산학협력단 광 센서의 삼각측량 배치 기반의 인터랙티브 디스플레이 장치
CN103902084B (zh) * 2012-12-27 2017-06-13 上海天马微电子有限公司 一种触摸屏
CN103941920A (zh) * 2013-01-21 2014-07-23 原相科技股份有限公司 光学触控系统

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4782328A (en) 1986-10-02 1988-11-01 Product Development Services, Incorporated Ambient-light-responsive touch screen data input method and system
US4811004A (en) * 1987-05-11 1989-03-07 Dale Electronics, Inc. Touch panel system and method for using same
US4868551A (en) 1983-10-28 1989-09-19 Thomson-Csf Sensitive display device comprising a scanned screen
US5164714A (en) 1988-06-20 1992-11-17 Amp Incorporated Modulated touch entry system and method with synchronous detection
US5317140A (en) 1992-11-24 1994-05-31 Dunthorn David I Diffusion-assisted position location particularly for visual pen detection
US5534917A (en) 1991-05-09 1996-07-09 Very Vivid, Inc. Video image based control system
EP0897161A1 (fr) * 1997-08-07 1999-02-17 Fujitsu Limited Panneau tactile optique à balayage
US20010026268A1 (en) * 2000-03-31 2001-10-04 Ricoh Company, Ltd. Coordiante input and detection device and information display and input apparatus
WO2001091043A2 (fr) 2000-05-24 2001-11-29 Next Holdings Limited Systeme d'affichage a ecran tactile
US6352351B1 (en) * 1999-06-30 2002-03-05 Ricoh Company, Ltd. Method and apparatus for inputting coordinates
US6504532B1 (en) * 1999-07-15 2003-01-07 Ricoh Company, Ltd. Coordinates detection apparatus

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4746770A (en) * 1987-02-17 1988-05-24 Sensor Frame Incorporated Method and apparatus for isolating and manipulating graphic objects on computer video monitor
JPH0521153Y2 (fr) * 1987-10-08 1993-05-31
JP2896183B2 (ja) * 1990-02-22 1999-05-31 株式会社ワコム 光学式複数2次元座標同時入力装置
JPH1031546A (ja) * 1996-07-17 1998-02-03 Yamaha Corp 座標入力装置
JP3624070B2 (ja) * 1997-03-07 2005-02-23 キヤノン株式会社 座標入力装置及びその制御方法
JPH11110116A (ja) * 1997-08-07 1999-04-23 Fujitsu Ltd 光学式位置検出装置
JP2000105671A (ja) * 1998-05-11 2000-04-11 Ricoh Co Ltd 座標入力/検出装置および電子黒板システム
JP4083941B2 (ja) * 1999-09-03 2008-04-30 株式会社リコー 座標入力装置
JP4208394B2 (ja) * 2000-08-09 2009-01-14 株式会社リコー 座標入力装置
JP4538933B2 (ja) * 2000-10-06 2010-09-08 三菱電機株式会社 位置検出機能つき液晶表示装置
JP2002196874A (ja) * 2000-12-27 2002-07-12 Ntt Docomo Inc 手書きデータ入力装置及び方法並びに個人認証装置及び方法
JP4357761B2 (ja) * 2001-03-23 2009-11-04 株式会社リコー 光学式座標入力装置

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4868551A (en) 1983-10-28 1989-09-19 Thomson-Csf Sensitive display device comprising a scanned screen
US4782328A (en) 1986-10-02 1988-11-01 Product Development Services, Incorporated Ambient-light-responsive touch screen data input method and system
US4811004A (en) * 1987-05-11 1989-03-07 Dale Electronics, Inc. Touch panel system and method for using same
US5164714A (en) 1988-06-20 1992-11-17 Amp Incorporated Modulated touch entry system and method with synchronous detection
US5534917A (en) 1991-05-09 1996-07-09 Very Vivid, Inc. Video image based control system
US5317140A (en) 1992-11-24 1994-05-31 Dunthorn David I Diffusion-assisted position location particularly for visual pen detection
EP0897161A1 (fr) * 1997-08-07 1999-02-17 Fujitsu Limited Panneau tactile optique à balayage
US6352351B1 (en) * 1999-06-30 2002-03-05 Ricoh Company, Ltd. Method and apparatus for inputting coordinates
US6504532B1 (en) * 1999-07-15 2003-01-07 Ricoh Company, Ltd. Coordinates detection apparatus
US20010026268A1 (en) * 2000-03-31 2001-10-04 Ricoh Company, Ltd. Coordiante input and detection device and information display and input apparatus
WO2001091043A2 (fr) 2000-05-24 2001-11-29 Next Holdings Limited Systeme d'affichage a ecran tactile

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP1599789A4 *

Cited By (49)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8055022B2 (en) 2000-07-05 2011-11-08 Smart Technologies Ulc Passive touch system and method of detecting user input
US8203535B2 (en) 2000-07-05 2012-06-19 Smart Technologies Ulc Passive touch system and method of detecting user input
US8378986B2 (en) 2000-07-05 2013-02-19 Smart Technologies Ulc Passive touch system and method of detecting user input
US8228304B2 (en) 2002-11-15 2012-07-24 Smart Technologies Ulc Size/scale orientation determination of a pointer in a camera-based touch system
US8466885B2 (en) 2003-02-14 2013-06-18 Next Holdings Limited Touch screen signal processing
US8508508B2 (en) 2003-02-14 2013-08-13 Next Holdings Limited Touch screen signal processing with single-point calibration
US8289299B2 (en) 2003-02-14 2012-10-16 Next Holdings Limited Touch screen signal processing
US8456447B2 (en) 2003-02-14 2013-06-04 Next Holdings Limited Touch screen signal processing
US8456451B2 (en) 2003-03-11 2013-06-04 Smart Technologies Ulc System and method for differentiating between pointers used to contact touch surface
US8325134B2 (en) 2003-09-16 2012-12-04 Smart Technologies Ulc Gesture recognition method and touch system incorporating the same
US8456418B2 (en) 2003-10-09 2013-06-04 Smart Technologies Ulc Apparatus for determining the location of a pointer within a region of interest
US8089462B2 (en) 2004-01-02 2012-01-03 Smart Technologies Ulc Pointer tracking across multiple overlapping coordinate input sub-regions defining a generally contiguous input region
US8576172B2 (en) 2004-01-02 2013-11-05 Smart Technologies Ulc Pointer tracking across multiple overlapping coordinate input sub-regions defining a generally contiguous input region
US8274496B2 (en) 2004-04-29 2012-09-25 Smart Technologies Ulc Dual mode touch systems
US8120596B2 (en) 2004-05-21 2012-02-21 Smart Technologies Ulc Tiled touch system
CN100370403C (zh) * 2004-10-29 2008-02-20 索尼株式会社 输入/输出装置和方法
US9442607B2 (en) 2006-12-04 2016-09-13 Smart Technologies Inc. Interactive input system and method
US8115753B2 (en) 2007-04-11 2012-02-14 Next Holdings Limited Touch screen system with hover and click input methods
WO2008128096A3 (fr) * 2007-04-11 2009-04-16 Next Holdings Inc Système à écran tactile avec procédés de saisie par effleurement et clic
US8094137B2 (en) 2007-07-23 2012-01-10 Smart Technologies Ulc System and method of detecting contact on a display
US8432377B2 (en) 2007-08-30 2013-04-30 Next Holdings Limited Optical touchscreen with improved illumination
US8384693B2 (en) 2007-08-30 2013-02-26 Next Holdings Limited Low profile touch panel systems
US9171454B2 (en) 2007-11-14 2015-10-27 Microsoft Technology Licensing, Llc Magic wand
US8405636B2 (en) 2008-01-07 2013-03-26 Next Holdings Limited Optical position sensing system and optical position sensor assembly
US8405637B2 (en) 2008-01-07 2013-03-26 Next Holdings Limited Optical position sensing system and optical position sensor assembly with convex imaging window
US8902193B2 (en) 2008-05-09 2014-12-02 Smart Technologies Ulc Interactive input system and bezel therefor
US8952894B2 (en) 2008-05-12 2015-02-10 Microsoft Technology Licensing, Llc Computer vision-based multi-touch sensing using infrared lasers
WO2009139971A3 (fr) * 2008-05-12 2010-01-07 Microsoft Corporation Détection tactile multipoint fondée sur la vision par ordinateur utilisant des lasers infrarouges
US8847739B2 (en) 2008-08-04 2014-09-30 Microsoft Corporation Fusing RFID and vision for surface object tracking
US8339378B2 (en) 2008-11-05 2012-12-25 Smart Technologies Ulc Interactive input system with multi-angle reflector
US10627973B2 (en) 2009-06-18 2020-04-21 Baanto International Ltd. Systems and sensors for sensing and tracking radiation blocking objects on a surface
CN102202172B (zh) * 2010-03-23 2016-03-09 北京汇冠新技术股份有限公司 用于带摄像头触摸屏的摄像头同步方法和系统
CN102202172A (zh) * 2010-03-23 2011-09-28 北京汇冠新技术股份有限公司 用于带摄像头触摸屏的摄像头同步方法和系统
WO2011117505A2 (fr) * 2010-03-25 2011-09-29 Intui Sense Dispositif interactif adaptable aux conditions de luminosite ambiante
WO2011117505A3 (fr) * 2010-03-25 2012-12-20 Intui Sense Dispositif interactif adaptable aux conditions de luminosite ambiante
CN102281395A (zh) * 2010-06-08 2011-12-14 北京汇冠新技术股份有限公司 摄像头同步方法、控制板、触摸屏、触摸系统和显示器
CN102281395B (zh) * 2010-06-08 2014-09-03 北京汇冠新技术股份有限公司 触摸屏、触摸系统和显示器
US9582116B2 (en) 2010-09-02 2017-02-28 Baanto International Ltd. Systems and methods for sensing and tracking radiation blocking objects on a surface
DE112011103849T5 (de) 2010-11-22 2013-10-02 Epson Norway Research And Development As Multitouch-Wechselwirkungs- und Beleuchtungssystem und -verfahren auf Kamerabasis
CN103329079A (zh) * 2010-11-22 2013-09-25 爱普生挪威研究和发展公司 基于相机的多点触摸交互以及照明系统和方法
US9996197B2 (en) 2010-11-22 2018-06-12 Seiko Epson Corporation Camera-based multi-touch interaction and illumination system and method
CN103329079B (zh) * 2010-11-22 2016-08-24 爱普生挪威研究和发展公司 基于相机的多点触摸交互以及照明系统和方法
WO2012070950A1 (fr) * 2010-11-22 2012-05-31 Epson Norway Research And Development As Système et procédé d'interaction à touchers multiples et d'éclairage à base de caméra
GB2487043B (en) * 2010-12-14 2013-08-14 Epson Norway Res And Dev As Camera-based multi-touch interaction and illumination system and method
GB2487043A (en) * 2010-12-14 2012-07-11 New Index As Determining touch or hover behaviour of an object interacting with a touch screen, using constant thickness light beams
US9168865B2 (en) 2011-10-11 2015-10-27 Hyundai Motor Company Organic electro-luminescence lighting device for vehicle
WO2013104061A1 (fr) * 2012-01-11 2013-07-18 Smart Technologies Ulc Étalonnage d'un rideau lumineux interactif
US9207812B2 (en) 2012-01-11 2015-12-08 Smart Technologies Ulc Interactive input system and method
US9880668B2 (en) 2013-09-11 2018-01-30 Beijing Lenovo Software Ltd. Method for identifying input information, apparatus for identifying input information and electronic device

Also Published As

Publication number Publication date
CA2515955A1 (fr) 2004-08-26
AU2004211738B2 (en) 2007-04-19
CN1774692A (zh) 2006-05-17
CA2515955C (fr) 2011-01-11
KR101035253B1 (ko) 2011-05-19
JP2006518076A (ja) 2006-08-03
CN100468303C (zh) 2009-03-11
JP4668897B2 (ja) 2011-04-13
KR20050111324A (ko) 2005-11-24
EP1599789A4 (fr) 2010-03-31
AU2004211738A1 (en) 2004-08-26
EP1599789A1 (fr) 2005-11-30

Similar Documents

Publication Publication Date Title
CA2515955C (fr) Traitement du signal d'un ecran tactile
US8466885B2 (en) Touch screen signal processing
US8456447B2 (en) Touch screen signal processing
US8508508B2 (en) Touch screen signal processing with single-point calibration
US20110199335A1 (en) Determining a Position of an Object Using a Single Camera
CA2848650C (fr) Systemes d'ecran tactile optique utilisant la lumiere reflechie
US8847924B2 (en) Reflecting light
EP2353069B1 (fr) Capteurs optiques stéréos pour résoudre un toucher multiple dans un système de détection de toucher
CN101663637B (zh) 利用悬浮和点击输入法的触摸屏系统
US20110090147A1 (en) Touchless pointing device
US20070018970A1 (en) Optical slider for input devices
US20110241984A1 (en) Illumination structure for an interactive input system
KR20110013459A (ko) 제어된 조광을 갖춘 대화형 입력 시스템
JP2013522801A (ja) 光ベースのタッチスクリーンのためのレンズ配列体
JP2012508913A (ja) 一体型タッチセンシングディスプレー装置およびその製造方法
JP2011043986A (ja) 光学式情報入力装置、光学式入力機能付き電子機器、および光学式情報入力方法
EP0592419A1 (fr) Dispositif d'affichage detectant des positions.
JP2011090604A (ja) 光学式位置検出装置および位置検出機能付き表示装置
US20110095977A1 (en) Interactive input system incorporating multi-angle reflecting structure
JP2006163751A (ja) 位置検出平面外に撮像部を有する光学式位置検出装置
JPH01150192A (ja) タッチ入力装置
KR101808523B1 (ko) 광학식 터치 입력 장치 및 이의 구동 방법
AU2013267005A1 (en) Optical touch screen systems using reflected light

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): BW GH GM KE LS MW MZ SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: 11033183

Country of ref document: US

WWE Wipo information: entry into national phase

Ref document number: 2515955

Country of ref document: CA

Ref document number: 2006502767

Country of ref document: JP

Ref document number: 1020057014885

Country of ref document: KR

WWE Wipo information: entry into national phase

Ref document number: 2004211738

Country of ref document: AU

WWE Wipo information: entry into national phase

Ref document number: 2004711522

Country of ref document: EP

ENP Entry into the national phase

Ref document number: 2004211738

Country of ref document: AU

Date of ref document: 20040216

Kind code of ref document: A

WWP Wipo information: published in national office

Ref document number: 2004211738

Country of ref document: AU

WWE Wipo information: entry into national phase

Ref document number: 20048099261

Country of ref document: CN

WWP Wipo information: published in national office

Ref document number: 1020057014885

Country of ref document: KR

WWP Wipo information: published in national office

Ref document number: 2004711522

Country of ref document: EP

DPEN Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed from 20040101)
WWG Wipo information: grant in national office

Ref document number: 2004211738

Country of ref document: AU