WO2013040691A1 - Interactive input system with variable reflector pen tool - Google Patents

Interactive input system with variable reflector pen tool Download PDF

Info

Publication number
WO2013040691A1
WO2013040691A1 PCT/CA2012/000882 CA2012000882W WO2013040691A1 WO 2013040691 A1 WO2013040691 A1 WO 2013040691A1 CA 2012000882 W CA2012000882 W CA 2012000882W WO 2013040691 A1 WO2013040691 A1 WO 2013040691A1
Authority
WO
WIPO (PCT)
Prior art keywords
pointer
pen tool
input system
interactive input
image frames
Prior art date
Application number
PCT/CA2012/000882
Other languages
French (fr)
Inventor
Technologies Ulc Smart
Sean Thompson
Grant Mcgibney
Gerald D. MORRISON
Original Assignee
Technologies Ulc Smart
Sean Thompson
Grant Mcgibney
Morrison Gerald D
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Technologies Ulc Smart, Sean Thompson, Grant Mcgibney, Morrison Gerald D filed Critical Technologies Ulc Smart
Publication of WO2013040691A1 publication Critical patent/WO2013040691A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0354Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
    • G06F3/03545Pens or stylus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0354Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
    • G06F3/03543Mice or pucks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/042Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means
    • G06F3/0421Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means by interrupting or reflecting a light beam, e.g. optical touch-screen

Definitions

  • the present invention relates to an interactive input system and to a pen tool therefor.
  • Interactive input systems that allow users to inject input into an application program using an active pointer (e.g. a pointer that emits light, sound or other signal), a passive pointer (e.g. a finger, cylinder or other object) or other suitable input device such as for example, a mouse or trackball, are well known.
  • active pointer e.g. a pointer that emits light, sound or other signal
  • passive pointer e.g. a finger, cylinder or other object
  • suitable input device such as for example, a mouse or trackball
  • touch systems comprising touch panels employing electromagnetic, capacitive, acoustic or other technologies to register pointer input; tablet and laptop personal computers (PCs); personal digital assistants (PDAs) and other handheld devices; and other similar devices.
  • U.S. Patent No. 6,803,906 to Morrison et al. discloses a touch system that employs machine vision to detect pointer interaction with a touch surface on which a computer-generated image is presented.
  • a rectangular bezel or frame surrounds the touch surface and supports digital cameras at its corners.
  • the digital cameras have overlapping fields of view that encompass and look generally across the touch surface.
  • the digital cameras acquire images looking across the touch surface from different vantages and generate image data.
  • Image data acquired by the digital cameras is processed by on-board digital signal processors to determine if a pointer exists in the captured image data.
  • the digital signal processors convey pointer characteristic data to a master controller, which in turn processes the pointer characteristic data to determine the location of the pointer in (x,y) coordinates relative to the touch surface using triangulation.
  • the pointer coordinates are conveyed to a computer executing one or more application programs.
  • the computer uses the pointer coordinates to update the computer-generated image that is presented on the touch surface. Pointer contacts on the touch surface can therefore be recorded as writing or drawing or used to control execution of application programs executed by the computer.
  • the illuminated bezel comprises infrared (IR) light emitting diodes (LEDs) that project infrared light onto diffusers.
  • IR infrared
  • LEDs light emitting diodes
  • diffusers diffuse the infrared light so that the intensity of backlighting provided over the touch surface by the illuminated bezel is generally even across the surfaces of the diffusers.
  • the backlight illumination provided by the bezel appears generally continuous to the digital cameras.
  • U.S. Patent No. 6,823,481 to Takekawa discloses a method and apparatus for preventing misdetection of coordinates of a pointing instrument.
  • a coordinate input/detection device compares a waveform of a signal corresponding to intensity distribution of rays received from a coordinate input area by a light-receiving unit and a standard waveform at fixed timing. If the waveform is not equal to the standard waveform, the coordinate input/detection device determines that an installation position of the light-receiving unit is off a correct position because of shock or vibration.
  • a new standard waveform is obtained and stored in a storage unit, thereby setting the waveform of the signal corresponding to the intensity distribution of rays received by the light-receiving unit located at an incorrect position as the standard waveform.
  • U.S. Patent Nos. 7,202,860 and 7,414,617 to Ogawa disclose a coordinate input device that includes a pair of cameras positioned in an upper left position and an upper right position of a display screen of a monitor lying close to a plane extending from the display screen of the monitor and views both a side face of an object in contact with a position on the display screen and a predetermined desktop coordinate detection area to capture the image of the object within the field of view.
  • the coordinate input device also includes a control circuit which calculates the coordinate value of a pointing tool, pointing to a position within a coordinate detection field, based on video signals output from the pair of cameras, and transfers the coordinate value to a program of a computer.
  • U.S. Patent No. 6,567,078 to Ogawa discloses a handwriting communication system, a handwriting input device and a handwriting display device used in the system, which can communicate by handwriting among a plurality of computers connected via a network.
  • the communication system includes a handwriting input device which is provided at a transmitting side for inputting the handwriting into a transmitting side computer, and a handwriting display device which is provided at a receiving side for displaying the handwriting based on information transmitted from the transmitting side to a receiving side computer.
  • the system transmits only a contiguous image around the handwritten portion, which reduces the communication volume compared to transmitting the whole image, and which makes the real time transmission and reception of handwriting trace possible.
  • U.S. Patent No. 6,441 ,362 to Ogawa discloses an optical digitizer for determining a position of a pointing object projecting a light and being disposed on a coordinate plane.
  • a detector is disposed on a periphery of the coordinate plane and has a view field covering the coordinate plane for receiving the light projected from the pointing object and for converting the received light into an electric signal.
  • a processor is provided for processing the electric signal fed from the detector to compute coordinates representing the position of the pointing object.
  • a collimator is disposed to limit the view field of the detector below a predetermined height relative to the coordinate plane such that through the limited view field the detector can receive only a parallel component of the light which is projected from the pointing object substantially in parallel to the coordinate plane.
  • a shield is disposed to enclose the periphery of the coordinate plane to block noise light other than the projected light from entering into the limited view field of the detector.
  • Improvements are generally desired. It is therefore an object of the present invention at least to provide a novel interactive input system and a pen tool therefor.
  • a pen tool for use with a machine vision interactive input system comprising an elongate body; a tip on the body; and a variable reflector disposed on the body, the variable reflector comprising a reflecting portion and an attenuating portion configured to control the amount of illumination reflected by the reflecting portion.
  • variable reflector is positioned adjacent the tip.
  • the attenuating portion comprises a liquid crystal device.
  • the liquid crystal device is a polymer dispersed liquid crystal device comprising a polymer dispersed liquid crystal film positioned in an annular region between coaxial electrodes.
  • the transparency of the attenuating element is proportional to an applied voltage and the variable reflector further may comprise a protective outer portion.
  • the pen tool further comprises a microcontroller configured to switch the attenuating portion between differing levels of transparency.
  • the pen tool further comprises an actuator protruding from the tip and being in communication with the controller.
  • the controller is configured to switch the attenuating portion between two or more levels of transparency in response to actuation of the actuator.
  • the pen tool further comprises a switch arrangement in communication with the controller for selecting the differing levels of transparency.
  • the reflecting portion comprises a retro-reflective material.
  • the attenuating portion covers only a portion of the retro-reflective material.
  • an interactive input system comprising at least one imaging assembly having a field of view looking into a region of interest and capturing image frames; at least one illumination source configured to emit illumination into said region of interest; and processing structure in communication with the at least one imaging assembly, when a pointer exists in captured image frames, said processing structure demodulating the captured image frames to determine frequency components thereof and examining the frequency components to determine at least one attribute of said pointer, the frequency components being representative of illumination reflected by said pointer.
  • the processing structure during demodulating the processing structure applies a transform to the captured image frames.
  • the illumination is infrared illumination.
  • the attribute comprises one of pointer input colour, pointer input line thickness and pointer functionality.
  • the pointer functionality may be one of right mouse click, left mouse click, and eraser.
  • the at least one imaging assembly captures a sequence of image frames, each sequence comprising one image frame captured when the at least one illumination source is in an off state and a plurality of image frames captured when the at least one illumination source is in an on state.
  • the processing structure may subtract the image frame captured when the at least one illumination source is in the off state from at least one image frame captured when at least one illumination source is in the on state to form a difference image frame, and determines a location of the pointer in the difference image frame.
  • the processing structure further defines an analysis region associated with the pointer location, and applies the analysis region to the captured image frames for carrying out the examining.
  • a method of inputting information into an interactive input system comprising emitting illumination into a region of interest from at least one illumination source; capturing image frames of the region of interest; when a pointer exists in captured image frames, demodulating the captured image frames to determine frequency components thereof and examining the frequency components to determine at least one attribute of the pointer, the frequency components being representative of illumination reflected by said pointer.
  • Figure 1 is a schematic perspective view of an interactive input system
  • Figure 2 is a schematic block diagram view of the interactive input system of Figure 1 ;
  • Figure 3 is a block diagram of an imaging assembly forming part of the interactive input system of Figure 1 ;
  • Figures 4A and 4B are front and rear perspective views, respectively, of a housing assembly forming part of the imaging assembly of Figure 3;
  • Figure 5 is a block diagram of a master controller forming part of the interactive input system of Figure 1 ;
  • Figure 6 is a perspective view of a pen tool for use with the interactive input system of Figure 1;
  • Figure 7 is a cross-sectional view of a portion of the pen tool of Figure
  • Figure 8 is cross-sectional view of the tip portion of the pen tool of
  • Figure 9 is a block diagram of components of the pen tool of Figure 6;
  • Figure 10 is a graphical plot of an image frame capture sequence used by the interactive input system of Figure 1 ;
  • Figures 11 A and 1 IB depict a sequence of nine (9) consecutive image frames captured by the imaging assembly of Figure 3 when the pen tool of Figure 6 is in the field of view of the imaging assembly and is reflecting illumination modulated according to a subcarrier frequency combination;
  • Figure 12 is a flowchart showing steps of an image processing method
  • Figure 13 is a graphical plot of the light intensity variation of the image frames of Figure 11A and 1 IB;
  • Figure 14 is a graphical plot of the frequency domain representation of the image frame light intensity variation of Figure 13;
  • Figures 15 A and 15B show frequency domain representations of the light intensity of eight (8) subcarrier frequency combinations
  • Figures 16A to 16C are perspective views of an input surface of the interactive input system of Figure 1, showing a pressure-sensitive pen tool applying light pressure input, medium pressure input, and heavy pressure input, respectively;
  • Figures 17A to 17C show frequency domain representations of light intensity corresponding to the light, medium and heavy pressure inputs of Figures 16A to 16C, respectively;
  • Figure 18 is a graphical plot of another embodiment of an image frame capture sequence used by the interactive input system of Figure 1 ;
  • Figure 19 is another embodiment of an interactive input system
  • Figure 20 is a schematic plan view of an imaging assembly
  • Figure 21 is a graphical plot of an image frame capture sequence used by the interactive input system of Figure 19;
  • Figure 22 is a schematic plan view of another embodiment of an imaging assembly arrangement for the interactive input system of Figure 19;
  • Figure 23 is another embodiment of an interactive input system
  • Figure 24 is another embodiment of an interactive input system
  • Figure 25 is a partial cross-sectional view of a portion of another embodiment of a pen tool;
  • Figure 26 is a perspective view of another embodiment of a pen tool;
  • Figures 27A to 27E are perspective views of the pen tool of Figure 26 applying input in accordance with various different selected attributes
  • Figure 28 is a perspective view of another embodiment of a pen tool
  • Figure 29 is a perspective view of another embodiment of a pen tool
  • Figure 30 is a perspective view of another embodiment of a pen tool
  • Figure 31 is a partial cross-sectional view of another embodiment of a pen tool.
  • Figure 32 is a perspective view of another embodiment of a pen tool.
  • interactive input system 20 that allows a user to inject input such as digital ink, mouse events etc. into an executing application program is shown and is generally identified by reference numeral 20.
  • interactive input system 20 comprises an interactive board 22 mounted on a vertical support surface such as for example, a wall surface or the like.
  • Interactive board 22 comprises a generally planar, rectangular interactive surface 24 that is surrounded about its periphery by a bezel 26.
  • An ultra-short throw projector (not shown) such as that sold by SMART Technologies ULC under the name SMART UX60 is also mounted on the support surface above the interactive board 22 and projects an image, such as for example a computer desktop, onto the interactive surface 24.
  • the interactive board 22 employs machine vision to detect one or more pointers brought into a region of interest in proximity with the interactive surface 24.
  • the interactive board 22 communicates with a general purpose computing device 28 executing one or more application programs via a universal serial bus (USB) cable 30 or other suitable wired or wireless connection.
  • General purpose computing device 28 processes the output of the interactive board 22 and, if required, adjusts image data being output to the projector so that the image presented on the interactive surface 24 reflects pointer activity.
  • the interactive board 22, general purpose computing device 28 and projector allow pointer activity proximate to the interactive surface 24 to be recorded as writing or drawing or used to control execution of one or more application programs executed by the general purpose computing device 28.
  • the bezel 26 in this embodiment is mechanically fastened to the interactive surface 24 and comprises four bezel segments 40, 42, 44, 46.
  • Bezel segments 40 and 42 extend along opposite side edges of the interactive surface 24 while bezel segments 44 and 46 extend along the top and bottom edges of the interactive surface 24 respectively.
  • the inwardly facing surface of each bezel segment 40, 42, 44 and 46 comprises a single, longitudinally extending strip or band of retro- reflective material.
  • the bezel segments 40, 42, 44 and 46 are oriented so that their inwardly facing surfaces extend in a plane generally normal to the plane of the interactive surface 24.
  • the tool tray 48 comprises a housing 48a having an upper surface 48b configured to define a plurality of receptacles or slots 48c.
  • the receptacles 48c are sized to receive one or more pen tools P as will be described as well as an eraser tool (not shown) that can be used to interact with the interactive surface 24.
  • Control buttons 48d are provided on the upper surface 48b of the housing 48a to enable a user to control operation of the interactive input system 20.
  • One end of the tool tray 48 is configured to receive a detachable tool tray accessory module 48e while the opposite end of the tool tray 48 is configured to receive a detachable communications module 48f for remote device communications.
  • the housing 48a accommodates a master controller 50 (see Figure 5) as will be described.
  • imaging assemblies 60 are accommodated by the bezel 26, with each imaging assembly 60 being positioned adjacent a different corner of the bezel.
  • the imaging assemblies 60 are oriented so that their fields of view overlap and look generally across the entire interactive surface 24.
  • any pointer such as for example a user's finger, a cylinder or other suitable object, or a pen tool or eraser tool lifted from a receptacle 48c of the tool tray 48, that is brought into proximity of the interactive surface 24 appears in the fields of view of the imaging assemblies 60.
  • a power adapter 62 provides the necessary operating power to the interactive board 22 when connected to a conventional AC mains power supply.
  • the imaging assembly 60 comprises an image sensor 70 such as that manufactured by Aptina (Micron) under Model No. MT9V034 having a resolution of 752x480 pixels, fitted with a two element, plastic lens (not shown) that provides the image sensor 70 with a field of view of approximately 104 degrees.
  • the other imaging assemblies 60 are within the field of view of the image sensor 70 thereby to ensure that the field of view of the image sensor 70 encompasses the entire interactive surface 24.
  • a digital signal processor (DSP) 72 such as that manufactured by Analog
  • a serial peripheral interface (SPI) flash memory 74 is connected to the DSP 72 via an SPI port and stores the firmware required for image assembly operation.
  • the imaging assembly 60 may optionally comprise synchronous dynamic random access memory (SDRAM) 76 to store additional temporary data as shown by the dotted lines.
  • SDRAM synchronous dynamic random access memory
  • the image sensor 70 also communicates with the DSP 72 via a two-wire interface (TWI) and a timer (TMR) interface. The control registers of the image sensor 70 are written from the DSP 72 via the TWI in order to configure parameters of the image sensor 70 such as the integration period for the image sensor 70.
  • the image sensor 70 operates in snapshot mode.
  • the image sensor 70 in response to an external trigger signal received from the DSP 72 via the TMR interface that has a duration set by a timer on the DSP 72, enters an integration period during which an image frame is captured.
  • the image sensor 70 enters a readout period during which time the captured image frame is available. With the image sensor in the readout period, the DSP 72 reads the image frame data acquired by the image sensor 70 over the image data bus 71 via the PPI.
  • the frame rate of the image sensor 70 in this embodiment is between about 900 and about 960 frames per second.
  • the DSP 72 in turn processes image frames received from the image sensor 70 and provides pointer information to the master controller 50 at a reduced rate of approximately 120 points/sec.
  • Those of skill in the art will however appreciate that other frame rates may be employed depending on the desired accuracy of pointer tracking and whether multi-touch and/or active pointer identification is employed.
  • Three strobe circuits 80 communicate with the DSP 72 via the TWI and via a general purpose input/output (GPIO) interface.
  • the IR strobe circuits 80 also communicate with the image sensor 70 and receive power provided on LED power line 82 via the power adapter 62.
  • Each strobe circuit 80 drives a respective illumination source in the form of an infrared (IR) light emitting diode (LED) 84 that provides infrared backlighting over the interactive surface 24.
  • IR infrared
  • LED light emitting diode
  • the DSP 72 also communicates with an RS-422 transceiver 86 via a serial port (SPORT0) and a non-maskable interrupt (NMI) port.
  • the transceiver 86 communicates with the master controller 50 over a differential synchronous signal (DSS) communications link 88 and a synch line 90.
  • Power for the components of the imaging assembly 60 is provided on power line 92 by the power adapter 62.
  • DSP 72 may also optionally be connected to a USB connector 94 via a USB port as indicated by the dotted lines.
  • the USB connector 94 can be used to connect the imaging assembly 60 to diagnostic equipment.
  • the image sensor 70 and its associated lens as well as the IR LEDs 84 are mounted on a housing assembly 100 that is shown in Figures 4A and 4B.
  • the housing assembly 100 comprises a polycarbonate housing body 102 having a front portion 104 and a rear portion 106 extending from the front portion.
  • An imaging aperture 108 is centrally formed in the housing body 102 and accommodates an IR- pass/visible light blocking filter 110.
  • the filter 110 has an IR-pass wavelength range of between about 830nm and about 880nm.
  • the image sensor 70 and associated lens are positioned behind the filter 110 and oriented such that the field of view of the image sensor 70 looks through the filter 110 and generally across the interactive surface 24.
  • the rear portion 106 is shaped to surround the image sensor 70.
  • Three passages 112a to 112c are formed through the housing body 102.
  • Passages 1 12a and 112b are positioned on opposite sides of the filter 1 10 and are in general horizontal alignment with the image sensor 70.
  • Passage 112c is centrally positioned above the filter 1 10.
  • Each tubular passage receives a light source socket 114 that is configured to receive a respective one of the IR LEDs 84.
  • Mounting flanges 116 are provided on opposite sides of the rear portion 106 to facilitate connection of the housing assembly 100 to the bezel 26 via suitable fasteners.
  • a label 118 formed of retro-reflective material overlies the front surface of the front portion 104. Further specifics concerning the housing assembly and its method of manufacture are described in U.S. Application Publication No.
  • master controller 50 comprises a DSP 200 such as that manufactured by Analog Devices under part number ADSP-BF522 Blackfin or other suitable processing device.
  • a serial peripheral interface (SPI) flash memory 202 is connected to the DSP 200 via an SPI port and stores the firmware required for master controller operation.
  • SPI serial peripheral interface
  • SDRAM synchronous dynamic random access memory
  • the DSP 200 communicates with the general purpose computing device 28 over the USB cable 30 via a USB port.
  • the DSP 200 communicates through its serial port (SPORT0) with the imaging assemblies 60 via an RS-422 transceiver 208 over the differential synchronous signal (DSS) communications link 88.
  • SPORT0 serial port
  • DSS differential synchronous signal
  • TDM time division multiplexed
  • the DSP 200 also communicates with the imaging assemblies 60 via the RS- 422 transceiver 208 over the camera synch line 90.
  • DSP 200 communicates with the tool tray accessory module 48e over an inter-integrated circuit (I 2 C) channel and communicates with the communications module 48f over universal asynchronous receiver/transmitter (UART), serial peripheral interface (SPI) and I C channels.
  • I 2 C inter-integrated circuit
  • UART universal asynchronous receiver/transmitter
  • SPI serial peripheral interface
  • the architectures of the imaging assemblies 60 and master controller 50 are similar. By providing a similar architecture between each imaging assembly 60 and the master controller 50, the same circuit board assembly and common components may be used for both thus reducing the part count and cost of the interactive input system 20. Differing components are added to the circuit board assemblies during manufacture dependent upon whether the circuit board assembly is intended for use in an imaging assembly 60 or in the master controller 50. For example, the master controller 50 may require a SDRAM 76 whereas the imaging assembly 60 may not.
  • the general purpose computing device 28 in this embodiment is a personal computer or other suitable processing device comprising, for example, a processing unit, system memory (volatile and/or non-volatile memory), other nonremovable or removable memory (e.g. a hard disk drive, RAM, ROM, EEPROM, CD- ROM, DVD, flash memory, etc.) and a system bus coupling the various computer components to the processing unit.
  • the computing device 28 may also comprise a network connection to access shared or remote drives, one or more networked computers, or other networked devices.
  • FIGS 6 to 9 show better illustrate a pen tool 220 for use with the interactive input system 20.
  • Pen tool 220 has a main body 222 terminating in a conical tip 224.
  • a variable reflector 226 is disposed on the body 222 adjacent the tip 224.
  • Variable reflector 226 has a multilayered structure, and comprises an inner layer 234 comprising a retro-reflective material, an attenuating layer 236 disposed on the inner layer 234, and an outer layer 238 disposed on the attenuating layer 236.
  • the attenuating layer 236 is configured to have an electrically controllable transparency, such that the transmittance of light therethrough is controllable.
  • the attenuating layer 236 is a polymer dispersed liquid crystal (PDLC) device, and comprises a PDLC film 240 positioned in the annular region between two co-axial, transparent electrodes 242 and 244.
  • the PDLC film 240 comprises a plurality of droplets of liquid crystals dispersed in a polymer matrix.
  • the liquid crystals within the droplets assume random orientations, causing the droplets to scatter light when it reaches the PDLC film 240.
  • the scattering of light reduces the amount of light than can pass straight through the PDLC film 240, thus reducing its transparency.
  • the liquid crystals within the droplets become partially aligned, causing fewer of the droplets to scatter light passing through the PDLC film 240.
  • the reduced scattering of light increases the amount of light that can pass straight through the PDLC film 240, thus increasing its transparency.
  • the degree of alignment of the droplets within the PDLC film 240, and consequently the level of transparency of the attenuating layer 236, is generally proportional to the applied voltage. Accordingly, the transparency of the attenuating layer 236 may be varied over a range corresponding to the working voltage range of the pen tool 220.
  • the variable reflector 226 is thus capable of controlling the amount of incident light that reaches the inner layer 234 and also the amount of light reflected from the inner layer 234.
  • the outer layer 238 is fabricated of a durable material that is transparent to infrared light, and functions as a protective barrier for the variable reflector 226.
  • the outer layer 238 is fabricated of plastic.
  • Protruding from the tip 224 is an actuator 225 that resembles a nib.
  • Actuator 225 is biased out of the tip 224 by a spring (not shown) but can be pushed into the tip against the bias of the spring upon application of pressure thereto.
  • Actuator 225 is connected to a microcontroller 250 housed within the main body 222. As shown in Figure 9, microcontroller 250 is in communication with the attenuating layer 236. When the actuator 225 is depressed, microcontroller 250 is configured to switch the attenuating layer 236 between levels of transparency at one or more modulating frequencies by applying alternating voltage levels to the electrodes using power supplied by a battery 254. This switching of the attenuating layer 236 modulates the amount of IR
  • the modulation of the amount of reflected IR illumination is used to identify the pen tool 220 although those of skill in the art will appreciate that the modulation may be performed for other purposes.
  • the DSP 200 of the master controller 50 outputs synchronization signals that are applied to the synch line 90 via the transceiver 208.
  • Each synchronization signal applied to the synch line 90 is received by the DSP 72 of each imaging assembly 60 via transceiver 86 and triggers a non-maskable interrupt (NMI) on the DSP 72.
  • NMI non-maskable interrupt
  • the DSP 72 of each imaging assembly 60 ensures that its local timers are within system tolerances and if not, corrects its local timers to match the master controller 50.
  • the DSP 72 Using one local timer, the DSP 72 initiates a pulse sequence via the snapshot line that is used to condition the image sensor to the snapshot mode and to control the integration period and frame rate of the image sensor 70 in the snapshot mode.
  • the DSP 72 also initiates a second local timer that is used to provide output on the LED control line 174 so that the IR LEDs 84 are properly powered during the image frame capture cycle.
  • the pulse sequences and the outputs on the LED control line 174 are generated so that the frame capture rate of each image sensor 70 is eight (8) times the desired image frame output rate.
  • the image sensor 70 of each imaging assembly 60 acquires image frames at the desired image frame rate. In this manner, image frames captured by the image sensor 70 of each imaging assembly can be referenced to the same point of time allowing the position of pointers brought into the fields of view of the image sensors 70 to be accurately triangulated.
  • Each imaging assembly 60 has its own local oscillator (not shown) and synchronization signals are distributed so that a lower frequency synchronization signal (e.g. the point rate, 120Hz) for each imaging assembly 60 is used to keep image frame capture synchronized. By distributing the synchronization signals for the imaging assemblies 60, rather than, transmitting a fast clock signal to each image assembly 60 from a central location, electromagnetic interference is reduced.
  • each IR LED 84 when each IR LED 84 is on, the IR LEDs flood the region of interest over the interactive surface 24 with infrared illumination. Infrared illumination that impinges on the retro-reflective bands of bezel segments 40, 42, 44 and 46 and on the retro-reflective labels 118 of the housing assemblies 100 is returned to the imaging assemblies 60. As a result, in the absence of a pointer, the image sensor 70 of each imaging assembly 60 sees a bright band having a substantially even intensity over its length, together with any ambient light artifacts.
  • the pointer When a pointer is brought into proximity with the interactive surface 24, the pointer occludes infrared illumination reflected by the retro-reflective bands of bezel segments 40, 42, 44 and 46 and/or the retro-reflective labels 118. As a result, the image sensor 70 of each imaging assembly 60 sees a dark region that interrupts the bright band in captured image frames. If the pointer is a pen tool 220, the image sensor 70 of each imaging assembly 60 also sees a bright region above the bright band corresponding to reflection of infrared illumination from the variable reflector 226. Additionally, the reflections of the illuminated retro-reflective bands of bezel segments 40, 42, 44 and 46 and the illuminated retro-reflective labels 118 appearing on the interactive surface 24 are also visible to the image sensor 70.
  • each imaging assembly 60 does not see the retro-reflective bands or the retro- reflective labels 118. If a pen tool 220 is in proximity with the display surface 24, no infrared illumination impinges on the variable reflector 226 of the pen tool 220 and consequently the image sensor 70 also does not see the variable reflector 226. Each imaging assembly 60 does however see artifacts resulting from ambient light on a dark background.
  • Figure 10 shows a portion of the image frame capture sequence used by the interactive input system 20, which is generally indicated using reference numeral 260.
  • each imaging assembly 60 continuously captures image frames that are grouped into image frame sequences and stored in buffers, with each image frame sequence comprising nine (9) image frames.
  • Figures 1 1 A and 1 IB show a sequence of image frames captured by one of the imaging assemblies 60 during the image frame capture sequence 260 when a pen tool is proximate the interactive surface 24.
  • the first image frame (“Frame #1”) of each sequence is captured with the IR LEDs 84 off so as to obtain a background image frame, and the following eight (8) image frames (Frame #2 to Frame #9) are captured with the IR LEDs 84 on so as to obtain eight (8) illuminated image frames.
  • the exposure of the image sensors 70 of the four (4) imaging assemblies 60 are staggered such that only one image frame is captured at one time.
  • the powering of the IR LEDs are also staggered during capture of the illuminated image frames such that each group of IR LEDs 84 is synchronized with its respective image sensor 70.
  • the cycle rate used by the four (4) image sensors 70 to capture a single image frame is 960 Hz, yielding an exposure time of 1/4 x 1/960 Hz, or about 260 Ds for each image sensor 70.
  • each DSP 72 subjects the image frames in the sequence to an image frame processing method, which is generally shown in Figure 12 and identified by reference numeral 270.
  • the background image frame is subtracted from the first illuminated image frame (i.e. Frame #2) so as to yield a difference image frame having ambient light removed (step 272).
  • the ambient light removed in this step comprises light originating from the operating environment surrounding the interactive input system 20, and infrared illumination emitted by the IR LEDs 84 that is scattered off of objects proximate to the imaging assemblies 60.
  • the difference image frame is then examined for values that represent the bezel and possibly one or more pointers (step 274).
  • Methods for determining pointer location within image frames have been described in U.S. Patent Application Publication No. 2009/0277697 to Bolt et al., entitled “Interactive Input System and Pen Tool Therefor", the disclosure of which is incorporated herein by reference in its entirety.
  • the locations of dark regions interrupting the bright band in the difference image frame are analyzed.
  • a pen tool 220 in proximity with the interactive surface 24 will give rise to a generally dark region that results from the occlusion of infrared illumination reflected from the retro-reflective band of the bezel segments by the tip 224.
  • the DSP 72 then defines one or more square-shaped pointer analysis regions that are generally centered on the area directly above the bezel at each pointer location (step 276). Image data within the pointer analysis region is used for determining if pointer information presented by a subcarrier frequency combination has been received. If the pointer is a pen tool 220, the pointer analysis region will encompass at least a portion of the region corresponding to the variable reflector 226 and will have a bright appearance as a result of the IR illumination reflected therefrom.
  • the transparency of the attenuating layer 236 of each pen tool 220, and therefore the intensity of IR illumination reflected by the variable reflector 226, is varied using a combination of three subcarrier frequencies, namely 120Hz, 240Hz, and 360Hz.
  • the transparency of the attenuating layer 236 is varied sinusoidally at each subcarrier frequency used.
  • the DSP 72 applies the one or more pointer analysis regions to each of the eight (8) illuminated image frames stored in the buffer (step 278).
  • the DSP 72 then carries out a
  • demodulation of the illuminated image frames (step 280) and measures the total light intensity within the pointer analysis region(s) for each illuminated image frame. This measured total light intensity is dealt with as a function of image frame number and gives rise to a light intensity variation. An example of such a light intensity variation is graphically plotted in Figure 13.
  • the DSP 72 then performs a time frequency transform on the light intensity variation to yield a frequency domain representation of light intensity (step 282), which is plotted graphically in Figure 14.
  • the transform is a discrete Fourier transform, but those of skill in the art will understand that any transform from the time domain to the frequency domain may be used.
  • an eight point discrete Fourier transform will report frequencies of OHz, 120Hz, 240Hz, 360Hz, and 480Hz.
  • the measurement at OHz represents light from all constant light sources, including ambient light and the bias applied to active pen tool IR illumination, and is discarded.
  • the measurement at 480Hz represents light from the oscillating IR LEDs 84 and is also discarded.
  • the remaining measurements at 120Hz, 240Hz, and 360Hz represent modulated subcarriers of IR illumination reflected by the pen tool 220.
  • the available carrier frequencies used by pen tools 220 are dependent on the frame rate (i.e. the frame capture rate) of the image sensors 70.
  • the frame rate is 960 frames per second. Therefore, to avoid aliasing at this frame rate, the upper limit of modulation frequency used for the attenuating layer 236 is 480Hz.
  • the three subcarrier frequencies, namely 120Hz, 240Hz, and 360Hz that are used by the pen tool 220 to modulate reflected IR illumination correspond to the normalized frequencies pi/4, pi/2 and 3pi/4 of the maximum modulation frequency (480Hz).
  • Each pen tool 220 may modulate the reflected IR light using a different combination of the three subcarrier frequencies.
  • the pen tool 220 operates each subcarrier frequency in either an "on" state or an "off state. This permits a total of eight (or 2", where n is the number of subcarrier frequencies available) different subcarrier frequency combinations each having a unique modulation pattern that can be recognized by the interactive input system 20.
  • the combination with all three subcarrier frequencies off is reserved by the interactive input system 20 for pointers that are not active pen tools, such as, for example a finger.
  • This provides seven other unique subcarrier frequency combinations for use with pen tools 220.
  • this information enables the identity of the active pen tool to be recognized by the interactive input system 20.
  • the identity information may be used to assign an attribute to the pen tool, such as pen colour, line thickness or functionality (e.g.
  • the frequency domain representations of each of these eight subcarrier frequency combinations are graphically plotted in Figures 15A and 15B.
  • the DSP 72 determines the identity of the pen tool 220 by analyzing the frequency domain representation of the light intensity variation and determining which frequencies are present above a certain threshold (step 284). If no frequencies are present in the frequency domain representation, as in the case of combination 1, then the interactive input system 20 determines that the pointer is not a pen tool 220.
  • each DSP 72 then conveys the pointer data representing the position of the pointer in the image frames as well as the pointer shape and contact status (actual contact or hover) to the DSP 200 of the master controller 50.
  • the DSP 200 uses the pointer data received from the DSPs 72 to calculate the position of each pointer relative to the interactive surface 24 in (x,y) coordinates using well known triangulation as described in above-incorporated U.S. Patent No. 6,803,906 to Morrison et al.
  • the pointer coordinate data along with pointer identity, pointer shape and contact status is conveyed to the general purpose computing device 28 allowing the image data presented on the interactive surface 24 to be updated.
  • additional information about the pen tool such as the level of pressure applied to the pen tool tip, may be communicated by the pen tool 220 by increasing both the number of subcarrier frequencies available for modulation as well as the frame rate of the imaging assemblies 60.
  • the amplitude of the subcarrier frequencies may be varied. This method can be used for the subcarrier frequency combinations that comprise more than one subcarrier frequency, namely combinations 5, 6, 7 and 8 in Figures 15A and 15B. Pen tools using these subcarrier frequency combinations may be encoded such that the level of pen tip pressure is manifested as a difference in amplitude of one or more subcarrier frequencies.
  • Figures 16A to 16C each schematically illustrate the motion of a pen tool 220 across the interactive surface 24, during which the pressure applied by the tip of the pen tool 220 to the interactive surface is light (Figure 16A), medium ( Figure 16B) and heavy ( Figure 16C).
  • the corresponding subcarrier frequency signal modulation is plotted graphically in Figures 17A to 17C.
  • the difference in amplitude of the subcarrier frequencies for different combinations is kept on the order of +/-25% to maintain the intensity of all subcarriers above the threshold for detection.
  • Pen tools using subcarrier frequency combinations comprising fewer than two subcarrier frequencies, namely subcarrier frequency combinations 1, 2, 3, and 4, could be assigned to pen tool functions that do not require pressure sensitivity, such as for example "eraser”, “right click”, “left click” and the like.
  • the pen tools 220 may comprise a pressure-sensitive tip, such as one that is configured to inform microcontroller 250 as to the depth within the pen tool it has been depressed.
  • the pressure-sensitive tip may comprise a pressure sensor coupled to the actuator 225. Pen tools without a pressure-sensitive tip could still be used with a pressure-sensitive capable interactive input system, but would give rise to an intensity profile in which the subcarrier frequencies have the same amplitude.
  • each image frame capture sequence comprises ten (10) image frames.
  • the first image frame (“Frame #1”) of each sequence is captured with the IR LEDs 84 on so as to obtain a preliminary illuminated image frame.
  • the second image frame (“Frame #2”) of each sequence is captured with the IR LEDs 84 off so as to obtain a background image frame.
  • the following eight (8) image frames (“Frame #3" to "Frame #10") are captured with the IR LEDs 84 on to obtain illuminated image frames.
  • the exposure of the image sensors 70 of the four (4) imaging assemblies 60 and the powering of the IR LEDs 84 are staggered to avoid any effects resulting from illumination of neighbouring IR LEDs.
  • each DSP 72 subjects the image frames in the sequence to an image frame processing method 370, which is similar to image frame processing method 270 described above.
  • the background image frame (“Frame #2”) is subtracted from the preliminary illuminated image frame (i.e. Frame #1) so as to yield a difference image frame, and one or more pointer analysis regions are applied to the following eight (8) illuminated image frames ("Frame #3" to "Frame #10") stored in the buffer.
  • the DSP 72 carries out a demodulation of these eight (8) illuminated image frames.
  • FIG. 19 and 20 show another embodiment of an interactive input system in the form of a touch table, and which is generally referred to using reference numeral 400.
  • Interactive input system 400 is similar to that described in U.S. Patent Application Publication No. 2011/0006981 to Chtchetinine et al. entitled “INTERACTIVE INPUT SYSTEM” filed on July 10, 2009, assigned to SMART Technologies, ULC, the disclosure of which is incorporated herein by reference in its entirety.
  • Interactive input system 400 comprises six (6) imaging assemblies 470a to 470f positioned about the periphery of the input area 462, and which look generally across the input area 462.
  • An illuminated bezel 472 surrounds the periphery of the input area 462 and generally overlies the imaging assemblies 470a to 470f.
  • the illuminated bezel 472 provides backlight illumination into the input area 462.
  • processing structure of interactive input system 400 utilizes a weight matrix method disclosed in PCT Application Publication No. WO 2011/003205 to Chtchetinine et al. entitled "INTERACTIVE INPUT SYSTEM AND METHOD" filed on July 12, 2010, assigned to SMART Technologies, ULC, the disclosure of which is incorporated herein by reference in its entirety.
  • Each imaging assembly 470a to 470f comprises a respective IR LED
  • the imaging assemblies 470a to 470f are grouped into four (4) imaging assembly banks, namely: a first imaging assembly bank 480a comprising imaging assemblies 470b and 470f; a second imaging assembly bank 480b comprising imaging assemblies 470a and 470e; a third imaging assembly bank 480c comprising imaging assembly 470d; and a fourth imaging assembly bank 480d comprising imaging assembly 470c.
  • the imaging assemblies within each bank capture image frames simultaneously.
  • the IR LEDs within each bank flood the input area 462 with infrared illumination simultaneously.
  • Figure 21 shows a portion of the image frame capture sequence used by the interactive input system 400, which is generally indicated using reference numeral 460.
  • Each imaging assembly bank continuously captures image frames that are grouped into sequences, with each sequence comprising ten (10) image frames.
  • the first image frame of each sequence is captured with the illuminated bezel 472 off and with the IR LEDs 474a to 474f off, so as to obtain a background image frame.
  • the second image frame of each sequence is captured with the illuminated bezel 472 on and with the IR LEDs 474a to 474f off, so as to obtain a preliminary illuminated image frame.
  • the following eight (8) image frames are captured with the illuminated bezel 472 off and with the IR LEDs 474a to 474f on, so as to obtain illuminated image frames for each respective imaging assembly bank 480a to 480d. Similar to image frame capture sequence 260 described above, the exposure of the image sensors of the four (4) imaging assembly banks 480a to 480d and the illumination of the IR LEDs 474a to 474f are staggered to avoid any effects resulting from illumination of neighbouring IR LEDs.
  • the image frames of the sequence are subjected to an image frame processing method similar to image frame processing method 270 illustrated in Figure 12 to determine the identity of one or more pen tools 220 by analyzing the frequency domain representation of the light intensity variation.
  • the background image frame (“Frame #1”) is subtracted from the preliminary illuminated image frame (i.e. Frame #2) so as to yield a difference image frame, and one or more pointer analysis regions are applied to the following eight (8) illuminated image frames ("Frame #3" to "Frame #10") stored in the buffer.
  • the DSP 72 carries out a demodulation of these eight (8) illuminated image frames.
  • FIG. 22 shows another embodiment of an interactive input system in the form of a touch table, and which is generally referred to using reference numeral 500.
  • Interactive input system 500 is generally similar to interactive input system 400 described above and with reference to Figures 20 and 21, however interactive input system 500 comprises twelve (12) imaging assemblies 570a to 5701 positioned about the periphery of the input area 562, and which look generally across the input area 562.
  • Each imaging assembly 570a to 5701 comprises a respective IR LED 574a to 5741 that is configured to flood the input area 562 with infrared illumination.
  • the imaging assemblies 570a to 5701 are grouped into four (4) imaging assembly banks, namely: a first imaging assembly bank 580a comprising imaging assemblies 570a to 570c; a second imaging assembly bank 580b comprising imaging assemblies 570d to 570f; a third imaging assembly bank 580c comprising imaging assemblies 570g to 570i; and a fourth imaging assembly bank 580d comprising imaging assembly 570j to 5701.
  • a first imaging assembly bank 580a comprising imaging assemblies 570a to 570c
  • a second imaging assembly bank 580b comprising imaging assemblies 570d to 570f
  • a third imaging assembly bank 580c comprising imaging assemblies 570g to 570i
  • a fourth imaging assembly bank 580d comprising imaging assembly 570j to 5701.
  • the imaging assemblies within each bank capture image frames simultaneously, and the IR LEDs within each bank flood the input area 562 with infrared illumination simultaneously.
  • Pen tool 220 may be used with still other interactive input systems.
  • Figure 23 shows another embodiment of an interactive input system 600 comprising an assembly 622 surrounding a display surface of a front projection system.
  • the front projection system utilizes a projector 698 that projects images on the display surface.
  • Imaging assemblies 660 positioned at the bottom corners of the assembly 622 look across the display surface.
  • Each imaging assembly 660 is generally similar to imaging assembly 60 described above and with reference to Figures 1 to 15, and comprises an image sensor and a set of IR LEDs mounted on a housing assembly.
  • a DSP unit receives image frames captured by the imaging assemblies 660 and carries out the image processing method described previously to locate the position of each pointer brought into proximity with the display surface and to determine if information is being communicated by a pen tool 220.
  • FIG 24 shows another embodiment of an interactive input system using a front projection system.
  • Interactive input system 700 comprises a single imaging assembly 760 positioned in proximity to a projector 798 and configured for viewing the display surface.
  • Imaging assembly 760 is generally similar to imaging assembly 60 described above and with reference to Figures 1 to 15, and comprises an image sensor and a set of IR LEDs mounted on a housing assembly.
  • a DSP unit receives image frames captured by the imaging assembly 760 and carries out the image processing method described previously to locate the position of a pointer brought into proximity with the display surface and to determine if information is being communicated by a pen tool 220.
  • Figure 25 shows another embodiment of a pen tool, and which is generally indicated using reference numeral 620.
  • Pen tool 620 is generally similar to pen tool 220 described above and with reference to Figures 6 to 9, and comprises a variable reflector 626 adjacent a conical tip 624. Similar to pen tool 620, variable reflector 626 has a multilayered structure, and comprises an inner layer 634 comprising a retro- reflective material, an attenuating layer 636 disposed on the inner portion 634, and an outer layer 638 disposed on the attenuating layer 636.
  • the attenuating layer 636 is a polymer dispersed liquid crystal (PDLC) device.
  • PDLC polymer dispersed liquid crystal
  • the outer layer 638 is fabricated of a durable material transparent to infrared light, and in this embodiment, the outer layer 638 is fabricated of a plastic. Protruding from the tip is an actuator that resembles a nib. The actuator is biased out of the tip 624 by a spring (not shown) but can be pushed into the tip upon application of pressure thereto. The actuator is connected to a microcontroller (not shown) housed within the main body of the pen tool 620. The microcontroller is in communication with the attenuating layer 636.
  • the microcontroller When the actuator is depressed, the microcontroller is configured to switch the attenuating layer 636 between alternating levels of transparency at one or more modulating frequencies using power supplied by a battery (not shown). This switching of the attenuating layer 636 modulates IR illumination reflected by the variable reflector 626, for enabling the interactive input system 20 to determine the identity of the pen tool 620.
  • the attenuating layer 636 does not extend the axial length of the inner layer 634, and covers only an upper portion of the inner layer 634. Attenuating layer 636 is therefore configured to attenuate illumination reflecting from only an upper portion of the variable reflector 626. As a result, illumination reflected from a lower portion of variable reflector 626 is not attenuated, enabling pen tools 620 to be readily distinguished from other forms of pointers by the interactive input system 20.
  • Figure 26 shows another example of a pen tool and which is generally identified by reference numeral 720.
  • Pen tool 720 is generally similar to pen tool 220 described above and with reference to Figures 6 to 9, and comprises a main body 722 that terminates in a conical tip 724.
  • Pen tool 720 further comprises a variable reflector 726 adjacent the conical tip 724 that is similar to variable reflector 226 described above and with reference to Figures 6 to 9.
  • the pen tool 720 further comprises a microcontroller (not shown) housed within the main body 722, which is configured to switch the attenuating layer of variable reflector 726 between differing levels of transparency at one or more modulating frequencies using power supplied by a battery (not shown).
  • Pen tool 720 further comprises a series of five (5) buttons 784a to 784e disposed on the surface of the main body 722 that are in communication with the microcontroller. Each of the buttons 784a to 784e is associated with a different combination of one or more modulating frequencies used to switch the attenuating layer. Selection of one of the buttons 784a to 784e enables a different attribute of the pen tool 720 to be selected by the user.
  • button 784a is associated with digital ink having a narrow width (Figure 27A); button 784b is associated with digital ink having a medium width (Figure 27B); button 784c is associated with digital ink having a wide width (Figure 27C); button 784d is associated with dashed digital ink (Figure 27D); and button 784e is associated with star-shaped digital ink (Figure 27E).
  • Figure 28 shows another pen tool and which is generally identified by reference numeral 820.
  • Pen tool 820 is generally similar to pen tool 720 described above and with reference to Figure 26, and comprises a combination of slidable switches 884a, 884b, 884c and 884d that are disposed on the surface of the main body 822 and that are in communication with a microcontroller (not shown).
  • a microcontroller not shown.
  • each of the slidable switches 884a to 884d is associated with a different set of one or more modulating frequencies used to switch the attenuating layer.
  • Selection of one of the slidable switches 884a to 884d enables a different attribute of the pen tool 820 to be selected by the user.
  • Figure 29 shows another example of a pen tool and which is generally identified by reference numeral 920.
  • Pen tool 920 is generally similar to pen tool 720 described above and with reference to Figure 26, and comprises a multi-position slidable switch 984 in communication with a microcontroller (not shown).
  • Multi-position slidable switch 984 has a set of four (4) different indexed switch positions 985a to 985d, each of which is associated with a different combination of one or more modulating frequencies used to switch the attenuating layer (not shown). Selection of one of the switch positions 985a to 985d enables a different attribute of the pen tool 920 to be selected by the user.
  • Figure 30 shows another example of a pen tool and which is generally identified by reference numeral 1020.
  • Pen tool 1020 is generally similar to pen tool 720 described above and with reference to Figure 26, and comprises a multi-position rotatable switch 1084 in communication with a microcontroller (not shown).
  • Multi- position rotatable switch 1084 has a set of five (5) different indexed switch positions 1085a to 1085e, each of which is associated with a different combination of one or more modulating frequencies used to switch the attenuating layer (not shown). Selection of one of the switch positions 1085a to 1085e enables a different attribute of the pen tool 1020 to be selected by the user.
  • FIG 31 shows another example of a pen tool and which is generally identified by reference numeral 1 120.
  • Pen tool 1120 is generally similar to pen tool 720 described above and with reference to Figure 26, and comprises a multi-position dial switch 1184 in communication with a microcontroller (not shown).
  • Multi-position dial switch 1184 has a set of five (5) different indexed switch positions (not shown), each of which is associated with a different combination of one or more modulating frequencies used to switch the attenuating layer (not shown). Selection of one of the switch positions 1 185a to 1185e enables a different attribute of the pen tool 1120 to be selected by the user.
  • Figure 32 shows another example of a pen tool and which is generally identified by reference numeral 1220.
  • Pen tool 1220 is generally similar to pen tool 720 described above and with reference to Figure 20, and comprises a button 1284 disposed on the body 1222 of pen tool 1220 and which is in communication with a
  • Button 1284 is configured to be pressed to enable cycling between a set of five (5) different sets of one or more modulating frequencies used to switch the attenuating layer (not shown). Each pressing of the button 1284 enables a different attribute of the pen tool 1220 to be selected by the user.
  • each pen tool modulates the reflected IR light using a different combination of the three subcarrier frequencies, where for each combination, the pen tool operates each subcarrier frequency in either an "on" state or an "off state
  • each combination may alternatively involve operating each subcarrier frequency in any one of a "full-power” (F) state, a "half-power” (H) state, and an "off state (O).
  • F full-power
  • H half-power
  • OF "off state
  • the twenty-seven (27) subcarrier frequency combinations are tabulated below in Table 2.
  • the frame rate of the imaging assemblies is 960Hz
  • the cycling rate of the IR light sources is 480Hz
  • the subcarrier frequencies are 120Hz, 240 Hz, and 360Hz
  • the imaging assemblies may be capable of very high frame rates, such as those on the order of 10 6 frames per second, or very low frame rates, such as 30 frames per second.
  • very high frame rates such as those on the order of 10 6 frames per second
  • very low frame rates such as 30 frames per second.
  • the IR light sources are cycled at a rate that is half of the frame rate
  • the IR light sources may alternatively be cycled at other rates, such as 1/3, 1/4 or 1/100 of the frame rate, for example.
  • the difference image frame is obtained by subtracting a background image frame from an illuminated image frame, where the background image frame and the illuminated image frame are captured successively
  • the difference image frame may be obtained using an alternative approach.
  • the difference image frame may be obtained by dividing the background image frame by the illuminated image frame, or vice versa.
  • non-successive image frames may alternatively be used for obtaining the difference image frame.
  • the pointer analysis region is square, it will be appreciated that this region is not limited to this shape. Accordingly, the pointer analysis region may be other shapes, such as rectangular, circular etc.
  • pen tip pressure may be communicated to the interactive input system by varying the amplitude of the subcarrier frequencies, in other embodiments, pen tip pressure may
  • the light sources emit infrared illumination
  • illumination of other wavelengths may alternatively be emitted.

Abstract

A pen tool for use with a machine vision interactive input system comprises an elongate body, a tip on the body and a variable reflector disposed on the body. The variable reflector comprises a reflecting portion and an attenuating portion configured to control the amount of illumination reflected by the reflecting portion.

Description

INTERACTIVE INPUT SYSTEM WITH VARIABLE REFLECTOR PEN TOOL
Field of the Invention
[0001] The present invention relates to an interactive input system and to a pen tool therefor.
Background of the Invention
[0001] Interactive input systems that allow users to inject input into an application program using an active pointer (e.g. a pointer that emits light, sound or other signal), a passive pointer (e.g. a finger, cylinder or other object) or other suitable input device such as for example, a mouse or trackball, are well known. These interactive input systems include but are not limited to: touch systems comprising touch panels employing analog resistive or machine vision technology to register pointer input such as those disclosed in U.S. Patent Nos. 5,448,263; 6,141,000; 6,337,681; 6,747,636;
6,803,906; 7,232,986; 7,236,162; and 7,274,356 and in U.S. Patent Application
Publication No. 2004/0179001 assigned to SMART Technologies ULC of Calgary, Alberta, Canada, assignee of the subject application, the entire disclosures of which are incorporated by reference; touch systems comprising touch panels employing electromagnetic, capacitive, acoustic or other technologies to register pointer input; tablet and laptop personal computers (PCs); personal digital assistants (PDAs) and other handheld devices; and other similar devices.
[0002] Above-incorporated U.S. Patent No. 6,803,906 to Morrison et al. discloses a touch system that employs machine vision to detect pointer interaction with a touch surface on which a computer-generated image is presented. A rectangular bezel or frame surrounds the touch surface and supports digital cameras at its corners. The digital cameras have overlapping fields of view that encompass and look generally across the touch surface. The digital cameras acquire images looking across the touch surface from different vantages and generate image data. Image data acquired by the digital cameras is processed by on-board digital signal processors to determine if a pointer exists in the captured image data. When it is determined that a pointer exists in the captured image data, the digital signal processors convey pointer characteristic data to a master controller, which in turn processes the pointer characteristic data to determine the location of the pointer in (x,y) coordinates relative to the touch surface using triangulation. The pointer coordinates are conveyed to a computer executing one or more application programs. The computer uses the pointer coordinates to update the computer-generated image that is presented on the touch surface. Pointer contacts on the touch surface can therefore be recorded as writing or drawing or used to control execution of application programs executed by the computer.
[00031 U.S. Patent No. 6,972,401 to Akitt et al. assigned to SMART
Technologies ULC, the disclosure of which is incorporated herein by reference in its entirety, discloses an illuminated bezel for use in a touch system such as that disclosed in above-incorporated U.S. Patent No. 6,803,906 to Morrison et al. The illuminated bezel comprises infrared (IR) light emitting diodes (LEDs) that project infrared light onto diffusers. The diffusers in turn, diffuse the infrared light so that the intensity of backlighting provided over the touch surface by the illuminated bezel is generally even across the surfaces of the diffusers. As a result, the backlight illumination provided by the bezel appears generally continuous to the digital cameras. Although this illuminated bezel works very well, it adds cost to the touch system.
[0004] U.S. Patent Publication Application No. 2011/0242060 to McGibney et al., entitled "INTERACTIVE INPUT SYSTEM AND INFORMATION INPUT METHOD THEREFOR", filed April 1, 2010, assigned to SMART Technologies ULC, the disclosure of which is incorporated herein by reference in its entirety, discloses an interactive input system comprising at least one imaging assembly having a field of view looking into a region of interest and capturing image frames and processing structure in communication with the at least one imaging assembly. When a pointer exists in captured image frames, the processing structure demodulates the captured image frames to determine frequency components thereof and examines the frequency components to determine at least one attribute of the pointer.
[0005] U.S. Patent Application Publication No. 2011/0242006 to Thompson et al., filed on April 1, 2010, entitled "INTERACTIVE INPUT SYSTEM AND PEN TOOL THEREFOR", and assigned to SMART Technologies ULC, the disclosure of which is incorporated herein by reference in its entirety, discloses a pen tool for use with a machine vision interactive input system comprising an elongate body and a tip arrangement at one end of the body. An end surface of the body at least partially about the tip arrangement carries light reflective material that is visible to at least one imaging assembly of the interactive input system when the pen tool is angled.
[0006] U.S. Patent No. 6,823,481 to Takekawa discloses a method and apparatus for preventing misdetection of coordinates of a pointing instrument. A coordinate input/detection device compares a waveform of a signal corresponding to intensity distribution of rays received from a coordinate input area by a light-receiving unit and a standard waveform at fixed timing. If the waveform is not equal to the standard waveform, the coordinate input/detection device determines that an installation position of the light-receiving unit is off a correct position because of shock or vibration. A new standard waveform is obtained and stored in a storage unit, thereby setting the waveform of the signal corresponding to the intensity distribution of rays received by the light-receiving unit located at an incorrect position as the standard waveform.
[0007] U.S. Patent Nos. 7,202,860 and 7,414,617 to Ogawa disclose a coordinate input device that includes a pair of cameras positioned in an upper left position and an upper right position of a display screen of a monitor lying close to a plane extending from the display screen of the monitor and views both a side face of an object in contact with a position on the display screen and a predetermined desktop coordinate detection area to capture the image of the object within the field of view. The coordinate input device also includes a control circuit which calculates the coordinate value of a pointing tool, pointing to a position within a coordinate detection field, based on video signals output from the pair of cameras, and transfers the coordinate value to a program of a computer.
[0008] U.S. Patent No. 6,567,078 to Ogawa discloses a handwriting communication system, a handwriting input device and a handwriting display device used in the system, which can communicate by handwriting among a plurality of computers connected via a network. The communication system includes a handwriting input device which is provided at a transmitting side for inputting the handwriting into a transmitting side computer, and a handwriting display device which is provided at a receiving side for displaying the handwriting based on information transmitted from the transmitting side to a receiving side computer. The system transmits only a contiguous image around the handwritten portion, which reduces the communication volume compared to transmitting the whole image, and which makes the real time transmission and reception of handwriting trace possible.
[0009] U.S. Patent No. 6,441 ,362 to Ogawa discloses an optical digitizer for determining a position of a pointing object projecting a light and being disposed on a coordinate plane. In the optical digitizer, a detector is disposed on a periphery of the coordinate plane and has a view field covering the coordinate plane for receiving the light projected from the pointing object and for converting the received light into an electric signal. A processor is provided for processing the electric signal fed from the detector to compute coordinates representing the position of the pointing object. A collimator is disposed to limit the view field of the detector below a predetermined height relative to the coordinate plane such that through the limited view field the detector can receive only a parallel component of the light which is projected from the pointing object substantially in parallel to the coordinate plane. A shield is disposed to enclose the periphery of the coordinate plane to block noise light other than the projected light from entering into the limited view field of the detector.
[0010] Improvements are generally desired. It is therefore an object of the present invention at least to provide a novel interactive input system and a pen tool therefor.
Summary of the Invention
[0011] Accordingly, in one aspect there is provided a pen tool for use with a machine vision interactive input system comprising an elongate body; a tip on the body; and a variable reflector disposed on the body, the variable reflector comprising a reflecting portion and an attenuating portion configured to control the amount of illumination reflected by the reflecting portion.
]0012] In one embodiment, the variable reflector is positioned adjacent the tip.
The attenuating portion comprises a liquid crystal device. In one form, the liquid crystal device is a polymer dispersed liquid crystal device comprising a polymer dispersed liquid crystal film positioned in an annular region between coaxial electrodes. The transparency of the attenuating element is proportional to an applied voltage and the variable reflector further may comprise a protective outer portion. [0013] In one embodiment, the pen tool further comprises a microcontroller configured to switch the attenuating portion between differing levels of transparency. In one embodiment, the pen tool further comprises an actuator protruding from the tip and being in communication with the controller. The controller is configured to switch the attenuating portion between two or more levels of transparency in response to actuation of the actuator. In another embodiment, the pen tool further comprises a switch arrangement in communication with the controller for selecting the differing levels of transparency.
[0014] In another embodiment, the reflecting portion comprises a retro-reflective material. In a further embodiment, the attenuating portion covers only a portion of the retro-reflective material.
[0015] In another aspect, there is provided an interactive input system comprising at least one imaging assembly having a field of view looking into a region of interest and capturing image frames; at least one illumination source configured to emit illumination into said region of interest; and processing structure in communication with the at least one imaging assembly, when a pointer exists in captured image frames, said processing structure demodulating the captured image frames to determine frequency components thereof and examining the frequency components to determine at least one attribute of said pointer, the frequency components being representative of illumination reflected by said pointer.
[0016] In one embodiment, during demodulating the processing structure applies a transform to the captured image frames. In another embodiment, the illumination is infrared illumination.
[0017] In another embodiment, the attribute comprises one of pointer input colour, pointer input line thickness and pointer functionality. The pointer functionality may be one of right mouse click, left mouse click, and eraser.
[0018] In another embodiment, the at least one imaging assembly captures a sequence of image frames, each sequence comprising one image frame captured when the at least one illumination source is in an off state and a plurality of image frames captured when the at least one illumination source is in an on state. The processing structure may subtract the image frame captured when the at least one illumination source is in the off state from at least one image frame captured when at least one illumination source is in the on state to form a difference image frame, and determines a location of the pointer in the difference image frame. The processing structure further defines an analysis region associated with the pointer location, and applies the analysis region to the captured image frames for carrying out the examining.
[0019] In another aspect, there is provided a method of inputting information into an interactive input system comprising emitting illumination into a region of interest from at least one illumination source; capturing image frames of the region of interest; when a pointer exists in captured image frames, demodulating the captured image frames to determine frequency components thereof and examining the frequency components to determine at least one attribute of the pointer, the frequency components being representative of illumination reflected by said pointer.
Brief Description of the Drawings
[0020] Embodiments will now be described more fully with reference to the accompanying drawings in which:
[0021] Figure 1 is a schematic perspective view of an interactive input system;
[0022] Figure 2 is a schematic block diagram view of the interactive input system of Figure 1 ;
[0023] Figure 3 is a block diagram of an imaging assembly forming part of the interactive input system of Figure 1 ;
[0024] Figures 4A and 4B are front and rear perspective views, respectively, of a housing assembly forming part of the imaging assembly of Figure 3;
[0025] Figure 5 is a block diagram of a master controller forming part of the interactive input system of Figure 1 ;
[0026] Figure 6 is a perspective view of a pen tool for use with the interactive input system of Figure 1;
[0027] Figure 7 is a cross-sectional view of a portion of the pen tool of Figure
6;
[0028] Figure 8 is cross-sectional view of the tip portion of the pen tool of
Figure 6;
[0029] Figure 9 is a block diagram of components of the pen tool of Figure 6; [0030] Figure 10 is a graphical plot of an image frame capture sequence used by the interactive input system of Figure 1 ;
[0031] Figures 11 A and 1 IB depict a sequence of nine (9) consecutive image frames captured by the imaging assembly of Figure 3 when the pen tool of Figure 6 is in the field of view of the imaging assembly and is reflecting illumination modulated according to a subcarrier frequency combination;
[0032] Figure 12 is a flowchart showing steps of an image processing method;
[0033] Figure 13 is a graphical plot of the light intensity variation of the image frames of Figure 11A and 1 IB;
[0034] Figure 14 is a graphical plot of the frequency domain representation of the image frame light intensity variation of Figure 13;
[0035] Figures 15 A and 15B show frequency domain representations of the light intensity of eight (8) subcarrier frequency combinations;
[0036] Figures 16A to 16C are perspective views of an input surface of the interactive input system of Figure 1, showing a pressure-sensitive pen tool applying light pressure input, medium pressure input, and heavy pressure input, respectively;
[0037] Figures 17A to 17C show frequency domain representations of light intensity corresponding to the light, medium and heavy pressure inputs of Figures 16A to 16C, respectively;
[0038] Figure 18 is a graphical plot of another embodiment of an image frame capture sequence used by the interactive input system of Figure 1 ;
[0039] Figure 19 is another embodiment of an interactive input system;
[0040] Figure 20 is a schematic plan view of an imaging assembly
arrangement forming part of the interactive input system of Figure 19;
[0041] Figure 21 is a graphical plot of an image frame capture sequence used by the interactive input system of Figure 19;
[0042] Figure 22 is a schematic plan view of another embodiment of an imaging assembly arrangement for the interactive input system of Figure 19;
[0043] Figure 23 is another embodiment of an interactive input system;
[0044] Figure 24 is another embodiment of an interactive input system;
[0045] Figure 25 is a partial cross-sectional view of a portion of another embodiment of a pen tool; [0046] Figure 26 is a perspective view of another embodiment of a pen tool;
[0047] Figures 27A to 27E are perspective views of the pen tool of Figure 26 applying input in accordance with various different selected attributes;
[0048] Figure 28 is a perspective view of another embodiment of a pen tool;
[0049] Figure 29 is a perspective view of another embodiment of a pen tool;
[0050] Figure 30 is a perspective view of another embodiment of a pen tool;
[0051] Figure 31 is a partial cross-sectional view of another embodiment of a pen tool; and
[0052] Figure 32 is a perspective view of another embodiment of a pen tool.
Detailed Description of the Embodiments
[0053] Turning now to Figures 1 and 2, an interactive input system that allows a user to inject input such as digital ink, mouse events etc. into an executing application program is shown and is generally identified by reference numeral 20. In this embodiment, interactive input system 20 comprises an interactive board 22 mounted on a vertical support surface such as for example, a wall surface or the like. Interactive board 22 comprises a generally planar, rectangular interactive surface 24 that is surrounded about its periphery by a bezel 26. An ultra-short throw projector (not shown) such as that sold by SMART Technologies ULC under the name SMART UX60 is also mounted on the support surface above the interactive board 22 and projects an image, such as for example a computer desktop, onto the interactive surface 24.
[0054] The interactive board 22 employs machine vision to detect one or more pointers brought into a region of interest in proximity with the interactive surface 24. The interactive board 22 communicates with a general purpose computing device 28 executing one or more application programs via a universal serial bus (USB) cable 30 or other suitable wired or wireless connection. General purpose computing device 28 processes the output of the interactive board 22 and, if required, adjusts image data being output to the projector so that the image presented on the interactive surface 24 reflects pointer activity. In this manner, the interactive board 22, general purpose computing device 28 and projector allow pointer activity proximate to the interactive surface 24 to be recorded as writing or drawing or used to control execution of one or more application programs executed by the general purpose computing device 28. [0055] The bezel 26 in this embodiment is mechanically fastened to the interactive surface 24 and comprises four bezel segments 40, 42, 44, 46. Bezel segments 40 and 42 extend along opposite side edges of the interactive surface 24 while bezel segments 44 and 46 extend along the top and bottom edges of the interactive surface 24 respectively. In this embodiment, the inwardly facing surface of each bezel segment 40, 42, 44 and 46 comprises a single, longitudinally extending strip or band of retro- reflective material. To take best advantage of the properties of the retro-reflective material, the bezel segments 40, 42, 44 and 46 are oriented so that their inwardly facing surfaces extend in a plane generally normal to the plane of the interactive surface 24.
[0056] A tool tray 48 of the type described in International PCT Application
Publication No. WO 2011/085486 is affixed to the interactive board 22 adjacent the bezel segment 46 using suitable fasteners such as for example, screws, clips, adhesive etc. As can be seen, the tool tray 48 comprises a housing 48a having an upper surface 48b configured to define a plurality of receptacles or slots 48c. The receptacles 48c are sized to receive one or more pen tools P as will be described as well as an eraser tool (not shown) that can be used to interact with the interactive surface 24. Control buttons 48d are provided on the upper surface 48b of the housing 48a to enable a user to control operation of the interactive input system 20. One end of the tool tray 48 is configured to receive a detachable tool tray accessory module 48e while the opposite end of the tool tray 48 is configured to receive a detachable communications module 48f for remote device communications. The housing 48a accommodates a master controller 50 (see Figure 5) as will be described.
[0057] As shown in Figure 2, imaging assemblies 60 are accommodated by the bezel 26, with each imaging assembly 60 being positioned adjacent a different corner of the bezel. The imaging assemblies 60 are oriented so that their fields of view overlap and look generally across the entire interactive surface 24. In this manner, any pointer such as for example a user's finger, a cylinder or other suitable object, or a pen tool or eraser tool lifted from a receptacle 48c of the tool tray 48, that is brought into proximity of the interactive surface 24 appears in the fields of view of the imaging assemblies 60. A power adapter 62 provides the necessary operating power to the interactive board 22 when connected to a conventional AC mains power supply. [0058] Turning now to Figure 3, components of one of the imaging assemblies
60 are shown. As can be seen, the imaging assembly 60 comprises an image sensor 70 such as that manufactured by Aptina (Micron) under Model No. MT9V034 having a resolution of 752x480 pixels, fitted with a two element, plastic lens (not shown) that provides the image sensor 70 with a field of view of approximately 104 degrees. In this manner, the other imaging assemblies 60 are within the field of view of the image sensor 70 thereby to ensure that the field of view of the image sensor 70 encompasses the entire interactive surface 24.
[0059] A digital signal processor (DSP) 72 such as that manufactured by Analog
Devices under part number ADSP-BF522 Blackfin or other suitable processing device, communicates with the image sensor 70 over an image data bus 71 via a parallel port interface (PPI). A serial peripheral interface (SPI) flash memory 74 is connected to the DSP 72 via an SPI port and stores the firmware required for image assembly operation. Depending on the size of captured image frames as well as the processing requirements of the DSP 72, the imaging assembly 60 may optionally comprise synchronous dynamic random access memory (SDRAM) 76 to store additional temporary data as shown by the dotted lines. The image sensor 70 also communicates with the DSP 72 via a two-wire interface (TWI) and a timer (TMR) interface. The control registers of the image sensor 70 are written from the DSP 72 via the TWI in order to configure parameters of the image sensor 70 such as the integration period for the image sensor 70.
[0060] In this embodiment, the image sensor 70 operates in snapshot mode. In the snapshot mode, the image sensor 70, in response to an external trigger signal received from the DSP 72 via the TMR interface that has a duration set by a timer on the DSP 72, enters an integration period during which an image frame is captured.
Following the integration period after the generation of the trigger signal by the DSP 72 has ended, the image sensor 70 enters a readout period during which time the captured image frame is available. With the image sensor in the readout period, the DSP 72 reads the image frame data acquired by the image sensor 70 over the image data bus 71 via the PPI. The frame rate of the image sensor 70 in this embodiment is between about 900 and about 960 frames per second. The DSP 72 in turn processes image frames received from the image sensor 70 and provides pointer information to the master controller 50 at a reduced rate of approximately 120 points/sec. Those of skill in the art will however appreciate that other frame rates may be employed depending on the desired accuracy of pointer tracking and whether multi-touch and/or active pointer identification is employed.
[0061] Three strobe circuits 80 communicate with the DSP 72 via the TWI and via a general purpose input/output (GPIO) interface. The IR strobe circuits 80 also communicate with the image sensor 70 and receive power provided on LED power line 82 via the power adapter 62. Each strobe circuit 80 drives a respective illumination source in the form of an infrared (IR) light emitting diode (LED) 84 that provides infrared backlighting over the interactive surface 24. Further specifics concerning the strobe circuits 80 and their operation are described in U.S. Application Publication No. 2011/0169727 to Akitt entitled "Interactive Input System and Illumination System Therefor" filed on February 19, 2010, the disclosure of which is incorporated herein by reference in its entirety.
[0062] The DSP 72 also communicates with an RS-422 transceiver 86 via a serial port (SPORT0) and a non-maskable interrupt (NMI) port. The transceiver 86 communicates with the master controller 50 over a differential synchronous signal (DSS) communications link 88 and a synch line 90. Power for the components of the imaging assembly 60 is provided on power line 92 by the power adapter 62. DSP 72 may also optionally be connected to a USB connector 94 via a USB port as indicated by the dotted lines. The USB connector 94 can be used to connect the imaging assembly 60 to diagnostic equipment.
[0063] The image sensor 70 and its associated lens as well as the IR LEDs 84 are mounted on a housing assembly 100 that is shown in Figures 4A and 4B. As can be seen, the housing assembly 100 comprises a polycarbonate housing body 102 having a front portion 104 and a rear portion 106 extending from the front portion. An imaging aperture 108 is centrally formed in the housing body 102 and accommodates an IR- pass/visible light blocking filter 110. The filter 110 has an IR-pass wavelength range of between about 830nm and about 880nm. The image sensor 70 and associated lens are positioned behind the filter 110 and oriented such that the field of view of the image sensor 70 looks through the filter 110 and generally across the interactive surface 24. The rear portion 106 is shaped to surround the image sensor 70. Three passages 112a to 112c are formed through the housing body 102. Passages 1 12a and 112b are positioned on opposite sides of the filter 1 10 and are in general horizontal alignment with the image sensor 70. Passage 112c is centrally positioned above the filter 1 10. Each tubular passage receives a light source socket 114 that is configured to receive a respective one of the IR LEDs 84. Mounting flanges 116 are provided on opposite sides of the rear portion 106 to facilitate connection of the housing assembly 100 to the bezel 26 via suitable fasteners. A label 118 formed of retro-reflective material overlies the front surface of the front portion 104. Further specifics concerning the housing assembly and its method of manufacture are described in U.S. Application Publication No.
201 1/0170253 to Liu et al. entitled "Housing Assembly for Interactive Input System and Fabrication Method" filed on February 19, 2010, the disclosure of which is incorporated herein by reference in its entirety.
[0064] Components of the master controller 50 are shown in Figure 5. As can be seen, master controller 50 comprises a DSP 200 such as that manufactured by Analog Devices under part number ADSP-BF522 Blackfin or other suitable processing device. A serial peripheral interface (SPI) flash memory 202 is connected to the DSP 200 via an SPI port and stores the firmware required for master controller operation. A
synchronous dynamic random access memory (SDRAM) 204 that stores temporary data necessary for system operation is connected to the DSP 200 via an SDRAM port. The DSP 200 communicates with the general purpose computing device 28 over the USB cable 30 via a USB port. The DSP 200 communicates through its serial port (SPORT0) with the imaging assemblies 60 via an RS-422 transceiver 208 over the differential synchronous signal (DSS) communications link 88. In this embodiment, as more than one imaging assembly 60 communicates with the master controller DSP 200 over the DSS communications link 88, time division multiplexed (TDM) communications is employed. The DSP 200 also communicates with the imaging assemblies 60 via the RS- 422 transceiver 208 over the camera synch line 90. DSP 200 communicates with the tool tray accessory module 48e over an inter-integrated circuit (I2C) channel and communicates with the communications module 48f over universal asynchronous receiver/transmitter (UART), serial peripheral interface (SPI) and I C channels.
[0065] As will be appreciated, the architectures of the imaging assemblies 60 and master controller 50 are similar. By providing a similar architecture between each imaging assembly 60 and the master controller 50, the same circuit board assembly and common components may be used for both thus reducing the part count and cost of the interactive input system 20. Differing components are added to the circuit board assemblies during manufacture dependent upon whether the circuit board assembly is intended for use in an imaging assembly 60 or in the master controller 50. For example, the master controller 50 may require a SDRAM 76 whereas the imaging assembly 60 may not.
[0066] The general purpose computing device 28 in this embodiment is a personal computer or other suitable processing device comprising, for example, a processing unit, system memory (volatile and/or non-volatile memory), other nonremovable or removable memory (e.g. a hard disk drive, RAM, ROM, EEPROM, CD- ROM, DVD, flash memory, etc.) and a system bus coupling the various computer components to the processing unit. The computing device 28 may also comprise a network connection to access shared or remote drives, one or more networked computers, or other networked devices.
[0067] Figures 6 to 9 show better illustrate a pen tool 220 for use with the interactive input system 20. Pen tool 220 has a main body 222 terminating in a conical tip 224. In this embodiment, a variable reflector 226 is disposed on the body 222 adjacent the tip 224. Variable reflector 226 has a multilayered structure, and comprises an inner layer 234 comprising a retro-reflective material, an attenuating layer 236 disposed on the inner layer 234, and an outer layer 238 disposed on the attenuating layer 236. The attenuating layer 236 is configured to have an electrically controllable transparency, such that the transmittance of light therethrough is controllable. In this embodiment, the attenuating layer 236 is a polymer dispersed liquid crystal (PDLC) device, and comprises a PDLC film 240 positioned in the annular region between two co-axial, transparent electrodes 242 and 244. The PDLC film 240 comprises a plurality of droplets of liquid crystals dispersed in a polymer matrix. When no voltage is applied between electrodes 242 and 244, the liquid crystals within the droplets assume random orientations, causing the droplets to scatter light when it reaches the PDLC film 240. The scattering of light reduces the amount of light than can pass straight through the PDLC film 240, thus reducing its transparency. However, when a voltage is applied across the electrodes 236 and 238, the liquid crystals within the droplets become partially aligned, causing fewer of the droplets to scatter light passing through the PDLC film 240. The reduced scattering of light increases the amount of light that can pass straight through the PDLC film 240, thus increasing its transparency. The degree of alignment of the droplets within the PDLC film 240, and consequently the level of transparency of the attenuating layer 236, is generally proportional to the applied voltage. Accordingly, the transparency of the attenuating layer 236 may be varied over a range corresponding to the working voltage range of the pen tool 220. The variable reflector 226 is thus capable of controlling the amount of incident light that reaches the inner layer 234 and also the amount of light reflected from the inner layer 234.
[0068] The outer layer 238 is fabricated of a durable material that is transparent to infrared light, and functions as a protective barrier for the variable reflector 226. In this embodiment, the outer layer 238 is fabricated of plastic.
[0069] Protruding from the tip 224 is an actuator 225 that resembles a nib.
Actuator 225 is biased out of the tip 224 by a spring (not shown) but can be pushed into the tip against the bias of the spring upon application of pressure thereto. Actuator 225 is connected to a microcontroller 250 housed within the main body 222. As shown in Figure 9, microcontroller 250 is in communication with the attenuating layer 236. When the actuator 225 is depressed, microcontroller 250 is configured to switch the attenuating layer 236 between levels of transparency at one or more modulating frequencies by applying alternating voltage levels to the electrodes using power supplied by a battery 254. This switching of the attenuating layer 236 modulates the amount of IR
illumination reaching the variable reflector 226 that is reflected back by the variable reflector 226 during use of the pen tool 220 with the interactive input system 20. In this embodiment, the modulation of the amount of reflected IR illumination is used to identify the pen tool 220 although those of skill in the art will appreciate that the modulation may be performed for other purposes.
[0070] During operation, the DSP 200 of the master controller 50 outputs synchronization signals that are applied to the synch line 90 via the transceiver 208. Each synchronization signal applied to the synch line 90 is received by the DSP 72 of each imaging assembly 60 via transceiver 86 and triggers a non-maskable interrupt (NMI) on the DSP 72. In response to the non-maskable interrupt triggered by the synchronization signal, the DSP 72 of each imaging assembly 60 ensures that its local timers are within system tolerances and if not, corrects its local timers to match the master controller 50. Using one local timer, the DSP 72 initiates a pulse sequence via the snapshot line that is used to condition the image sensor to the snapshot mode and to control the integration period and frame rate of the image sensor 70 in the snapshot mode. The DSP 72 also initiates a second local timer that is used to provide output on the LED control line 174 so that the IR LEDs 84 are properly powered during the image frame capture cycle. In this embodiment, the pulse sequences and the outputs on the LED control line 174 are generated so that the frame capture rate of each image sensor 70 is eight (8) times the desired image frame output rate.
[0071] In response to the pulse sequence output on the snapshot line, the image sensor 70 of each imaging assembly 60 acquires image frames at the desired image frame rate. In this manner, image frames captured by the image sensor 70 of each imaging assembly can be referenced to the same point of time allowing the position of pointers brought into the fields of view of the image sensors 70 to be accurately triangulated. Each imaging assembly 60 has its own local oscillator (not shown) and synchronization signals are distributed so that a lower frequency synchronization signal (e.g. the point rate, 120Hz) for each imaging assembly 60 is used to keep image frame capture synchronized. By distributing the synchronization signals for the imaging assemblies 60, rather than, transmitting a fast clock signal to each image assembly 60 from a central location, electromagnetic interference is reduced.
[0072] During the image frame capture sequence, when each IR LED 84 is on, the IR LEDs flood the region of interest over the interactive surface 24 with infrared illumination. Infrared illumination that impinges on the retro-reflective bands of bezel segments 40, 42, 44 and 46 and on the retro-reflective labels 118 of the housing assemblies 100 is returned to the imaging assemblies 60. As a result, in the absence of a pointer, the image sensor 70 of each imaging assembly 60 sees a bright band having a substantially even intensity over its length, together with any ambient light artifacts. When a pointer is brought into proximity with the interactive surface 24, the pointer occludes infrared illumination reflected by the retro-reflective bands of bezel segments 40, 42, 44 and 46 and/or the retro-reflective labels 118. As a result, the image sensor 70 of each imaging assembly 60 sees a dark region that interrupts the bright band in captured image frames. If the pointer is a pen tool 220, the image sensor 70 of each imaging assembly 60 also sees a bright region above the bright band corresponding to reflection of infrared illumination from the variable reflector 226. Additionally, the reflections of the illuminated retro-reflective bands of bezel segments 40, 42, 44 and 46 and the illuminated retro-reflective labels 118 appearing on the interactive surface 24 are also visible to the image sensor 70.
[0073] When the IR light sources 82 are off, no infrared illumination impinges on the retro-reflective bands of bezel segments 40, 42, 44 and 46 or on the retro- reflective labels 1 18 of the housing assemblies 100. Consequently, the image sensor 70 of each imaging assembly 60 does not see the retro-reflective bands or the retro- reflective labels 118. If a pen tool 220 is in proximity with the display surface 24, no infrared illumination impinges on the variable reflector 226 of the pen tool 220 and consequently the image sensor 70 also does not see the variable reflector 226. Each imaging assembly 60 does however see artifacts resulting from ambient light on a dark background.
[0074] Figure 10 shows a portion of the image frame capture sequence used by the interactive input system 20, which is generally indicated using reference numeral 260. In this embodiment, each imaging assembly 60 continuously captures image frames that are grouped into image frame sequences and stored in buffers, with each image frame sequence comprising nine (9) image frames. Figures 1 1 A and 1 IB show a sequence of image frames captured by one of the imaging assemblies 60 during the image frame capture sequence 260 when a pen tool is proximate the interactive surface 24. In this embodiment, at each imaging assembly 60, the first image frame ("Frame #1") of each sequence is captured with the IR LEDs 84 off so as to obtain a background image frame, and the following eight (8) image frames (Frame #2 to Frame #9) are captured with the IR LEDs 84 on so as to obtain eight (8) illuminated image frames. To avoid any effects resulting from illumination of neighbouring IR LEDs during image frame capture sequence 260, the exposure of the image sensors 70 of the four (4) imaging assemblies 60 are staggered such that only one image frame is captured at one time. Similarly, the powering of the IR LEDs are also staggered during capture of the illuminated image frames such that each group of IR LEDs 84 is synchronized with its respective image sensor 70. In this embodiment, the cycle rate used by the four (4) image sensors 70 to capture a single image frame is 960 Hz, yielding an exposure time of 1/4 x 1/960 Hz, or about 260 Ds for each image sensor 70. [0075] Once the sequence of image frames has been captured, each DSP 72 subjects the image frames in the sequence to an image frame processing method, which is generally shown in Figure 12 and identified by reference numeral 270. To reduce the effects ambient light may have on pointer discrimination, the background image frame is subtracted from the first illuminated image frame (i.e. Frame #2) so as to yield a difference image frame having ambient light removed (step 272). The ambient light removed in this step comprises light originating from the operating environment surrounding the interactive input system 20, and infrared illumination emitted by the IR LEDs 84 that is scattered off of objects proximate to the imaging assemblies 60.
[0076] The difference image frame is then examined for values that represent the bezel and possibly one or more pointers (step 274). Methods for determining pointer location within image frames have been described in U.S. Patent Application Publication No. 2009/0277697 to Bolt et al., entitled "Interactive Input System and Pen Tool Therefor", the disclosure of which is incorporated herein by reference in its entirety. Thus, the locations of dark regions interrupting the bright band in the difference image frame are analyzed. As will be appreciated, a pen tool 220 in proximity with the interactive surface 24 will give rise to a generally dark region that results from the occlusion of infrared illumination reflected from the retro-reflective band of the bezel segments by the tip 224.
[0077] Once the locations of one or more pointers in the difference image frame have been determined, the DSP 72 then defines one or more square-shaped pointer analysis regions that are generally centered on the area directly above the bezel at each pointer location (step 276). Image data within the pointer analysis region is used for determining if pointer information presented by a subcarrier frequency combination has been received. If the pointer is a pen tool 220, the pointer analysis region will encompass at least a portion of the region corresponding to the variable reflector 226 and will have a bright appearance as a result of the IR illumination reflected therefrom.
[0078] In this embodiment, the transparency of the attenuating layer 236 of each pen tool 220, and therefore the intensity of IR illumination reflected by the variable reflector 226, is varied using a combination of three subcarrier frequencies, namely 120Hz, 240Hz, and 360Hz. In this embodiment, the transparency of the attenuating layer 236 is varied sinusoidally at each subcarrier frequency used. Turning again to Figures 11 A and 1 IB, it can be seen that the intensity of the bright region representing the variable reflector 226 in the image frames varies as a result of the modulation of the IR illumination reflected therefrom.
[0079] Having defined the one or more pointer analysis regions, the DSP 72 applies the one or more pointer analysis regions to each of the eight (8) illuminated image frames stored in the buffer (step 278). The DSP 72 then carries out a
demodulation of the illuminated image frames (step 280) and measures the total light intensity within the pointer analysis region(s) for each illuminated image frame. This measured total light intensity is dealt with as a function of image frame number and gives rise to a light intensity variation. An example of such a light intensity variation is graphically plotted in Figure 13. The DSP 72 then performs a time frequency transform on the light intensity variation to yield a frequency domain representation of light intensity (step 282), which is plotted graphically in Figure 14. In this embodiment, the transform is a discrete Fourier transform, but those of skill in the art will understand that any transform from the time domain to the frequency domain may be used. For a frame rate of 960 frames per second, an eight point discrete Fourier transform will report frequencies of OHz, 120Hz, 240Hz, 360Hz, and 480Hz. The measurement at OHz represents light from all constant light sources, including ambient light and the bias applied to active pen tool IR illumination, and is discarded. The measurement at 480Hz represents light from the oscillating IR LEDs 84 and is also discarded. The remaining measurements at 120Hz, 240Hz, and 360Hz represent modulated subcarriers of IR illumination reflected by the pen tool 220.
[0080] The available carrier frequencies used by pen tools 220 are dependent on the frame rate (i.e. the frame capture rate) of the image sensors 70. In this embodiment as mentioned previously, the frame rate is 960 frames per second. Therefore, to avoid aliasing at this frame rate, the upper limit of modulation frequency used for the attenuating layer 236 is 480Hz.
[0081] The three subcarrier frequencies, namely 120Hz, 240Hz, and 360Hz that are used by the pen tool 220 to modulate reflected IR illumination correspond to the normalized frequencies pi/4, pi/2 and 3pi/4 of the maximum modulation frequency (480Hz). Each pen tool 220 may modulate the reflected IR light using a different combination of the three subcarrier frequencies. In this embodiment, the pen tool 220 operates each subcarrier frequency in either an "on" state or an "off state. This permits a total of eight (or 2", where n is the number of subcarrier frequencies available) different subcarrier frequency combinations each having a unique modulation pattern that can be recognized by the interactive input system 20. The combination with all three subcarrier frequencies off is reserved by the interactive input system 20 for pointers that are not active pen tools, such as, for example a finger. This provides seven other unique subcarrier frequency combinations for use with pen tools 220. As each unique subcarrier frequency combination can be assigned to a different respective pen tool 220, this information enables the identity of the active pen tool to be recognized by the interactive input system 20. As will be appreciated, the identity information may be used to assign an attribute to the pen tool, such as pen colour, line thickness or functionality (e.g.
passive pointer/no functionality, left mouse click, right mouse click, black, red, green, blue, and eraser, respectively), or the identify information may be used to represent a different user (e.g. users 1 through 8, respectively). The eight subcarrier frequency combinations are tabulated below in Table 1.
Table 1
Combination Subcarrier #1 Subcarrier #2 Subcarrier #3
no.
1 off off off
2 on off off
3 off on off
4 off off on
5 on on off
6 on off on
7 ' off on on
8 on on on
[0082] The frequency domain representations of each of these eight subcarrier frequency combinations are graphically plotted in Figures 15A and 15B. The DSP 72 determines the identity of the pen tool 220 by analyzing the frequency domain representation of the light intensity variation and determining which frequencies are present above a certain threshold (step 284). If no frequencies are present in the frequency domain representation, as in the case of combination 1, then the interactive input system 20 determines that the pointer is not a pen tool 220. [0083] After the location, the identity and the tip pressure (if available) of the pointer have been determined by the DSP 72 of each imaging assembly 60, each DSP 72 then conveys the pointer data representing the position of the pointer in the image frames as well as the pointer shape and contact status (actual contact or hover) to the DSP 200 of the master controller 50. The DSP 200 uses the pointer data received from the DSPs 72 to calculate the position of each pointer relative to the interactive surface 24 in (x,y) coordinates using well known triangulation as described in above-incorporated U.S. Patent No. 6,803,906 to Morrison et al. The pointer coordinate data along with pointer identity, pointer shape and contact status is conveyed to the general purpose computing device 28 allowing the image data presented on the interactive surface 24 to be updated.
[0084] If desired, additional information about the pen tool, such as the level of pressure applied to the pen tool tip, may be communicated by the pen tool 220 by increasing both the number of subcarrier frequencies available for modulation as well as the frame rate of the imaging assemblies 60. Alternatively, the amplitude of the subcarrier frequencies may be varied. This method can be used for the subcarrier frequency combinations that comprise more than one subcarrier frequency, namely combinations 5, 6, 7 and 8 in Figures 15A and 15B. Pen tools using these subcarrier frequency combinations may be encoded such that the level of pen tip pressure is manifested as a difference in amplitude of one or more subcarrier frequencies. Figures 16A to 16C each schematically illustrate the motion of a pen tool 220 across the interactive surface 24, during which the pressure applied by the tip of the pen tool 220 to the interactive surface is light (Figure 16A), medium (Figure 16B) and heavy (Figure 16C). The corresponding subcarrier frequency signal modulation is plotted graphically in Figures 17A to 17C. The difference in amplitude of the subcarrier frequencies for different combinations is kept on the order of +/-25% to maintain the intensity of all subcarriers above the threshold for detection. Pen tools using subcarrier frequency combinations comprising fewer than two subcarrier frequencies, namely subcarrier frequency combinations 1, 2, 3, and 4, could be assigned to pen tool functions that do not require pressure sensitivity, such as for example "eraser", "right click", "left click" and the like.
[0085] To utilize such a pressure sensitive feature, the pen tools 220 may comprise a pressure-sensitive tip, such as one that is configured to inform microcontroller 250 as to the depth within the pen tool it has been depressed. In particular, the pressure-sensitive tip may comprise a pressure sensor coupled to the actuator 225. Pen tools without a pressure-sensitive tip could still be used with a pressure-sensitive capable interactive input system, but would give rise to an intensity profile in which the subcarrier frequencies have the same amplitude.
[0086] The image frame capture sequence is not limited to that described above and, in other embodiments, other image frame capture sequences may be used. For example, Figure 18 shows a portion of another embodiment of an image frame capture sequence used by the interactive input system 20, and which is generally indicated using reference numeral 360. In this embodiment, each image frame capture sequence comprises ten (10) image frames. The first image frame ("Frame #1") of each sequence is captured with the IR LEDs 84 on so as to obtain a preliminary illuminated image frame. The second image frame ("Frame #2") of each sequence is captured with the IR LEDs 84 off so as to obtain a background image frame. The following eight (8) image frames ("Frame #3" to "Frame #10") are captured with the IR LEDs 84 on to obtain illuminated image frames. The exposure of the image sensors 70 of the four (4) imaging assemblies 60 and the powering of the IR LEDs 84 are staggered to avoid any effects resulting from illumination of neighbouring IR LEDs.
[0087] Once the sequence of image frames has been captured and stored in the buffers, each DSP 72 subjects the image frames in the sequence to an image frame processing method 370, which is similar to image frame processing method 270 described above. In this embodiment, the background image frame ("Frame #2") is subtracted from the preliminary illuminated image frame (i.e. Frame #1) so as to yield a difference image frame, and one or more pointer analysis regions are applied to the following eight (8) illuminated image frames ("Frame #3" to "Frame #10") stored in the buffer. The DSP 72 carries out a demodulation of these eight (8) illuminated image frames.
[0088] The pen tool 220 described above is not only for use with interactive input system 20 described above, and may alternatively be used with other interactive input systems employing machine vision. For example, Figures 19 and 20 show another embodiment of an interactive input system in the form of a touch table, and which is generally referred to using reference numeral 400. Interactive input system 400 is similar to that described in U.S. Patent Application Publication No. 2011/0006981 to Chtchetinine et al. entitled "INTERACTIVE INPUT SYSTEM" filed on July 10, 2009, assigned to SMART Technologies, ULC, the disclosure of which is incorporated herein by reference in its entirety. Interactive input system 400 comprises six (6) imaging assemblies 470a to 470f positioned about the periphery of the input area 462, and which look generally across the input area 462. An illuminated bezel 472 surrounds the periphery of the input area 462 and generally overlies the imaging assemblies 470a to 470f. The illuminated bezel 472 provides backlight illumination into the input area 462. To detect targets, processing structure of interactive input system 400 utilizes a weight matrix method disclosed in PCT Application Publication No. WO 2011/003205 to Chtchetinine et al. entitled "INTERACTIVE INPUT SYSTEM AND METHOD" filed on July 12, 2010, assigned to SMART Technologies, ULC, the disclosure of which is incorporated herein by reference in its entirety.
[0089] Each imaging assembly 470a to 470f comprises a respective IR LED
474a to 474f that is configured to flood the input area 462 with infrared illumination. In this embodiment, the imaging assemblies 470a to 470f are grouped into four (4) imaging assembly banks, namely: a first imaging assembly bank 480a comprising imaging assemblies 470b and 470f; a second imaging assembly bank 480b comprising imaging assemblies 470a and 470e; a third imaging assembly bank 480c comprising imaging assembly 470d; and a fourth imaging assembly bank 480d comprising imaging assembly 470c. The imaging assemblies within each bank capture image frames simultaneously. Similarly, the IR LEDs within each bank flood the input area 462 with infrared illumination simultaneously.
[0090] Figure 21 shows a portion of the image frame capture sequence used by the interactive input system 400, which is generally indicated using reference numeral 460. Each imaging assembly bank continuously captures image frames that are grouped into sequences, with each sequence comprising ten (10) image frames. In this embodiment, the first image frame of each sequence is captured with the illuminated bezel 472 off and with the IR LEDs 474a to 474f off, so as to obtain a background image frame. The second image frame of each sequence is captured with the illuminated bezel 472 on and with the IR LEDs 474a to 474f off, so as to obtain a preliminary illuminated image frame. The following eight (8) image frames are captured with the illuminated bezel 472 off and with the IR LEDs 474a to 474f on, so as to obtain illuminated image frames for each respective imaging assembly bank 480a to 480d. Similar to image frame capture sequence 260 described above, the exposure of the image sensors of the four (4) imaging assembly banks 480a to 480d and the illumination of the IR LEDs 474a to 474f are staggered to avoid any effects resulting from illumination of neighbouring IR LEDs.
[0091] Once the sequence of image frames has been captured and stored in the buffers, the image frames of the sequence are subjected to an image frame processing method similar to image frame processing method 270 illustrated in Figure 12 to determine the identity of one or more pen tools 220 by analyzing the frequency domain representation of the light intensity variation. In this embodiment, the background image frame ("Frame #1") is subtracted from the preliminary illuminated image frame (i.e. Frame #2) so as to yield a difference image frame, and one or more pointer analysis regions are applied to the following eight (8) illuminated image frames ("Frame #3" to "Frame #10") stored in the buffer. The DSP 72 carries out a demodulation of these eight (8) illuminated image frames.
[0092] The pen tool 220 may be used with still other interactive input systems employing machine vision. For example, Figure 22 shows another embodiment of an interactive input system in the form of a touch table, and which is generally referred to using reference numeral 500. Interactive input system 500 is generally similar to interactive input system 400 described above and with reference to Figures 20 and 21, however interactive input system 500 comprises twelve (12) imaging assemblies 570a to 5701 positioned about the periphery of the input area 562, and which look generally across the input area 562. Each imaging assembly 570a to 5701 comprises a respective IR LED 574a to 5741 that is configured to flood the input area 562 with infrared illumination.
[0093] In this embodiment, the imaging assemblies 570a to 5701 are grouped into four (4) imaging assembly banks, namely: a first imaging assembly bank 580a comprising imaging assemblies 570a to 570c; a second imaging assembly bank 580b comprising imaging assemblies 570d to 570f; a third imaging assembly bank 580c comprising imaging assemblies 570g to 570i; and a fourth imaging assembly bank 580d comprising imaging assembly 570j to 5701. Similar to interactive input system 400 described above, the imaging assemblies within each bank capture image frames simultaneously, and the IR LEDs within each bank flood the input area 562 with infrared illumination simultaneously.
[0094] Pen tool 220 may be used with still other interactive input systems. For example, Figure 23 shows another embodiment of an interactive input system 600 comprising an assembly 622 surrounding a display surface of a front projection system. The front projection system utilizes a projector 698 that projects images on the display surface. Imaging assemblies 660 positioned at the bottom corners of the assembly 622 look across the display surface. Each imaging assembly 660 is generally similar to imaging assembly 60 described above and with reference to Figures 1 to 15, and comprises an image sensor and a set of IR LEDs mounted on a housing assembly. A DSP unit receives image frames captured by the imaging assemblies 660 and carries out the image processing method described previously to locate the position of each pointer brought into proximity with the display surface and to determine if information is being communicated by a pen tool 220.
[0095] Figure 24 shows another embodiment of an interactive input system using a front projection system. Interactive input system 700 comprises a single imaging assembly 760 positioned in proximity to a projector 798 and configured for viewing the display surface. Imaging assembly 760 is generally similar to imaging assembly 60 described above and with reference to Figures 1 to 15, and comprises an image sensor and a set of IR LEDs mounted on a housing assembly. A DSP unit receives image frames captured by the imaging assembly 760 and carries out the image processing method described previously to locate the position of a pointer brought into proximity with the display surface and to determine if information is being communicated by a pen tool 220.
[0096] Figure 25 shows another embodiment of a pen tool, and which is generally indicated using reference numeral 620. Pen tool 620 is generally similar to pen tool 220 described above and with reference to Figures 6 to 9, and comprises a variable reflector 626 adjacent a conical tip 624. Similar to pen tool 620, variable reflector 626 has a multilayered structure, and comprises an inner layer 634 comprising a retro- reflective material, an attenuating layer 636 disposed on the inner portion 634, and an outer layer 638 disposed on the attenuating layer 636. In this embodiment, the attenuating layer 636 is a polymer dispersed liquid crystal (PDLC) device. The outer layer 638 is fabricated of a durable material transparent to infrared light, and in this embodiment, the outer layer 638 is fabricated of a plastic. Protruding from the tip is an actuator that resembles a nib. The actuator is biased out of the tip 624 by a spring (not shown) but can be pushed into the tip upon application of pressure thereto. The actuator is connected to a microcontroller (not shown) housed within the main body of the pen tool 620. The microcontroller is in communication with the attenuating layer 636.
When the actuator is depressed, the microcontroller is configured to switch the attenuating layer 636 between alternating levels of transparency at one or more modulating frequencies using power supplied by a battery (not shown). This switching of the attenuating layer 636 modulates IR illumination reflected by the variable reflector 626, for enabling the interactive input system 20 to determine the identity of the pen tool 620.
[0097] In the embodiment shown in Figure 25, the attenuating layer 636 does not extend the axial length of the inner layer 634, and covers only an upper portion of the inner layer 634. Attenuating layer 636 is therefore configured to attenuate illumination reflecting from only an upper portion of the variable reflector 626. As a result, illumination reflected from a lower portion of variable reflector 626 is not attenuated, enabling pen tools 620 to be readily distinguished from other forms of pointers by the interactive input system 20.
[0098] Still other configurations of the pen tool are possible. For example,
Figure 26 shows another example of a pen tool and which is generally identified by reference numeral 720. Pen tool 720 is generally similar to pen tool 220 described above and with reference to Figures 6 to 9, and comprises a main body 722 that terminates in a conical tip 724. Pen tool 720 further comprises a variable reflector 726 adjacent the conical tip 724 that is similar to variable reflector 226 described above and with reference to Figures 6 to 9. The pen tool 720 further comprises a microcontroller (not shown) housed within the main body 722, which is configured to switch the attenuating layer of variable reflector 726 between differing levels of transparency at one or more modulating frequencies using power supplied by a battery (not shown). This switching of the attenuating layer modulates the IR illumination reflected by the variable reflector 726 during use of the pen tool 720. [0099] Pen tool 720 further comprises a series of five (5) buttons 784a to 784e disposed on the surface of the main body 722 that are in communication with the microcontroller. Each of the buttons 784a to 784e is associated with a different combination of one or more modulating frequencies used to switch the attenuating layer. Selection of one of the buttons 784a to 784e enables a different attribute of the pen tool 720 to be selected by the user. In the embodiment shown, button 784a is associated with digital ink having a narrow width (Figure 27A); button 784b is associated with digital ink having a medium width (Figure 27B); button 784c is associated with digital ink having a wide width (Figure 27C); button 784d is associated with dashed digital ink (Figure 27D); and button 784e is associated with star-shaped digital ink (Figure 27E).
[00100] Still other variations are possible. For example, Figure 28 shows another pen tool and which is generally identified by reference numeral 820. Pen tool 820 is generally similar to pen tool 720 described above and with reference to Figure 26, and comprises a combination of slidable switches 884a, 884b, 884c and 884d that are disposed on the surface of the main body 822 and that are in communication with a microcontroller (not shown). As with buttons 784a to 784e described above with respect to pen tool 720, each of the slidable switches 884a to 884d is associated with a different set of one or more modulating frequencies used to switch the attenuating layer.
Selection of one of the slidable switches 884a to 884d enables a different attribute of the pen tool 820 to be selected by the user.
[00101] Figure 29 shows another example of a pen tool and which is generally identified by reference numeral 920. Pen tool 920 is generally similar to pen tool 720 described above and with reference to Figure 26, and comprises a multi-position slidable switch 984 in communication with a microcontroller (not shown). Multi-position slidable switch 984 has a set of four (4) different indexed switch positions 985a to 985d, each of which is associated with a different combination of one or more modulating frequencies used to switch the attenuating layer (not shown). Selection of one of the switch positions 985a to 985d enables a different attribute of the pen tool 920 to be selected by the user.
[00102] Figure 30 shows another example of a pen tool and which is generally identified by reference numeral 1020. Pen tool 1020 is generally similar to pen tool 720 described above and with reference to Figure 26, and comprises a multi-position rotatable switch 1084 in communication with a microcontroller (not shown). Multi- position rotatable switch 1084 has a set of five (5) different indexed switch positions 1085a to 1085e, each of which is associated with a different combination of one or more modulating frequencies used to switch the attenuating layer (not shown). Selection of one of the switch positions 1085a to 1085e enables a different attribute of the pen tool 1020 to be selected by the user.
[00103] Figure 31 shows another example of a pen tool and which is generally identified by reference numeral 1 120. Pen tool 1120 is generally similar to pen tool 720 described above and with reference to Figure 26, and comprises a multi-position dial switch 1184 in communication with a microcontroller (not shown). Multi-position dial switch 1184 has a set of five (5) different indexed switch positions (not shown), each of which is associated with a different combination of one or more modulating frequencies used to switch the attenuating layer (not shown). Selection of one of the switch positions 1 185a to 1185e enables a different attribute of the pen tool 1120 to be selected by the user.
[00104] Figure 32 shows another example of a pen tool and which is generally identified by reference numeral 1220. Pen tool 1220 is generally similar to pen tool 720 described above and with reference to Figure 20, and comprises a button 1284 disposed on the body 1222 of pen tool 1220 and which is in communication with a
microcontroller (not shown). Button 1284 is configured to be pressed to enable cycling between a set of five (5) different sets of one or more modulating frequencies used to switch the attenuating layer (not shown). Each pressing of the button 1284 enables a different attribute of the pen tool 1220 to be selected by the user.
[00105] Although in embodiments described above, each pen tool modulates the reflected IR light using a different combination of the three subcarrier frequencies, where for each combination, the pen tool operates each subcarrier frequency in either an "on" state or an "off state, in other embodiments, each combination may alternatively involve operating each subcarrier frequency in any one of a "full-power" (F) state, a "half-power" (H) state, and an "off state (O). This permits a total of twenty-seven (or 3", where n is the number of subcarrier frequencies available) different subcarrier frequency combinations each having a unique modulation pattern that can be recognized by the interactive input system. The twenty-seven (27) subcarrier frequency combinations are tabulated below in Table 2.
Table 2
Combination Subcarrier #1 Subcarrier #2 Subcarrier #3
no.
1 O O O
2 F O O
3 O F O
4 F F O
5 0 O F
6 F O F
7 O F F
8 F F F
9 F H O
10 H F O
1 1 F O H
12 F H H
13 O F H
14 H F H
15 F F H
16 H O F
17 O H F
18 H H F
19 F H F
20 H F F
21 H O O
22 O H O
23 H H O
24 O O H
25 H O H
26 O H H
27 H H H
[00118] Although in embodiments described above, the frame rate of the imaging assemblies is 960Hz, the cycling rate of the IR light sources is 480Hz, and the subcarrier frequencies are 120Hz, 240 Hz, and 360Hz, those of skill in the art will appreciate that the interactive input system is not limited to these frequencies. For example, the imaging assemblies may be capable of very high frame rates, such as those on the order of 106 frames per second, or very low frame rates, such as 30 frames per second. Although the amount of information that may be transmitted increases as the frame rate increases, the systems and methods described above are feasible using imaging assemblies having low frame rates.
[00119] Although in embodiments described above the IR light sources are cycled at a rate that is half of the frame rate, in other embodiments, the IR light sources may alternatively be cycled at other rates, such as 1/3, 1/4 or 1/100 of the frame rate, for example.
[00120] Although in embodiments described above the difference image frame is obtained by subtracting a background image frame from an illuminated image frame, where the background image frame and the illuminated image frame are captured successively, in other embodiments, the difference image frame may be obtained using an alternative approach. For example, the difference image frame may be obtained by dividing the background image frame by the illuminated image frame, or vice versa. In still other embodiments, non-successive image frames may alternatively be used for obtaining the difference image frame.
[00121] While in embodiments described above the pointer analysis region is square, it will be appreciated that this region is not limited to this shape. Accordingly, the pointer analysis region may be other shapes, such as rectangular, circular etc.
[00122] Although in embodiments described above a single pointer analysis region is associated with each located pointer, in other embodiments, multiple pointer analysis regions may be used.
[00123] Although in embodiments described above, information relating to pen tip pressure is communicated to the interactive input system by varying the amplitude of the subcarrier frequencies, in other embodiments, pen tip pressure may
alternatively be communicated by being assigned to a respective subcarrier frequency.
[00124] Although in the embodiments described above, the light sources emit infrared illumination, in other embodiments, illumination of other wavelengths may alternatively be emitted.
[00125] Although preferred embodiments have been described, those of skill in the art will appreciate that variations and modifications may be made with departing from the scope thereof as defined by the appended claims.

Claims

What is claimed is:
1. A pen tool for use with a machine vision interactive input system comprising:
an elongate body;
a tip on the body; and
a variable reflector disposed on the body, the variable reflector comprising a reflecting portion and an attenuating portion configured to control the amount of illumination reflected by the reflecting portion.
2. The pen tool of claim 1 , wherein the variable reflector is positioned adjacent the tip.
3. The pen tool of claim 1 or 2, wherein the attenuating portion comprises a liquid crystal device.
4. The pen tool of claim 3, wherein the liquid crystal device comprises a polymer dispersed liquid crystal film positioned in an annular region between coaxial electrodes.
5. The pen tool of any one of claims 1 to 4, wherein the transparency of the attenuating portion is proportional to an applied voltage.
6. The pen tool of any one of claims 1 to 5, wherein the variable reflector further comprises a protective outer portion.
7. The pen tool of any one of claims 1 to 6, further comprising a controller configured to switch the attenuating portion between differing levels of transparency.
8. The pen tool of claim 7, further comprising an actuator protruding from the tip, the controller being configured to switch the attenuating portion between two or more levels of transparency in response to actuation of said actuator.
9. The pen tool of claim 7, further comprising a switch arrangement on said body in communication with the controller for selecting the differing levels of transparency.
10. The pen tool of claim 9 wherein said switch arrangement comprises a plurality of switch positions, each switch position being associated with a different pattern of transparency levels.
1 1. The pen tool of any one of claims 1 to 10, wherein the reflecting portion comprises a retro -reflective material.
12. The pen tool of claim 11, wherein the attenuating portion covers only a portion of the retro-reflective material.
13. The pen tool of claim 1 wherein said variable reflector is a layered structure adjacent said tip comprising an inner reflective layer over which an attenuating layer is disposed.
14. The pen tool of claim 13 wherein said attenuating layer either fully or partially covers said reflective layer.
15. The pen tool of claim 13 or 14 wherein said reflective layer is formed of retro-reflecting material.
16. The pen tool of any one of claims 13 to 15, wherein the attenuating layer comprises a liquid crystal device.
17. The pen tool of claim 16, wherein the liquid crystal device comprises a polymer dispersed liquid crystal film positioned in an annular region between coaxial electrodes.
18. The pen tool of any one of claims 14 to 17, further comprising a controller configured to switch the attenuating layer between differing levels of transparency.
19. The pen tool of claim 18, further comprising an actuator protruding from the tip, the controller being configured to switch the attenuating layer between two or more levels of transparency in response to actuation of said actuator.
20. The pen tool of claim 18, further comprising a switch arrangement on said body in communication with the controller for selecting the differing levels of transparency.
21. An interactive input system comprising:
at least one imaging assembly having a field of view looking into a region of interest and capturing image frames;
at least one illumination source configured to emit illumination into said region of interest; and
processing structure in communication with the at least one imaging assembly, when a pointer exists in captured image frames, said processing structure demodulating the captured image frames to determine frequency components thereof and examining the frequency components to determine at least one attribute of said pointer, the frequency components being representative of illumination reflected by said pointer.
22. The interactive input system of claim 21, wherein during demodulating the processing structure applies a transform to the captured image frames.
23. The interactive input system of claim 21 or 22, wherein the
illumination is infrared illumination.
24. The interactive input system of any one of claims 21 to 23, wherein the attribute comprises at least one of pointer input colour, pointer input line thickness and pointer functionality.
25. The interactive input system of claim 24, wherein the pointer functionality is one of right mouse click, left mouse click, and eraser.
26. The interactive input system of any one of claims 21 to 25, wherein the at least one imaging assembly captures a sequence of image frames, each sequence comprising one image frame captured when the at least one illumination source is in an off state and a plurality of image frames captured when the at least one
illumination source is in an on state.
27. The interactive input system of claim 26, wherein said processing structure subtracts the image frame captured when the at least one illumination source is in the off state from at least one image frame captured when the at least one illumination source is in the on state to form a difference image frame, and determines a location of the pointer in the difference image frame.
28. The interactive input system of claim 27, wherein said processing structure further defines an analysis region associated with the pointer location, and applies the analysis region to the captured image frames for carrying out the examining.
29. The interactive input system of claim 28, wherein said analysis region is applied to the plurality of image frames captured when the at least one illumination source is in the on state.
30. The interactive input system of any one of claims 21 to 29, comprising a plurality of imaging assemblies having overlapping fields of view and capturing image frames of said region of interest from different vantages.
31. The interactive input system of claim 30, comprising an illumination source associated with each of said imaging assemblies.
32. The interactive input system of claim 30 or 31 , wherein said processing structure communicates with each of said imaging assemblies, said processing structure demodulating the images frames captured by each imaging assembly to determine the frequency components thereof and examining the frequency
components to determine the at least one attribute of said pointer.
33. The interactive input system of claim 21, wherein each of the imaging assemblies comprises at least two light sensors having generally identical views of the region of interest and capturing the image frames, the at least two light sensors having exposure periods that are generally staggered.
34. The interactive input system of any one of claims 21 to 33, further comprising a bezel at least partially surrounding the region of interest and having a surface in the field of view of said at least two imaging assemblies.
35. A method of inputting information into an interactive input system comprising:
emitting illumination into a region of interest from at least one illumination source;
capturing image frames of the region of interest;
when a pointer exists in captured image frames, demodulating the captured image frames to determine frequency components thereof and examining the frequency components to determine at least one attribute of the pointer, the frequency components being representative of illumination reflected by said pointer.
36. The method of claim 35, further comprising varying the illumination reflected by the pointer.
37. The method of claim 35 or 36, wherein the at least one illumination source is cycled between on and off states.
38. The method of claim 37, wherein the captured image frames comprise image frames captured during the on and off states of the at least one illumination source.
39. The method of claim 37 or 38, further comprising:
generating a difference image frame by subtracting an image frame captured while the at least one illumination source is in the off state from an image frame captured while the at least one illumination source in the on state;
determining a location of the pointer in the difference image frame; and
defining an analysis region associated with the pointer location, wherein the examining is carried out within the analysis region applied to the captured image frames.
40. The method of claim 39, wherein the examining is carried out within the analysis region applied to image frames captured during the on state of the at least one illumination source.
41. The method of any one of claims 35 to 40, wherein the at least one attribute is selected from the group consisting of pointer input colour, pointer input line thickness and pointer functionality.
42. The method of claim 41 , wherein the pointer functionality one of right mouse click, left mouse click, and eraser.
PCT/CA2012/000882 2011-09-22 2012-09-24 Interactive input system with variable reflector pen tool WO2013040691A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201161537754P 2011-09-22 2011-09-22
US61/537,754 2011-09-22

Publications (1)

Publication Number Publication Date
WO2013040691A1 true WO2013040691A1 (en) 2013-03-28

Family

ID=47913709

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CA2012/000882 WO2013040691A1 (en) 2011-09-22 2012-09-24 Interactive input system with variable reflector pen tool

Country Status (2)

Country Link
US (1) US9292109B2 (en)
WO (1) WO2013040691A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3177983A4 (en) * 2014-08-05 2018-06-27 Hewlett-Packard Development Company, L.P. Determining a position of an input object

Families Citing this family (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI549098B (en) * 2012-07-24 2016-09-11 原相科技股份有限公司 Image sensing method, and image sensing apparatus, light source determining system utilizing the image sensing method
KR20140020108A (en) * 2012-08-08 2014-02-18 삼성전자주식회사 Method for recognizing touch pen and an electronic device thereof
JP6060581B2 (en) * 2012-09-18 2017-01-18 セイコーエプソン株式会社 Interactive system, interactive system control method, and projector
KR101910578B1 (en) * 2012-11-09 2018-10-22 삼성전자주식회사 Color optical pen for e-board or display
US20140232699A1 (en) * 2013-02-20 2014-08-21 Microvision, Inc. Interactive Projection System with Actuated Stylus
US9766723B2 (en) 2013-03-11 2017-09-19 Barnes & Noble College Booksellers, Llc Stylus sensitive device with hover over stylus control functionality
US9946365B2 (en) 2013-03-11 2018-04-17 Barnes & Noble College Booksellers, Llc Stylus-based pressure-sensitive area for UI control of computing device
US9785259B2 (en) * 2013-03-11 2017-10-10 Barnes & Noble College Booksellers, Llc Stylus-based slider functionality for UI control of computing device
CN104345986B (en) * 2013-08-08 2017-09-29 联想(北京)有限公司 A kind of input unit, alternative projection system and method
TWI511006B (en) * 2014-02-07 2015-12-01 Wistron Corp Optical imaging system and imaging processing method for optical imaging system
US9733728B2 (en) * 2014-03-03 2017-08-15 Seiko Epson Corporation Position detecting device and position detecting method
US9965052B2 (en) * 2014-06-17 2018-05-08 Apple Inc. Antenna for computer stylus
US10108301B2 (en) 2014-09-02 2018-10-23 Rapt Ip Limited Instrument detection with an optical touch sensitive device, with associating contacts with active instruments
US9965101B2 (en) 2014-09-02 2018-05-08 Rapt Ip Limited Instrument detection with an optical touch sensitive device
US9791976B2 (en) 2014-09-02 2017-10-17 Rapt Ip Limited Instrument detection with an optical touch sensitive device
US9791977B2 (en) 2014-12-16 2017-10-17 Rapt Ip Limited Transient deformation detection for a touch-sensitive surface
TW201628400A (en) * 2015-01-20 2016-08-01 原相科技股份有限公司 Image data sensing system and image data sensing method
CN105988639B (en) * 2015-01-27 2019-05-21 原相科技股份有限公司 Image data sensing system and image data method for sensing
CN108415599A (en) * 2017-05-03 2018-08-17 重庆津油纳米光学科技有限公司 A kind of touch screen answered including 3D camera shootings and distance perspective

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6714311B2 (en) * 2000-08-04 2004-03-30 Xiroku Inc. Position detection device, position pointing device, position detecting method and pen-down detecting method
US20040140963A1 (en) * 2003-01-21 2004-07-22 David Kim Stylus having variable reflectivity and method for data input therewith
EP1447706B1 (en) * 2003-02-13 2009-03-11 Industrial Technology Research Institute Polymer dispersed cholesteric liquid crystal display device and method of manufacturing thereof
WO2009135321A1 (en) * 2008-05-09 2009-11-12 Smart Technologies Ulc Interactive input system with optical bezel
US7619600B2 (en) * 2005-07-05 2009-11-17 Fuji Xerox Co., Ltd. Driving method of liquid crystal device and driving device of liquid crystal device
US20090284670A1 (en) * 2005-03-15 2009-11-19 Jiuzhi Xue Windows with electrically controllable transmission and reflection
US20110080553A1 (en) * 2008-03-14 2011-04-07 Gang Sun Electrically controlled medium for modulating light
WO2011047459A1 (en) * 2009-10-23 2011-04-28 Smart Technologies Ulc Touch-input system with selectively reflective bezel
US20110221706A1 (en) * 2008-09-15 2011-09-15 Smart Technologies Ulc Touch input with image sensor and signal processor

Family Cites Families (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5448263A (en) 1991-10-21 1995-09-05 Smart Technologies Inc. Interactive display system
US6141000A (en) 1991-10-21 2000-10-31 Smart Technologies Inc. Projection display system with touch sensing on screen, computer assisted alignment correction and network conferencing
US5793360A (en) * 1995-05-05 1998-08-11 Wacom Co., Ltd. Digitizer eraser system and method
US6411362B2 (en) 1999-01-04 2002-06-25 International Business Machines Corporation Rotational mask scanning exposure method and apparatus
JP2001209487A (en) 2000-01-25 2001-08-03 Uw:Kk Handwriting communication system, and handwriting input and handwriting display device used for the system
EP1128318A3 (en) * 2000-02-21 2002-01-23 Cyberboard A/S Position detection device
JP3934846B2 (en) 2000-03-06 2007-06-20 株式会社リコー Coordinate input / detection device, electronic blackboard system, light receiving element positional deviation correction method, and storage medium
US6803906B1 (en) 2000-07-05 2004-10-12 Smart Technologies, Inc. Passive touch system and method of detecting user input
US7126590B2 (en) * 2001-10-04 2006-10-24 Intel Corporation Using RF identification tags in writing instruments as a means for line style differentiation
JP3920067B2 (en) 2001-10-09 2007-05-30 株式会社イーアイティー Coordinate input device
US6972401B2 (en) 2003-01-30 2005-12-06 Smart Technologies Inc. Illuminated bezel and touch system incorporating the same
US7532206B2 (en) 2003-03-11 2009-05-12 Smart Technologies Ulc System and method for differentiating between pointers used to contact touch surface
US7274356B2 (en) 2003-10-09 2007-09-25 Smart Technologies Inc. Apparatus for determining the location of a pointer within a region of interest
US7232986B2 (en) 2004-02-17 2007-06-19 Smart Technologies Inc. Apparatus for detecting a pointer within a region of interest
US20070177880A1 (en) * 2004-03-15 2007-08-02 Nir Karasikov Retromodulation-based data communication
US20070165007A1 (en) * 2006-01-13 2007-07-19 Gerald Morrison Interactive input system
US20090278794A1 (en) * 2008-05-09 2009-11-12 Smart Technologies Ulc Interactive Input System With Controlled Lighting
US20090277697A1 (en) * 2008-05-09 2009-11-12 Smart Technologies Ulc Interactive Input System And Pen Tool Therefor
US20110170253A1 (en) 2010-01-13 2011-07-14 Smart Technologies Ulc Housing assembly for imaging assembly and fabrication method therefor
US8964103B2 (en) * 2010-02-16 2015-02-24 Blackberry Limited Method and apparatus for reducing continuous autofocus power consumption
US9189086B2 (en) 2010-04-01 2015-11-17 Smart Technologies Ulc Interactive input system and information input method therefor
US8872772B2 (en) 2010-04-01 2014-10-28 Smart Technologies Ulc Interactive input system and pen tool therefor

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6714311B2 (en) * 2000-08-04 2004-03-30 Xiroku Inc. Position detection device, position pointing device, position detecting method and pen-down detecting method
US20040140963A1 (en) * 2003-01-21 2004-07-22 David Kim Stylus having variable reflectivity and method for data input therewith
EP1447706B1 (en) * 2003-02-13 2009-03-11 Industrial Technology Research Institute Polymer dispersed cholesteric liquid crystal display device and method of manufacturing thereof
US20090284670A1 (en) * 2005-03-15 2009-11-19 Jiuzhi Xue Windows with electrically controllable transmission and reflection
US7619600B2 (en) * 2005-07-05 2009-11-17 Fuji Xerox Co., Ltd. Driving method of liquid crystal device and driving device of liquid crystal device
US20110080553A1 (en) * 2008-03-14 2011-04-07 Gang Sun Electrically controlled medium for modulating light
WO2009135321A1 (en) * 2008-05-09 2009-11-12 Smart Technologies Ulc Interactive input system with optical bezel
US20110221706A1 (en) * 2008-09-15 2011-09-15 Smart Technologies Ulc Touch input with image sensor and signal processor
WO2011047459A1 (en) * 2009-10-23 2011-04-28 Smart Technologies Ulc Touch-input system with selectively reflective bezel

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
OLTEAN, M.: "Switchable Glass: A possible medium for Evolvable Hardware", PROCEEDINGS OF THE FIRST NASA/ESA CONFERENCE ON ADAPTIVE HARDWARE AND SYSTEMS (AHS'06) 0-7695-2614-4/06 © 2006 IEEE * ABSTRACT; ''1. INTRODUCTION''; ''4. ADVANTAGES AND WEAKNESSES'' *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3177983A4 (en) * 2014-08-05 2018-06-27 Hewlett-Packard Development Company, L.P. Determining a position of an input object
US10318023B2 (en) 2014-08-05 2019-06-11 Hewlett-Packard Development Company, L.P. Determining a position of an input object

Also Published As

Publication number Publication date
US9292109B2 (en) 2016-03-22
US20130100022A1 (en) 2013-04-25

Similar Documents

Publication Publication Date Title
US9292109B2 (en) Interactive input system and pen tool therefor
EP2553553B1 (en) Active pointer attribute determination by demodulating image frames
US8872772B2 (en) Interactive input system and pen tool therefor
US8941620B2 (en) System and method for a virtual multi-touch mouse and stylus apparatus
US8902193B2 (en) Interactive input system and bezel therefor
US20130257825A1 (en) Interactive input system and pen tool therefor
EP2676179B1 (en) Interactive input system and tool tray therefor
CA2786338C (en) Interactive system with synchronous, variable intensity of illumination
US20150029165A1 (en) Interactive input system and pen tool therefor
KR20120058594A (en) Interactive input system with improved signal-to-noise ratio (snr) and image capture method
CA2786318A1 (en) Whiteboard with tool tray incorporating a processor
US9329700B2 (en) Interactive system with successively activated illumination sources
US20110095989A1 (en) Interactive input system and bezel therefor
US20150248189A1 (en) Touch Sensing Systems
US20110241987A1 (en) Interactive input system and information input method therefor
WO2011120145A1 (en) Interactive input device with palm reject capabilities
JP2004094569A (en) Position detecting method, position detecting device and electronic blackboard device using the same
US20120249479A1 (en) Interactive input system and imaging assembly therefor
CA2899677A1 (en) Interactive input system and pen tool therefor

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12833438

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 12833438

Country of ref document: EP

Kind code of ref document: A1