WO2007020549A2 - Method of calibrating a control system for controlling a device - Google Patents

Method of calibrating a control system for controlling a device Download PDF

Info

Publication number
WO2007020549A2
WO2007020549A2 PCT/IB2006/052652 IB2006052652W WO2007020549A2 WO 2007020549 A2 WO2007020549 A2 WO 2007020549A2 IB 2006052652 W IB2006052652 W IB 2006052652W WO 2007020549 A2 WO2007020549 A2 WO 2007020549A2
Authority
WO
WIPO (PCT)
Prior art keywords
device
display
pointing device
test pattern
image
Prior art date
Application number
PCT/IB2006/052652
Other languages
French (fr)
Other versions
WO2007020549A3 (en
Inventor
Jan Kneissler
Original Assignee
Koninklijke Philips Electronics N.V.
Philips Intellectual Property & Standards Gmbh
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to EP05107427 priority Critical
Priority to EP05107427.6 priority
Application filed by Koninklijke Philips Electronics N.V., Philips Intellectual Property & Standards Gmbh filed Critical Koninklijke Philips Electronics N.V.
Publication of WO2007020549A2 publication Critical patent/WO2007020549A2/en
Publication of WO2007020549A3 publication Critical patent/WO2007020549A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0346Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/0304Detection arrangements using opto-electronic means

Abstract

The invention describes a method of calibrating a control system (1) for controlling a device (10), comprising a pointing device (2) with a camera (3), in which method a test pattern (TP) is presented on a display (4) associated with the device (10) to be controlled. The pointing device (2) is aimed in the direction of the display (4). Using the camera (3) of the pointing device (2), image data (5) of the test pattern (TP) shown on the display (4) are generated and subsequently analyzed to determine image characteristic information (6). The control system (1) and/or the display (4) associated with the device (10) to be controlled is then calibrated according to the image characteristic information (6). The invention further describes a control system (1) for controlling a device (10). Furthermore, the invention describes a control interface (20) for use in the control system (1), a pointing device (2), and an electrically or electronically controllable device (10) comprising such a control interface (20).

Description

Method of calibrating a control system for controlling a device

This invention relates to a method of calibrating a control system for controlling a device. Furthermore, the invention relates to a control system for controlling a device. The invention also relates to a control interface for use in a control system, to a pointing device, and to an electrically or electronically controllable device comprising such a control system.

Up until now, a commonly known means of interaction between a user and a device, such as a home entertainment device or other electronic device, has been the remote control. With the aid of a remote control, a user can issue one of a number of predefined commands to the device with which that remote control can communicate. However, the commands available to the user are limited, and the buttons associated with the commands are often non-intuitive. Furthermore, each device generally has its own remote control. This can be annoying, particularly in modern homes, where a user might have several home entertainment devices such a television, CD player, DVD player, tuner, etc., and therefore also a variety of remote controls.

Recent developments in modes of interaction between a user and a device have led to a more intuitive and practical hand-held pointing device, which the user simply aims at a device in order to choose an option from a number of options on a screen, or to make a certain gesture, which is then interpreted as a particular command. The large array of buttons usually to be found on a typical remote control is, for such a pointing device, entirely superfluous.

A user interface system based on a pointing device is known from WO 2004/047011 A2, which disclosure is herewith included by reference. The concept of such a system is that a pointing device connected to a camera can be used to control any device in its surroundings by being aimed at an object, e.g. a housing of the device, a screen, or any other application or accessory associated with the device. Such a pointing device can be used to control a device by pointing at, for example, the display of the device, upon which might be shown a number of options. One possible use of such a pointing device can be to allow a user to select one of the options by showing a cursor at a point in the display which appears to correspond to a point at which the pointing device is being aimed. For example, once the user has positioned the cursor over the desired option, he can proceed to select that option. The camera of the pointing device generates images of the target area at which it is being aimed, and the image data are subsequently analyzed to determine the option selected by the user, the actual device aimed at by the user, a gesture performed by the user, etc. The pointing device's camera is typically a digital camera based on CMOS (complementary metal oxide semiconductor) or CCD (charge-coupled device) technology.

However, it may be that discrepancies between that which is displayed on the screen and that which is "seen" by the camera of the pointing device result in erroneous interpretation of the image data. For example, the update rate of a screen and the capturing rate of the camera might lead to a "flickering" in the image data, or to errors in the image data that subsequently lead to errors in the image processing. Furthermore, incorrect or non- optimal color or grey-scale rendering can lead to incorrect or ambiguous image interpretation results.

US 6,618,076 Bl offers a system for determining the relationship between an image, which is to be rendered on a screen by a projector and the image as it actually appears on the screen and captured by a camera. However, this approach ultimately offers no more than a mapping between points in an image and points on a display in order to determine the projection aspect. The other problems mentioned above are not addressed by US 6,618,076 Bl.

Therefore, an object of the present invention is to provide an easy way of compensating for the differences between images as presented on a display and as captured by a camera of a pointing device.

To this end, the present invention provides a method of calibrating a control system for controlling a device, comprising a pointing device with a camera, in which method a test pattern is presented on a display associated with the device. The pointing device is aimed in the direction of the display in the usual manner, and, using the camera of the pointing device, image data is generated of the test pattern shown on the display. The test pattern image data is analyzed to determine image characteristic information. The control system and/or the display associated with the device to be controlled is then calibrated according to the image characteristic information. The device to be controlled may be any electrically or electronically controllable device associated with a display or monitor for rendering an image, such as the display of a television, personal computer, projector system, etc. In the following, the term "display" describes both types of image rendering area, whether monitor or backdrop. The device to be controlled can thus be any type of home entertainment device, e.g. DVD recorder, television, etc., or any other device for which user options can be presented or rendered graphically on a display. To avoid confusion in the following text, the image shown in a display is referred to as the "rendered image", whereas an image of this rendered image, captured by the camera of the pointing device, is referred to as the "captured image" or "image data". Depending on the image rendering circuitry and technology, the images will be rendered on the display within a particular region given by the aspect ratio of the display, and at a certain frame rate. Furthermore, the rendered images will exhibit a certain amount of flicker or afterglow, generally imperceptible to a human viewer, but which may have detrimental effects in image processing of the captured images. The test pattern presented on the display of the device need be presented only briefly, long enough for the camera of the pointing device to capture one or more images of the test pattern. These captured images, or image data, can subsequently be analyzed to deduce or compute image characteristic information. Here, the term "image characteristic information" is intended to mean any information, relevant to a calibration procedure, describing qualities of the captured image, or discrepancies between the captured image (that which is actually "seen") and the test image (that which is intended to be seen). To this end, the known information describing the test pattern - for example, the pattern data supplied to the graphics card of the display - can be compared in a suitable way to the captured image of the test pattern to obtain the image characteristic information. Using the image characteristic information thus acquired, it is possible to determine any measures which need to be taken to adapt the step of image analysis to adjust for discrepancies between that which is presented on the display and that which is captured by the camera of the pointing device, i.e. it is possible to calibrate the control system. For instance, the image data might be altered to compensate for a discrepancy in brightness between the test pattern and the captured image of the test pattern, for example by simply adjusting the brightness level of the image data. In another example, the image characteristic information might indicate that the frame rate of the display and the capture rate of the camera are incompatible. In this case, the calibration step might comprise adjusting the capture rate of the camera of the pointing device to give better results, or the step of image analysis might be adjusted, for example, by ignoring or discarding one or more of a sequence of captured images.

An appropriate control system for controlling a device comprises a test pattern source for supplying test pattern data for a test pattern to be presented on a display of the device, and a pointing device, which pointing device comprises a camera for generating image data of the display in a target area in front of the pointing device. Furthermore, the control system comprises an image analysis unit for analyzing the image data to determine image characteristic information. The control system also comprises a calibration unit for determining, on the basis of the image characteristic information, calibration adjustments required for performing adjustments to the image data and/or the step of image analysis, and/or the display associated with the device to be controlled.

An obvious advantage of the invention is that, by performing the calibration process described above, the interaction between the pointing device and the device to be controlled can be easily optimized without requiring any significant user participation. The acquisition of the image characteristic information for a pointing device and a particular display need only be carried out once. Since the test pattern need only be shown briefly in the display, the user will not be inconvenienced in any significant way. Thereafter, the information acquired can be put to good use during normal interaction between the pointing device and the device to be controlled, in order to enable or improve the interaction by allowing correct interpretation of the images of the display captured by the camera of the pointing device. The method of calibration can be repeated as often as necessary, for example, whenever the user switches between displays by aiming pointing device at a different display, such as when the user switches from interacting with a device associated with a television display in the living room to a device associated with a display in, say, the kitchen.

The dependent claims and the subsequent description disclose particularly advantageous embodiments and features of the invention.

The test pattern presented on the display of the device preferably comprises a number of different elements. One of these might be an orientation marker, which can have a distinct shape and be located in, say, the bottom left corner of the test pattern as it appears in the display, so that image analysis can determine the orientation of the captured images. Another element might be a high-contrast element (e.g. a checkerboard pattern of black and white squares). The high-contrast element can serve to provide information as to the difference in brightness between "light" and "dark". The test pattern might also comprise color elements in various shades of the primary colors red, green and blue, as well as, for example, elements showing different grey-scale levels. The color elements can be used to estimate the color rendering characteristics of the phosphors or light-emitting substrate of the display, and the response curves of the camera's sensors. Furthermore, the test pattern might comprise one or more temporal elements, such as patches that oscillate at different known frequencies, or a time progress bar, showing the remaining duration of the test pattern in the display.

The temporal elements can preferably be used to determine, for example, the latency of the system, i.e., the time which elapses between rendering an image on the display, capturing the image by the camera of the pointing device, and processing the image.

Additional delays might incur if the image data must be transmitted from the pointing device to an external control interface prior to the image processing. If the latency is too long, an undesirable result might be that a cursor on the screen does not smoothly follow the user's movements with the pointing device, or that the reaction of the system to the user's movements is delayed. If the image characteristic information indicates that the latency is too long, the calibration procedure might react by, for example, reducing the image-processing effort to at least increase the speed at which the image processing is carried out.

In a preferred embodiment of the invention, the step of analyzing the image data comprises application of known dimensional data pertaining to the pointing device. This known dimensional data can typically be data describing the display resolution of the captured image, e.g. the image size in pixels of the camera's CCD or CMOS sensor array, the position of the principle point - i.e. the intersection of the optical axis of the camera lens with the sensor array - the capture rate of the camera, etc. Such known dimensional data can be stored in a suitable memory chip in the pointing device and simply supplied, as header information, along with the image data to the image analysis unit. Alternatively, a source of the known dimensional data of the pointing device can be external to the pointing device itself, and can be accessed in some way by the control system and forwarded to the image analysis unit. By application of the known dimensional data to the step of image analysis, it is possible to determine, among others, the aspect ratio of the display. By comparing the test pattern directly to the image of the test pattern in the method according to the invention, it is possible to determine discrepancies between the rendering of the test pattern in the display and its appearance in a captured image. These discrepancies are characteristic of the system, i.e. they are caused by physical properties of the system such as color rendering characteristics of the display, the detection spectrum of the camera of the pointing device, the frame rate of the display, the capture rate of the camera, illumination flicker of the display, etc. A comparison of the test pattern as it should appear and as it does appear in the image data gives the image characteristic data which can be used to generate calibration information for performing appropriate adjustments in order to obtain accurate image analysis in an actual interaction mode, since the actual purpose of the pointing device, as described in WO 2004/047011 A2, is allow interaction between a user and a device. For example, such an adjustment on the basis of the calibration information might involve providing the graphics card of the display with corrective information to improve the image rendering. Alternatively, the control interface might communicate with the pointing device to effect a change in the image acquisition by, for example, altering the camera's shutter time. In another possible approach, the calibration information might be applied to adapt the image processing step, for example by first processing the image data by means of suitable algorithms to compensate for undesirable discrepancies between the test pattern as it should appear and as it does appear in the image. To use the pointing device to interact with a device to be controlled, the user can aim the pointing device in the direction of the display associated with the device, upon which is presented a number of "options" such as, for example, a number of menu items. Subsequently - as explained above - an image of a target area aimed at by the pointing device is generated, and the target area image, or image data, is processed to determine the target point at which the pointing device is aimed. The chosen option is determined depending on the position of the target point in the display, or a cursor which is positioned in the display according to the movements of the pointing device. An option might be a menu item presented on a display, and aiming the pointing device at the option allows the user to easily select the option, by having a cursor seem to follow the movements of the pointing device.

Computer vision algorithms can be applied to determine the target point in the target area image when the pointing device is being aimed at a display in a calibration or interaction mode. A method of processing the image data of the target area image using computer vision algorithms might comprise detecting distinctive points in the target image data, determining corresponding points in a template of the component, e.g. the screen, of the device or in the (known) surroundings of the component, and developing a transformation for mapping the points in the image data to the corresponding points in the template. This transformation can then be used to determine the position and aspect of the pointing device relative to the component, so that the point of intersection of the pointing axis with the component can be located in the template. The position of this intersection in the template corresponds to the target point in the component at which the user is aiming the pointing device, and which can be used to easily identify the area occupied by a test pattern, or the option which has been targeted by the user. Comparing the image data with the pre-defined template may thereby be restricted to identifying and comparing only salient points such as distinctive corner points. The term "comparing" is to be understood in a broad sense, i.e. by only comparing sufficient features in order to quickly identify the option at which the user is aiming, or to identify the boundaries or the actual elements of a test pattern in the display. The processing of the image data might be carried out within the pointing device itself. However, the computing power of such a pointing device, which is preferably realized to be held comfortably in the hand, is necessarily limited by the power required by complicated computing processes. Therefore, in a preferred embodiment of the invention, the image data might be transmitted for further processing to an external control interface. Such a control interface, pursuant to the invention, comprises an image analysis unit for analyzing image data of a display of a device to determine image characteristic information and a calibration unit for determining, on the basis of the image characteristic information, calibration adjustments required for the image data and/or for the step of image analysis to compensate for discrepancies between the test pattern as it is intended to be seen in the display, and the image of the test pattern as captured by the camera of the pointing device. Such a control interface may be, for example, incorporated in the device to be controlled. For example, a DVD player sold together with a pointing device might incorporate such a control system and a means for communicating a test pattern to the display of a television connected to the DVD player. Alternatively, this control interface may be realized as a separate entity which can communicate in any appropriate manner with a device to be controlled and an associated display, whereby such a control interface may be capable of communicating with more than one device in the system, and with more than one system or type of system.

To communicate image data from the pointing device to an external control interlace, the pointing device is preferably equipped with a transmitter, and the control interlace with a suitable receiver.

The control interface can also determine when a test pattern is to be presented on a display of a device. For example, if the control interface does not yet have any calibration information for a particular display, it may cause a test pattern to be briefly shown when the pointing device is being aimed at the display. Alternatively, the control interface might automatically cause the test pattern to be displayed briefly whenever the device being controlled by the pointing device - e.g. a DVD player - has been turned on and the pointing device is being aimed at the display connected to the DVD player. In another alternative, the user can trigger a calibration procedure by means of a suitable command input to the control system.

The control interface might be provided with a test pattern stored, for example, as a suitable data file in a memory storage unit. The control interface can preferably communicate this test pattern data to the display, for example by means of the graphics card of the display, via a suitable interface. Alternatively, the control interlace may acquire the test pattern from an external source, for example from the device itself.

Naturally, it is conceivable that the control interface be incorporated in the pointing device, but, as long as the dimensions of the necessary units and modules are too large to be incorporated in a hand-held device, the control interface is preferably realized as an entity separate from the pointing device. The pointing device and control system described above combine to give a practical means for calibrating the interaction between the pointing device and any device, which avails of a display for presenting user options. Application of the invention is conceivable in various kinds of environment, such as a home, office, hotel environment, etc. The method according to the invention can be applied to any electrically or electronically controllable device of a system, which avails of a display to present its user options.

Furthermore, the control system and the device being controlled can comprise any number of modules, components or units, and can be distributed in any manner.

Other objects and features of the present invention will become apparent from the following detailed descriptions considered in conjunction with the accompanying drawing. It is to be understood, however, that the drawings are designed solely for the purposes of illustration and not as a definition of the limits of the invention.

Fig. 1 is a schematic representation of a control system comprising a pointing device, and a control interface according to an embodiment of the invention, and a display associated with a device to be controlled;

Fig. 2 shows a test pattern presented in a display of a device according to an embodiment of the invention. In the diagrams, like numbers refer to like objects throughout.

In Fig. 1, the essential elements of a control system according to the invention are shown in a schematic manner: a pointing device 2, with a camera 3 in its tip, is being aimed by a user (not shown in the diagram) at a display 4 associated with a device to be controlled 10, such that a test pattern TP, being shown in the display 4, falls within the target area A aimed at by the pointing device 2.

The test pattern TP being rendered on the display 4 of the device 10 has been supplied to the device 10 by a control interface 20, in the form of a suitable signal 24 comprising test pattern data 21 , retrieved from a test pattern source 11. The test pattern TP features several different elements, each of which serves a different purpose.

The various elements of the test pattern will now be explained with the aid of Fig. 2.

A high-contrast element 30 consists of alternating squares of black and white to give a checkerboard pattern. Towards the middle of the test pattern TP, color elements 33 are depicted as squares or patches of different colors and grey-scale levels. Due to the limitations of such a diagram, these patches can only be shown as patterns in black and white, but in reality, the patches are in color or grey-scale.

Another element, an orientation marker 34, serves to define the orientation of an image of the test pattern TP since it is located at a predefined position in the test pattern TP. Here, the orientation marker 34 is located towards the lower left of the test pattern TP.

Along the lower edge of the test pattern TP, a series of oscillating elements 32 is shown. Each of these oscillating elements 32 flickers or blinks at a certain frequency, at IHz, 2Hz, 4Hz, 8Hz and 16Hz, as seen from left to right in the diagram. Another element whose content changes over time is the temporal element 31, which is a type of progress indicator, and indicates the elapsed and remaining time for presentation of the test pattern TP in the display 4 by gradually "filling up". Since the entire duration of presentation of the test pattern TP in the display 4 is predefined, and therefore also the time taken for the temporal element 31 to fill up, the control system can determine the overall latency of the system by means of the temporal element 31 together with the oscillating elements 32.

The camera 3 of the pointing device 2 generates a sequence of images of the test pattern TP in the display 4. The resulting image data 5 of the target area A are transmitted via a transmitter 18 to the control interface 20, where they are picked up by a receiver 17 and forwarded to an image analysis unit 12 in the control interface 20 (Fig. 1). To correctly interpret the image data 5, the image analysis unit 12 also avails of known dimensional data 7 pertaining to the pointing device 2, which it can retrieve from a memory 19. An alternative means of supplying the image analysis unit 12 with the required dimensional data of the pointing device 2 would be for the pointing device 2 to transmit the dimensional data to the control interface 20 along with the image data 5. However, for the sake of clarity, the dimensional data 7 is shown as being retrieved from a physical source 19. The pointing device dimensional data 7 can be information describing the size of the camera's sensor array, its capture rate, the location of the principle point, etc., and assists the image analysis unit 12 in correctly processing the image data 5. The image analysis unit 12 compares the image data 5 of the test pattern TP, under consideration of the dimensional data 7 of the pointing device 2, to the actual TP data 21. By knowing what it should see - the actual test pattern data 21 - and by analyzing what it actually does see - the image data 5 - the image analysis unit 12 can determine several characteristics of the system, such as color rendering characteristics, delay times, contrast levels, etc. This information is provided as image characteristic information 6 to a calibration unit 13, which in turn determines any measures, which can be taken to adjust for the characteristics mentioned above. For example, the calibration unit 13 might determine that certain adjustments should be made to the image data 5 prior to image analysis in order to correct for certain system characteristics, such as non-optimal color rendering, etc. The calibration unit 13 provides the relevant unit, in this case the image analysis unit 12, with the necessary information in the form of calibration information 14. It is also conceivable that the calibration unit 13 provide another unit or module with calibration information 14. For example, it might provide the graphics card of the display with corrective information to improve the image rendering, or it might communicate with the pointing device 2 to effect a change in the image acquisition by, for example, altering the camera's shutter time.

Once the calibration procedure has been carried out, the user can continue to interact with the device 10 by means of the pointing device 2, in the usual manner. This involves interpretation of the image data 5 to determine the point at which the user is aiming the pointing device 2, in order to determine any option he may have chosen. This is described exhaustively in WO 2004/047011 A2, and only briefly indicated here with the aid of the image interpretation block 9, which performs the necessary interpretation of the image analysis results 8 in order to generate corresponding control commands 18 which can be forwarded to the device, for example by means of the communication interface 23. For the sake of simplicity, the control interlace has been shown as a separate block. Generally, however, a control interface could be incorporated in the device to be controlled, in this case the television. If the device to be controlled does not have its own display, for example a DVD player, its control interface can send the test pattern TP to the television or monitor connected to the DVD player so that the display of the television or monitor can present the test pattern for the calibration process.

Although the present invention has been disclosed in the form of preferred embodiments and variations thereon, it will be understood that numerous additional modifications and variations could be made thereto without departing from the scope of the invention. For example, the calibration procedure according to the invention can be activated whenever the pointing device is aimed at a display previously "unknown" to the control system. The control system can supply the display with a test pattern so that the test pattern is rendered in the display, allowing the control system to calibrate itself for this display. The control system might store information, acquired in this way, for a number of different displays, so that the pointing device of the control system can be used to successfully interact with devices associated with any number of displays.

For the sake of clarity, it is also to be understood that the use of "a" or "an" throughout this application does not exclude a plurality, and "comprising" does not exclude other steps or elements. A "unit" may comprise a number of blocks or devices, unless explicitly described as a single entity.

Claims

CLAIMS:
1. A method of calibrating a control system (1) for controlling a device (10), comprising a pointing device (2) with a camera (3), which method comprises the steps of presenting a test pattern (TP) on a display (4) associated with the device (10) to be controlled; - aiming the pointing device (2) in the direction of the display (4); using the camera (3) of the pointing device (2) to generate image data (5) of the test pattern (TP) shown on the display (4); analyzing the image data (5) to determine image characteristic information (6); calibrating the control system (1) and/or the display (4) according to the image characteristic information (6).
2. A method according to claim 1 , wherein the step of analyzing the image data (5) comprises application of known dimensional data (7) pertaining to the pointing device (2).
3. A method according to claim 1 or claim 2, wherein the step of calibration of the control system (1) comprises performing adjustments to the image data (5) and/or to the step of image analysis to compensate for the image characteristic information (6).
4. A method according to any of the preceding claims, wherein the test pattern
(TP) comprises a high-contrast element (30) and/or a temporal element (31), and/or an oscillating element (32) and/or a color element (33), and/or an orientation marker (34).
5. A control system (1) for controlling a device (10), comprising - a test pattern source (11) for supplying test pattern data (21) for a test pattern
(TP) to be presented on a display (4) of the device (10) to be controlled; a pointing device (2), which pointing device (2) comprises a camera (3) for generating image data (5) of the display (4) in a target area (A) in front of the pointing device
(2); an image analysis unit (12) for analyzing the image data (5) to determine image characteristic information (6); a calibration unit (13) for determining, on the basis of image characteristic information (6), calibration adjustments (14) required for the image data (4) and/or for the step of image analysis.
6. A control interface (20), comprising an image analysis unit (12) for analyzing image data (5) of a display (4) of a device (10) to be controlled to determine image characteristic information (6); - a calibration unit (13) for determining, on the basis of image characteristic (6), calibration adjustments (14) required for the image data (4) and/or for the step of image analysis and/or for the display (4) associated with the device (10) to be controlled.
7. A control interface according to claim 6, comprising a source (19) of known dimensional data (7) pertaining to the pointing device (2).
8. A control interface according to claim 6 or claim 7, comprising a test pattern acquisition unit (22) for acquiring from a test pattern source (11) test pattern data (21) to be presented as a test pattern (TP) on a display (4) of a device (10) to be controlled.
9. A control interface according to claim 8, comprising a communication interlace (23) for communicating the test pattern data (21) to the device (10) to be controlled.
10. A control interface according to any of claims 6 to 9, comprising a receiver (17) for receiving image data (5) from a pointing device (2).
11. A pointing device (2) for use in a control system (1), comprising a camera (3) for generating image data (5) of a target area (A) in front of the pointing device (2), and a control interface (20) according to any of claims 6 to 9.
12. An electrically or electronically controllable device (10) comprising a control interlace (20) according to any of claims 6 to 10.
PCT/IB2006/052652 2005-08-12 2006-08-02 Method of calibrating a control system for controlling a device WO2007020549A2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
EP05107427 2005-08-12
EP05107427.6 2005-08-12

Publications (2)

Publication Number Publication Date
WO2007020549A2 true WO2007020549A2 (en) 2007-02-22
WO2007020549A3 WO2007020549A3 (en) 2007-08-02

Family

ID=37545343

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2006/052652 WO2007020549A2 (en) 2005-08-12 2006-08-02 Method of calibrating a control system for controlling a device

Country Status (1)

Country Link
WO (1) WO2007020549A2 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2448034A (en) * 2007-03-26 2008-10-01 Avago Tech Ecbu Ip System and method for tracking an input device using a display screen in captured frames of image data
US8836796B2 (en) 2010-11-23 2014-09-16 Dolby Laboratories Licensing Corporation Method and system for display characterization or calibration using a camera device
US20160134841A1 (en) * 2014-11-10 2016-05-12 David Christopher Round Verifying information on an electronic display with an incorporated monitoring device
US10448006B2 (en) 2016-02-11 2019-10-15 Signify Holding B.V. People sensing system

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1087327A2 (en) * 1999-09-21 2001-03-28 Seiko Epson Corporation Interactive display presentation system
WO2001047285A1 (en) * 1999-12-23 2001-06-28 Justsystem Corporation Method and apparatus for calibrating projector-camera system
WO2003056505A1 (en) * 2001-12-21 2003-07-10 British Telecommunications Public Limited Company Device and method for calculating a location on a display
WO2004047011A2 (en) * 2002-11-20 2004-06-03 Koninklijke Philips Electronics N.V. User interface system based on pointing device

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1087327A2 (en) * 1999-09-21 2001-03-28 Seiko Epson Corporation Interactive display presentation system
WO2001047285A1 (en) * 1999-12-23 2001-06-28 Justsystem Corporation Method and apparatus for calibrating projector-camera system
WO2003056505A1 (en) * 2001-12-21 2003-07-10 British Telecommunications Public Limited Company Device and method for calculating a location on a display
WO2004047011A2 (en) * 2002-11-20 2004-06-03 Koninklijke Philips Electronics N.V. User interface system based on pointing device

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2448034A (en) * 2007-03-26 2008-10-01 Avago Tech Ecbu Ip System and method for tracking an input device using a display screen in captured frames of image data
GB2448034B (en) * 2007-03-26 2009-08-12 Avago Tech Ecbu Ip System and method for tracking an input device using a display screen in captured frames of image data
US8836796B2 (en) 2010-11-23 2014-09-16 Dolby Laboratories Licensing Corporation Method and system for display characterization or calibration using a camera device
US20160134841A1 (en) * 2014-11-10 2016-05-12 David Christopher Round Verifying information on an electronic display with an incorporated monitoring device
US10448006B2 (en) 2016-02-11 2019-10-15 Signify Holding B.V. People sensing system

Also Published As

Publication number Publication date
WO2007020549A3 (en) 2007-08-02

Similar Documents

Publication Publication Date Title
US9800846B2 (en) Gestural control of visual projectors
CN104076512B (en) The control method of head-mount type display unit and head-mount type display unit
JP6062000B2 (en) Display device and user interface
US9195323B2 (en) Pointer control system
KR101637126B1 (en) Ambient light adaptive displays with paper-like appearance
TWI566216B (en) Ambient light adaptive displays
KR101861393B1 (en) Integrated low power depth camera and projection device
JP2017191792A (en) Light control method and lighting device using the same
US9679392B2 (en) Hosted camera remote control
US9313439B2 (en) User adaptive display device and method thereof
KR20140066638A (en) Apparatus and method for obtaining 3d image
JP5825561B2 (en) Interactive lighting control system and method
JP2017515133A (en) Color temperature adjusting method and apparatus
JP5845783B2 (en) Display device, display control method, and program
US8830227B2 (en) Depth-based gain control
KR101764372B1 (en) Apparatus and method for compositing image in a portable terminal
US7864159B2 (en) Handheld vision based absolute pointing system
US9836639B2 (en) Systems and methods of light modulation in eye tracking devices
JP5091857B2 (en) System control method
JP4764832B2 (en) Apparatus and method for optical input device
US10209515B2 (en) Filtering devices and filtering methods
US8736708B2 (en) Information processing apparatus allowing remote operation of an image capturing apparatus and control method therefor
US7969413B2 (en) Storage medium having stored thereon program for adjusting pointing device, and pointing device
JP5235075B2 (en) Display system
RU2158059C1 (en) Method and device for automatic regulation of parameters of image on display screen

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application
NENP Non-entry into the national phase in:

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 06780287

Country of ref document: EP

Kind code of ref document: A2