WO2004066213A1 - Position and orientation sensing with a projector - Google Patents

Position and orientation sensing with a projector Download PDF

Info

Publication number
WO2004066213A1
WO2004066213A1 PCT/JP2003/016777 JP0316777W WO2004066213A1 WO 2004066213 A1 WO2004066213 A1 WO 2004066213A1 JP 0316777 W JP0316777 W JP 0316777W WO 2004066213 A1 WO2004066213 A1 WO 2004066213A1
Authority
WO
WIPO (PCT)
Prior art keywords
projector
orientation
camera
planar surface
pattern
Prior art date
Application number
PCT/JP2003/016777
Other languages
English (en)
French (fr)
Inventor
Paul A. Beardsley
Ramesh Raskar
Vaibhav Vaish
Original Assignee
Mitsubishi Denki Kabushiki Kaisha
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mitsubishi Denki Kabushiki Kaisha filed Critical Mitsubishi Denki Kabushiki Kaisha
Priority to DE60323945T priority Critical patent/DE60323945D1/de
Priority to JP2004567159A priority patent/JP4437748B2/ja
Priority to EP03768261A priority patent/EP1483740B1/en
Publication of WO2004066213A1 publication Critical patent/WO2004066213A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/74Projection arrangements for image reproduction, e.g. using eidophor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/10Image acquisition
    • G06V10/12Details of acquisition arrangements; Constructional details thereof
    • G06V10/14Optical characteristics of the device performing the acquisition or on the illumination arrangements
    • G06V10/145Illumination specially adapted for pattern recognition, e.g. using gratings
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/10Image acquisition
    • G06V10/12Details of acquisition arrangements; Constructional details thereof
    • G06V10/14Optical characteristics of the device performing the acquisition or on the illumination arrangements
    • G06V10/147Details of sensors, e.g. sensor lenses
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/12Picture reproducers
    • H04N9/31Projection devices for colour picture display, e.g. using electronic spatial light modulators [ESLM]
    • H04N9/3179Video signal processing therefor
    • H04N9/3185Geometric adjustment, e.g. keystone or convergence
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/12Picture reproducers
    • H04N9/31Projection devices for colour picture display, e.g. using electronic spatial light modulators [ESLM]
    • H04N9/3191Testing thereof
    • H04N9/3194Testing thereof including sensor feedback

Definitions

  • the invention relates generally to computer vision, and more particularly to determining the position and orientation of an object using a projected image.
  • Position sensors are useful for many applications in which it is necessary to dete ⁇ nine the position and orientation of an object.
  • the term 'position sensor' means both position and orientation sensor.
  • the position is the location of the object in a three-dimensional coordinate frame, and the orientation of the object is its rotation relative to that coordinate frame.
  • a position sensor can be attached to a moving camera that is being used for 3D scanning. The position and orientation of the camera are then known, to aid creation of a 3D reconstruction of a scene from the h ⁇ ages acquired by the moving camera.
  • a position sensor can be attached to a projector to aid correct projection of an image onto a surface or an object.
  • a position sensor can also be attached to a tool so that the tool can be properly positioned with respect to a part or a working surface. If a user is manipulating the tool, the position sensor enables haptic or other feedback, for example in tele-operation, or for computer-aided surgery.
  • a position sensor can also be used to track the location of a moving object such as a vehicle or person.
  • position sensors have been used in prior art applications is to place visual markers or active emitters, such as LEDs, in known positions in the environment.
  • the position sensor can be a camera that observes the markers or emitters in order to infer its own position and orientation.
  • Other applications use ultrasound instead of light.
  • visual markers or active emitters can be attached to the object of interest so that the object's position and orientation can be determined directly using a camera observing the object.
  • a system which uses LED emitters in the environment is described in "The HiBall Tracker: High-Performance Wide- Area Tracking for Virtual and Augmented Environments," by Welch et al, Proc. of the ACM Symposium on Virtual Reality Software and Technology, 1999.
  • a system which uses ultrasound emitters in the environment is the IS-600 motion tracker by Intersense Corporation, Burlington, Massachusetts.
  • a position and orientation sensor for an object, which works without having to modify the environment, and which does not require a complicated calibration.
  • a position and orientation sensor that is compact, i.e., the sensor does not require markers or emitters that are distributed across the environment, or across the surface of an object of interest.
  • a position and orientation of a projector are dete ⁇ nined from projected images. First, a homography between a camera and a planar surface, with the camera and planar surface in a fixed relationship, is determined. Next, a known pattern is projected on the planar surface using a projector having known intrinsic parameters and an unknown position and orientation.
  • An image of the known pattern is acquired with the camera, and a position and orientation of the projector is determined from the image, using the homography and the intrinsic parameters.
  • Figure 1 is a block diagram of a position and orientation sensing system according to the invention.
  • Figure 2 is a flow diagram of a calibration process according to the invention.
  • Figure 3 is a flow diagram of a position and orientation sensing method according to the invention.
  • Figure 1 shows a position and orientation sensing system 100 according to the invention.
  • the system includes a fixed camera 110 and a movable projector 120. Both the camera 110 and the projector 120 view a fixed, planar (2D) surface 130.
  • the camera and the planar surface have a fixed relative position with respect to each other.
  • the camera and projector can be on the same side of the surface, or alternatively, if the surface is translucent, the camera and the projector can be on opposite sides of the planar surface.
  • the camera 110 and the projector 120 are digital devices. Analog devices, with A/D converters, are also possible.
  • a processor (CPU) 140 is coupled to the camera 110.
  • the processor 140 is equipped with a memory and I/O ports as are known in the art.
  • the processor 140 is capable of executing a position and orientation sensing method 300 according to the invention, as described below in greater detail.
  • a calibration procedure 200 of the system 100 has two stages 210-220.
  • the second is optional.
  • the fixed camera 110 is calibrated relative to the fixed planar surface 130 to determine a homography H 211 between a coordinate frame of an image plane of the camera 110, and a Euclidean (metric) coordinate frame of the fixed surface 130. This can be achieved, using well known geometric relationships, given four points 131 of known position on the fixed surface 130, e.g., the four corners of a square on the fixed planar surface 130.
  • the homography 211 relates the coordinate frame of the camera 110 to the coordinate frame of the fixed surface 130 in the following way.
  • an overall scale of the coordinate frame of the surface 130 is unknown, e.g., the four points 131 on the surface 130 are known to be the corners of a square but the size of the square is unknown, then the position and orientation of the projector 120 is determined up to an unknown scale in the subsequent processing. This is sufficient for many applications e.g. for 3D scanning of a scene in which the goal is to create a 3D reconstruction but the overall scale may be immaterial.
  • the calibration procedure 200 determines the intrinsic parameters IP 221 of the projector 120. Determining the intrinsic parameters of a pinhole device, such as the projector 120, is a well- known process. Alternatively, the intrinsic parameters of the projector 120 can be determined while the system 100 is operating, as described in further detail below.
  • the position sensing system 100 can be used for sensing the position and orientation of an object using the projector 120 and the method 300.
  • the homography 211 between the camera 110 and the planar surface 130 is dete ⁇ nined 310, as described above.
  • a known pattern 230 is projected 320 onto the planar surface 210 by the projector 120 having known intrinsic parameters 221, but an unknown position and orientation.
  • the pattern 230 includes at least three distinct non-collinear points.
  • the camera 110 acquires 330 an image 331 of the known pattern 230, and the features of the pattern 230 are measured in image coordinates of the projector 120.
  • the homography 211 is used to transfonn features of the pattern from the coordinate frame of the image 331 to the coordinate frame of the planar surface 130.
  • the position and orientation of the projector 120 are determined.
  • the approach is an adaptation of a method described by DeMenthon et al., "Exact and Approximate Solutions of the Perspective-Three-Point Problem," IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 14, no. 11, pp. 1100-1105, November 1992.
  • the DeMenthon method is for computing the position and orientation of a camera with known intrinsic parameters observing a physical pattern with known coordinates.
  • that method is adapted for the case of a projector which is projecting a pattern onto a plane, with a coordinate frame imposed on the projected pattern by means of a fixed camera observing the plane.
  • the projected pattern can be adapted as necessary during operation, to provide the best possible view to the fixed camera(s).
  • Any system that can be modeled as a pinhole projector is suitable for making the projection, e.g., three different-colored laser pointers with distinct beam directions and with the beams concurrent at a point, or with the beams non-concurrent but with calibration data to take account of the non-concurrency.
  • the intrinsic parameters 221 of the projector 210 can be determined during operation of the system 100.
  • the pattern 230 is projected for two different positions of the projector 220.
  • the minimum requirements on the pattern in this case are that there are four distinct points in general positions, or any equivalent pattern.
  • the projector's intrinsic parameters can be detenriined as follows. First, the homography H 211 is used to dete ⁇ nine coordinates of the projected pattern features on the planar surface 130. Then, a first homography Gl is dete ⁇ nined between the image plane of the projector 120, and the planar surface for a first position PI of the projector. Similarly a second homography G2 is computed for a second position P2 of the projector. Given Gl and G2, it is possible to dete ⁇ nine the intrinsic parameters 221 of the projector 120 as described by Zhang in "A flexible new technique for camera calibration," IEEE Transactions on Pattern Analysis and Machine Intelligence, 22(11): 1330-1334, 2000.
  • That technique was developed for intrinsic parameters of a camera. However, we apply that technique to determine the intrinsic parameters of a projector.
  • the operation 200 of the system 100 described above uses one fixed camera 110. If the projector 120 is moving across an extended area, so that the projected pattern is likely to move out of the field of view of any one camera, then multiple fixed cameras can be used, each observing a different part of the fixed planar surface 130. In this case, the calibration of the fixed cameras requires the dete ⁇ nination of multiple homographies HI, H2, etc., between the image plane of each camera, and a common coordinate frame of the fixed surface 130.
  • the projector 120 can be moving inside a room, and the ceiling, floor, or walls provide the fixed planar surfaces.
  • a wide-angle or omni-directional projector is projecting simultaneously on multiple surfaces for a wide field of view, all of the projection information can be used simultaneously for determining the position and orientation of the projector 120 to a greater accuracy.
  • the projector 120 can be in a fixed relationship to an object 160 of interest. Then, the position and orientation of the projector can be used to infer a position and orientation of the object 160. If the object has its own coordinate frame, then the projector and object can be calibrated in another calibration step.
  • the object 160 can be a second moving camera.
  • the projector 120 can track the position and orientation of the camera 160. This is useful when the second camera 160 is used for 3D scanning.
  • the inferred positions and orientations of second camera can be used to determine a 3D reconstruction from the scanned images.
  • the object can be another projector.
  • the second projector projects images into the environment according to the inferred position and orientation of the second projector. This is useful for augmenting virtual reality scenes, or active display areas.
  • the object can also be a tool so that feedback dependent on the inferred position and orientation of the tool can be provided to a user.
  • the feedback can be haptic.
  • the projector can be attached to an object, which is moving in 2D, e.g., a car or robot, in which case only the position is relevant, and the orientation is irrelevant.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Signal Processing (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Vascular Medicine (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • Geometry (AREA)
  • Length Measuring Devices By Optical Means (AREA)
  • Image Processing (AREA)
  • Studio Devices (AREA)
  • Image Analysis (AREA)
  • Projection Apparatus (AREA)
PCT/JP2003/016777 2003-01-17 2003-12-25 Position and orientation sensing with a projector WO2004066213A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
DE60323945T DE60323945D1 (de) 2003-01-17 2003-12-25 Bestimmung von position und ausrichtung mit einem projektor
JP2004567159A JP4437748B2 (ja) 2003-01-17 2003-12-25 プロジェクタによる位置および向きの読み取り
EP03768261A EP1483740B1 (en) 2003-01-17 2003-12-25 Position and orientation sensing with a projector

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US10/346,442 US7242818B2 (en) 2003-01-17 2003-01-17 Position and orientation sensing with a projector
US10/346,442 2003-01-17

Publications (1)

Publication Number Publication Date
WO2004066213A1 true WO2004066213A1 (en) 2004-08-05

Family

ID=32712151

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2003/016777 WO2004066213A1 (en) 2003-01-17 2003-12-25 Position and orientation sensing with a projector

Country Status (6)

Country Link
US (1) US7242818B2 (ja)
EP (1) EP1483740B1 (ja)
JP (1) JP4437748B2 (ja)
CN (1) CN100343876C (ja)
DE (1) DE60323945D1 (ja)
WO (1) WO2004066213A1 (ja)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9305361B2 (en) 2011-09-12 2016-04-05 Qualcomm Incorporated Resolving homography decomposition ambiguity based on orientation sensors

Families Citing this family (40)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2004054250A1 (en) * 2002-12-11 2004-06-24 Koninklijke Philips Electronics N.V. Projection system with contrast homogeneity correction
CA2464569A1 (en) * 2003-04-16 2004-10-16 Universite De Montreal Single or multi-projector for arbitrary surfaces without calibration nor reconstruction
DE102004045280A1 (de) * 2004-09-16 2006-04-06 Harold Larivière Kalibrierungsvorrichtung für projizierte Bilder
US8147066B2 (en) * 2004-09-21 2012-04-03 Nikon Corporation Portable information device having a projector and an imaging device
NO323926B1 (no) * 2004-11-12 2007-07-23 New Index As Visuelt system samt styringsobjekt og apparat til bruk i systemet.
KR100714722B1 (ko) * 2005-06-17 2007-05-04 삼성전자주식회사 광원의 신호를 사용하여 포인팅 사용자 인터페이스를구현하는 장치 및 방법
US20070282564A1 (en) * 2005-12-06 2007-12-06 Microvision, Inc. Spatially aware mobile projection
US20070176851A1 (en) * 2005-12-06 2007-08-02 Willey Stephen R Projection display with motion compensation
US20110111849A1 (en) * 2005-12-06 2011-05-12 Microvision, Inc. Spatially Aware Mobile Projection
US7794090B2 (en) * 2006-01-24 2010-09-14 Seiko Epson Corporation Efficient dual photography
US7768527B2 (en) * 2006-05-31 2010-08-03 Beihang University Hardware-in-the-loop simulation system and method for computer vision
JP4697087B2 (ja) * 2006-08-09 2011-06-08 富士ゼロックス株式会社 画像処理装置
JP5281015B2 (ja) * 2006-12-19 2013-09-04 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ 自動立体的表示デバイス及び該デバイスを使用したシステム
US8994757B2 (en) 2007-03-15 2015-03-31 Scalable Display Technologies, Inc. System and method for providing improved display quality by display adjustment and image processing using optical feedback
CN100461220C (zh) * 2007-03-30 2009-02-11 东南大学 三维扫描系统中的基于固定参数与可变参数的标定方法
CN100491903C (zh) * 2007-09-05 2009-05-27 北京航空航天大学 一种结构光视觉传感器结构参数标定方法
KR101265950B1 (ko) * 2007-09-18 2013-05-23 삼성전자주식회사 프로젝터 및 상기 프로젝터의 프로젝션 제어 방법
US20090280901A1 (en) * 2008-05-09 2009-11-12 Dell Products, Lp Game controller device and methods thereof
US8427424B2 (en) 2008-09-30 2013-04-23 Microsoft Corporation Using physical objects in conjunction with an interactive surface
US20100321382A1 (en) 2009-06-18 2010-12-23 Scalable Display Technologies, Inc. System and method for injection of mapping functions
TWI398160B (zh) * 2009-12-01 2013-06-01 Ind Tech Res Inst 攝影機校正系統與座標資料產生系統及其方法
US8730309B2 (en) 2010-02-23 2014-05-20 Microsoft Corporation Projectors and depth cameras for deviceless augmented reality and interaction
EP2641400B1 (en) 2010-11-15 2021-07-21 Scalable Display Technologies, Inc. System for geometrically calibrating a plurality of projectors, method for geometrically calibrating a plurality of projectors and corresponding computer program
US9329469B2 (en) 2011-02-17 2016-05-03 Microsoft Technology Licensing, Llc Providing an interactive experience using a 3D depth camera and a 3D projector
US9480907B2 (en) 2011-03-02 2016-11-01 Microsoft Technology Licensing, Llc Immersive display with peripheral illusions
US9597587B2 (en) * 2011-06-08 2017-03-21 Microsoft Technology Licensing, Llc Locational node device
US9497447B2 (en) 2011-06-15 2016-11-15 Scalable Display Technologies, Inc. System and method for color and intensity calibrating of a display system for practical usage
US8976249B2 (en) 2011-11-04 2015-03-10 Empire Technology Development Llc IR signal capture for images
JP2015519601A (ja) 2012-04-19 2015-07-09 スケーラブル ディスプレイ テクノロジーズ インコーポレイテッド システム入力解像度の変動なくディスプレイシステムを校正するシステムおよび方法
JP6172495B2 (ja) 2012-12-28 2017-08-02 株式会社リコー 校正装置、装置、プロジェクタ、3次元スキャナ、校正方法、方法、プログラム、及び記憶媒体
EP2972585A4 (en) 2013-03-15 2016-10-19 Scalable Display Technologies Inc SYSTEM AND METHOD FOR CALIBRATING DISPLAY SYSTEM USING SHORT-DISTANCE PROJECTION VIEWING APPARATUS
JP2014206634A (ja) 2013-04-12 2014-10-30 船井電機株式会社 電子機器
CN103644842B (zh) * 2013-11-22 2017-01-25 大连日佳电子有限公司 一种侧窥环视的方法和装置
KR20150115455A (ko) 2014-04-04 2015-10-14 삼성전자주식회사 휴대단말기의 프로젝터 초점 제어 방법 및 장치
CN105629199A (zh) * 2014-11-04 2016-06-01 Tcl集团股份有限公司 一种室内定位方法及系统
US10080004B2 (en) 2014-11-06 2018-09-18 Disney Enterprises, Inc. Method and system for projector calibration
EP3238447B1 (en) * 2014-12-22 2021-10-27 Cyberoptics Corporation Updating calibration of a three-dimensional measurement system
EP3158502B1 (en) * 2015-08-10 2018-07-25 Koninklijke Philips N.V. Furniture occupancy monitoring
US9930320B2 (en) 2016-05-04 2018-03-27 Apple Inc. Resolving three dimensional spatial information using time-shared structured lighting that embeds digital communication
US9992464B1 (en) * 2016-11-11 2018-06-05 Christie Digital Systems Usa, Inc. Method and system for screen correction

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4443705A (en) * 1982-10-01 1984-04-17 Robotic Vision Systems, Inc. Method for locating points on a three-dimensional surface using light intensity variations
US4724480A (en) * 1985-05-02 1988-02-09 Robotic Vision Systems, Inc. Method for optical alignment of one object with respect to another
JPH11166818A (ja) * 1997-12-04 1999-06-22 Suzuki Motor Corp 三次元形状計測装置の校正方法及び校正装置
JP4453119B2 (ja) * 1999-06-08 2010-04-21 ソニー株式会社 カメラ・キャリブレーション装置及び方法、画像処理装置及び方法、プログラム提供媒体、並びに、カメラ
JP2001008231A (ja) * 1999-06-22 2001-01-12 Gen Tec:Kk 3次元空間における物体の多視点画像伝送方法及びシステム
JP4507307B2 (ja) * 1999-09-16 2010-07-21 独立行政法人科学技術振興機構 映像投影装置
US6618076B1 (en) * 1999-12-23 2003-09-09 Justsystem Corporation Method and apparatus for calibrating projector-camera system
JP4169464B2 (ja) * 1999-12-28 2008-10-22 株式会社リコー 画像処理方法と画像処理装置及びコンピュータ読み取り可能な記録媒体
US20010041073A1 (en) * 2000-02-03 2001-11-15 Noam Sorek Active aid for a handheld camera
JP2001320652A (ja) * 2000-05-11 2001-11-16 Nec Corp プロジェクタ装置
US20020167537A1 (en) * 2001-05-11 2002-11-14 Miroslav Trajkovic Motion-based tracking with pan-tilt-zoom camera

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
OKATANI T ET AL: "Autocalibration of a projector-screen-camera system: theory and algorithm for screen-to-camera homography estimation", COMPUTER VISION 2003 PROCEEDINGS, 13 October 2003 (2003-10-13), pages 774 - 781, XP010662440 *
RASKAR R ET AL: "Multi-Projector Displays Using Camera-Based Registration", VISUALIZATION '99. PROCEEDINGS SAN FRANCISCO, CA, USA 24-29 OCT. 1999, PISCATAWAY, NJ, USA,IEEE, US, 24 October 1999 (1999-10-24), pages 161 - 522, XP010364967, ISBN: 0-7803-5897-X *
SURATI R J: "Scalable Self-Calibrating Display Technology for Seamless Large-Scale Displays", PH. D. THESIS, MIT, 24 January 1999 (1999-01-24), XP002275230 *
YANG R, GOTZ D, HENSLEY J, TOWLES H: "PixelFlex: A reconfigurable Multi-Projector Display System", VISUALIZATION '01 PROCEEDINGS, 21 October 2001 (2001-10-21) - 26 October 2001 (2001-10-26), pages 167 - 174, XP002275229 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9305361B2 (en) 2011-09-12 2016-04-05 Qualcomm Incorporated Resolving homography decomposition ambiguity based on orientation sensors

Also Published As

Publication number Publication date
US7242818B2 (en) 2007-07-10
JP2006513504A (ja) 2006-04-20
US20040141156A1 (en) 2004-07-22
DE60323945D1 (de) 2008-11-20
CN100343876C (zh) 2007-10-17
JP4437748B2 (ja) 2010-03-24
EP1483740A1 (en) 2004-12-08
CN1745397A (zh) 2006-03-08
EP1483740B1 (en) 2008-10-08

Similar Documents

Publication Publication Date Title
US7242818B2 (en) Position and orientation sensing with a projector
US5856844A (en) Method and apparatus for determining position and orientation
US6417836B1 (en) Computer input device having six degrees of freedom for controlling movement of a three-dimensional object
ES2279823T3 (es) Sistema tactil basado en camaras.
JP4822643B2 (ja) 無線ポインタの光学トラッキングを備えるコンピュータ・プレゼンテーション・システムおよび方法
EP2333639B1 (en) Camera-based touch system and method
EP1215017B1 (en) Robot teaching apparatus
US5818424A (en) Rod shaped device and data acquisition apparatus for determining the position and orientation of an object in space
US7256772B2 (en) Auto-aligning touch system and method
US7502018B2 (en) Projector, electronic whiteboard system using projector and method of acquiring indicated point
US7623250B2 (en) Enhanced shape characterization device and method
JPH03282203A (ja) ターゲット及びこれを用いた三次元位置姿勢計測システム
US20210374978A1 (en) Capturing environmental scans using anchor objects for registration
Ansar et al. Visual and haptic collaborative tele-presence
JP4227037B2 (ja) 撮像システム及び校正方法
Zhu et al. Wii remote–based low-cost motion capture for automated assembly simulation
EP0716389A1 (en) Rod-shaped device for spatial data acquisition and method of determining the position and orientation of an object in space
WO2022172471A1 (ja) 支援システム、画像処理装置、支援方法およびプログラム
Puljiz et al. Inside-out Infrared Marker Tracking via Head Mounted Displays for Smart Robot Programming
JPH07113535B2 (ja) 面傾度測定装置
JPH05329793A (ja) 視覚センサ
KR20230085865A (ko) 위치 추적기
Wieghardt et al. Full Self-calibration of a Hand-Mounted Projector Using Structured Light
Labuz et al. Camera and projector motion for range mapping
JP2001175405A (ja) 3次元位置・姿勢計測装置及びその方法並びに記録媒体

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 2004567159

Country of ref document: JP

AK Designated states

Kind code of ref document: A1

Designated state(s): CN JP

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PT RO SE SI SK TR

WWE Wipo information: entry into national phase

Ref document number: 2003768261

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 20038A03098

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWP Wipo information: published in national office

Ref document number: 2003768261

Country of ref document: EP