EP3627289A1 - Nachverfolgungssystem und nachverfolgungsverfahren mit verwendung davon - Google Patents

Nachverfolgungssystem und nachverfolgungsverfahren mit verwendung davon Download PDF

Info

Publication number
EP3627289A1
EP3627289A1 EP18213371.0A EP18213371A EP3627289A1 EP 3627289 A1 EP3627289 A1 EP 3627289A1 EP 18213371 A EP18213371 A EP 18213371A EP 3627289 A1 EP3627289 A1 EP 3627289A1
Authority
EP
European Patent Office
Prior art keywords
information
relation
ankle
coordinate
sensor
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP18213371.0A
Other languages
English (en)
French (fr)
Inventor
Peter Chou
Chun-Wei Lin
Yi-Kang Hsieh
Chia-Wei Wu
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
XRspace Co Ltd
Original Assignee
XRspace Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by XRspace Co Ltd filed Critical XRspace Co Ltd
Publication of EP3627289A1 publication Critical patent/EP3627289A1/de
Withdrawn legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/012Head tracking input arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • G02B27/0172Head mounted characterised by optical features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0334Foot operated pointing devices

Definitions

  • the present disclosure relates to a tracking system and tracking method.
  • HMDs head-mounted displays
  • the HMDs, the motion controller peripherals or cameras may work together to react to gestures made by the user in the VR/AR/MR/XR environment, so as to simultaneously provide interactions to the user in the VR/AR/MR/XR environment.
  • the body movements of the user are free to change positions as forward/backward (i.e. surge), up/down (i.e. heave), left/right (i.e. sway) translations in three perpendicular axes, which are combined with changes in orientation through rotation about three perpendicular axes.
  • the conventional tracking system in the VR/AR/MR/XR environment is needed to be equipped with at least four sensors, outside-in trackers and cameras to sense the body movements of the user, which results in inconveniences to the user when experiencing the VR/AR/MR/XR environment.
  • the present disclosure provides a tracking system and tracking method to provide a better user experience.
  • the tracking system for tracking a body movement of a user, includes a head-mounted display, abbreviated to HMD, worn on a head of a user for collecting a head information of the user; a first sensor, worn on an ankle of the user for collecting a first ankle information of the user; and a second sensor, worn on another ankle of the user for collecting a second ankle information; wherein a first relation exists between the head information and the first ankle information, a second relation exists between the head information and the second ankle information and a third relation exists between the first ankle information and the second ankle information; wherein the tracking system determines the body movement of the user by the head information, the first ankle information, the second ankle information, the first relation, the second relation and the third relation.
  • HMD head-mounted display
  • the claimed tracking method for a head-mounted display (HMD) system.
  • the HMD virtualizes a body movement of a user in a virtual environment.
  • the tracking method comprises collecting a head information, a first ankle information, a second ankle information of the user by a HMD and a plurality of sensors, wherein a first relation exists between the head information and the first ankle information, a second relation exists between the head information and the second ankle information and a third relation exists between the first ankle information and the second ankle information; and determining the body movement of the user according to the head information, the first ankle information, the second ankle information, the first relation, the second relation and the third relation.
  • FIG. 1 is a schematic diagram of a tracking system 10 according to an embodiment of the present disclosure.
  • the tracking system 10 includes a head-mounted display (HMD) 102 and a plurality of sensors 104.
  • the HMD 102 may be worn on a head of a user for collecting a head information of the user.
  • the sensors 104 may be worn on both ankles of the user for collecting a first ankle information and a second ankle information of the user.
  • the tracking system 10 determines the body movement of the user by the head information, the first ankle information, the second ankle information and relations among them.
  • the determined body movement may be transmitted to the HMD 102, such that the HMD 102 may display the determined body movement in a scene via a virtual avatar.
  • the tracking system 10 of the present disclosure not only obtains the head information and ankle information of the user by the HMD and the sensors, but also utilizes a first relation between the head information and the first ankle information, a second relation between the head information and the second ankle information and a third relation between the first ankle information and the second ankle information as judging factors for determining the body movement, e.g. for calibrating the head information and the ankle information, so as to improve the accuracy of tracking results and improve the user experience.
  • the sensors 104 of the present disclosure includes a first sensor 104_1 and a second sensor 104_2.
  • the first sensor 104_1 and the second sensor 104_2 respectively includes a first inertial measurement unit (IMU) 106_1 and a second IMU 106_2 to simulate a first coordinate of the first sensor 104_1 and simulate a second coordinate of the second sensor 104_2.
  • IMU inertial measurement unit
  • a relative distance and a relative position of the sensors may be obtained by the first coordinate and the second coordinate.
  • the relative distance and the relative position of the sensors may be taken as the third relation of the first and the second ankle information.
  • the third relation may include a vector difference or a momentum difference in 3D movement between the sensors.
  • the HMD includes a third IMU to simulate a third coordinate of the HMD.
  • a relative position and a relative distance are obtained between the HMD and the first sensor by the first coordinate and the third coordinate.
  • the relative position and the relative distance between the HMD and the first sensor may be the first relation.
  • a relative position and a relative distance are obtained between the HMD and the second sensor by the first coordinate and the second coordinate.
  • the relative position and the relative distance between the HMD and the second sensor may be the second relation.
  • the first relation and the second relation may include a vector difference or a momentum difference in 3D movement among the HMD and the two sensors.
  • a magnetic meter or a spherical antenna array may be added to the sensors to calibrate their locations and improve the accuracy of the tracking results of the tracking system 10.
  • the HMD 102 may connect to the sensors 104 respectively via a first connection C1, such that the HMD 102 and the sensors 104 may share locations, 3D movement vectors and momentums to each other.
  • the sensors 104 may connect to each other via a second connection C2, such that the sensors 104 may share locations, 3D movement vectors and momentums to each other.
  • the HMD 102 may calculate the first relation, the second relation and the third relation according to the first connection C1 and the second connection C2, anyone of the sensors 104 may be utilized for calculating the first relation, the second relation and the third relation, such that the tracking system 10 may allocate calculation tasks on the HMD 102 and the sensors 104 based on individual calculation abilities to improve the efficiency of real-time calculation.
  • the first connection C1 and the second connection C2 may be implemented by any one of the wireless communication (e.g. Bluetooth), ultrasonic, laser, or magnetic force.
  • the wireless communication e.g. Bluetooth
  • ultrasonic ultrasonic
  • laser laser
  • magnetic force any one of the wireless communication (e.g. Bluetooth), ultrasonic, laser, or magnetic force.
  • the sensors 104 may be six degrees-of-freedom (DOF) sensors or 3 DOF sensors. Since the present disclosure not only collects the head information and ankle information, but also collects the first relation, the second relation and the third relation to determine the body movement, more accurate tracking results may be obtained no matter by utilizing 6 DOF sensors or 3 DOF sensors.
  • DOF degrees-of-freedom
  • the tracking system 10 of the present disclosure may be implemented in all kinds of methods. Furthermore, the operating process of the tracking system 10 may be concluded to a tracking process 30 as shown in FIG. 3 , which includes the following steps:
  • the details of the tracking process 30 may be referred to the above mentioned embodiments of the tracking system 10 and are not narrated herein for brevity.
  • the present disclosure provides a tracking system 10 and related tracking method, which is utilized for obtaining more precise the tracking results and provides a better user experience when the user experiences the VR/AR/MR/XR environment.

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Optics & Photonics (AREA)
  • User Interface Of Digital Computer (AREA)
  • Position Input By Displaying (AREA)
  • Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
  • Processing Or Creating Images (AREA)
EP18213371.0A 2018-09-19 2018-12-18 Nachverfolgungssystem und nachverfolgungsverfahren mit verwendung davon Withdrawn EP3627289A1 (de)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US16/136,198 US10817047B2 (en) 2018-09-19 2018-09-19 Tracking system and tacking method using the same

Publications (1)

Publication Number Publication Date
EP3627289A1 true EP3627289A1 (de) 2020-03-25

Family

ID=64744605

Family Applications (1)

Application Number Title Priority Date Filing Date
EP18213371.0A Withdrawn EP3627289A1 (de) 2018-09-19 2018-12-18 Nachverfolgungssystem und nachverfolgungsverfahren mit verwendung davon

Country Status (5)

Country Link
US (1) US10817047B2 (de)
EP (1) EP3627289A1 (de)
JP (1) JP6986003B2 (de)
CN (1) CN110928404B (de)
TW (1) TWI735830B (de)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10496808B2 (en) 2016-10-25 2019-12-03 Apple Inc. User interface for managing access to credentials for use in an operation
CN112416125A (zh) 2020-11-17 2021-02-26 青岛小鸟看看科技有限公司 Vr头戴式一体机
US20220269333A1 (en) * 2021-02-19 2022-08-25 Apple Inc. User interfaces and device settings based on user identification
CN117180720B (zh) * 2023-11-07 2024-01-05 成都孚谦科技有限公司 一种基于体感追踪器技术的虚拟动作游戏交互系统及方法

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110009241A1 (en) * 2009-04-10 2011-01-13 Sovoz, Inc. Virtual locomotion controller apparatus and methods
EP3067783A1 (de) * 2015-03-11 2016-09-14 ETH Zurich Verfahren und system zum verfolgen menschlicher fortbewegung durch relative positionsverfolgung von menschlichen körperteilen
EP3270266A2 (de) * 2016-07-15 2018-01-17 HTC Corporation Verfahren, elektronische vorrichtung und aufzeichnungsmedium zur automatischen konfiguration von sensoren

Family Cites Families (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU2001233019A1 (en) 2000-01-28 2001-08-07 Intersense, Inc. Self-referenced tracking
US7542040B2 (en) * 2004-08-11 2009-06-02 The United States Of America As Represented By The Secretary Of The Navy Simulated locomotion method and apparatus
US7636645B1 (en) * 2007-06-18 2009-12-22 Ailive Inc. Self-contained inertial navigation system for interactive control using movable controllers
KR101483713B1 (ko) 2008-06-30 2015-01-16 삼성전자 주식회사 모션 캡쳐 장치 및 모션 캡쳐 방법
JP2010240185A (ja) 2009-04-07 2010-10-28 Kanazawa Inst Of Technology 動作学習支援装置
US20110270135A1 (en) * 2009-11-30 2011-11-03 Christopher John Dooley Augmented reality for testing and training of human performance
US10052035B2 (en) * 2013-10-25 2018-08-21 Qualcomm Incorporated System and method for obtaining bodily function measurements using a mobile device
US10415975B2 (en) 2014-01-09 2019-09-17 Xsens Holding B.V. Motion tracking with reduced on-body sensors set
US9649558B2 (en) * 2014-03-14 2017-05-16 Sony Interactive Entertainment Inc. Gaming device with rotatably placed cameras
WO2016033717A1 (zh) * 2014-09-01 2016-03-10 北京诺亦腾科技有限公司 一种组合式运动捕捉系统
JP6391465B2 (ja) 2014-12-26 2018-09-19 Kddi株式会社 ウェアラブル端末装置およびプログラム
US10073516B2 (en) * 2014-12-29 2018-09-11 Sony Interactive Entertainment Inc. Methods and systems for user interaction within virtual reality scene using head mounted display
WO2016118344A1 (en) * 2015-01-20 2016-07-28 Microsoft Technology Licensing, Llc Fixed size augmented reality objects
US10248188B2 (en) * 2015-06-03 2019-04-02 James M. O'Neil System and method for generating wireless signals and controlling digital responses from physical movement
US10324522B2 (en) 2015-11-25 2019-06-18 Jakob Balslev Methods and systems of a motion-capture body suit with wearable body-position sensors
US10353489B2 (en) * 2016-04-13 2019-07-16 Seiko Epson Corporation Foot input device and head-mounted display device
DE102016212236A1 (de) * 2016-07-05 2018-01-11 Siemens Aktiengesellschaft Interaktionssystem und -verfahren
JPWO2018143360A1 (ja) 2017-02-03 2019-12-26 良夫 川又 相対位置検出システム及び画像表示システム
CN206819290U (zh) * 2017-03-24 2017-12-29 苏州创捷传媒展览股份有限公司 一种虚拟现实多人互动的系统
WO2018208924A1 (en) * 2017-05-09 2018-11-15 Google Llc Augmented and/or virtual reality footwear
US10379613B2 (en) * 2017-05-16 2019-08-13 Finch Technologies Ltd. Tracking arm movements to generate inputs for computer systems
US11016116B2 (en) * 2018-01-11 2021-05-25 Finch Technologies Ltd. Correction of accumulated errors in inertial measurement units attached to a user

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110009241A1 (en) * 2009-04-10 2011-01-13 Sovoz, Inc. Virtual locomotion controller apparatus and methods
EP3067783A1 (de) * 2015-03-11 2016-09-14 ETH Zurich Verfahren und system zum verfolgen menschlicher fortbewegung durch relative positionsverfolgung von menschlichen körperteilen
EP3270266A2 (de) * 2016-07-15 2018-01-17 HTC Corporation Verfahren, elektronische vorrichtung und aufzeichnungsmedium zur automatischen konfiguration von sensoren

Also Published As

Publication number Publication date
US20200089311A1 (en) 2020-03-19
CN110928404A (zh) 2020-03-27
JP2020047239A (ja) 2020-03-26
US10817047B2 (en) 2020-10-27
TW202013147A (zh) 2020-04-01
CN110928404B (zh) 2024-04-19
JP6986003B2 (ja) 2021-12-22
TWI735830B (zh) 2021-08-11

Similar Documents

Publication Publication Date Title
EP3627289A1 (de) Nachverfolgungssystem und nachverfolgungsverfahren mit verwendung davon
JP6690014B2 (ja) ヘッドマウントディスプレイ追跡
US11353709B2 (en) Sensor fusion for electromagnetic tracking
US10540006B2 (en) Tracking torso orientation to generate inputs for computer systems
EP3486707B1 (de) Wahrnehmungsbasierte prädiktive verfolgung für kopfmontierte anzeigen
US9785249B1 (en) Systems and methods for tracking motion and gesture of heads and eyes
US20150070274A1 (en) Methods and systems for determining 6dof location and orientation of head-mounted display and associated user movements
WO2016041088A1 (en) System and method for tracking wearable peripherals in augmented reality and virtual reality applications
WO2018140129A1 (en) Systems and methods for tracking a controller
CN108700941A (zh) 用于在虚拟现实环境中对准组件的方法和设备
US20210183343A1 (en) Content Stabilization for Head-Mounted Displays
CN108427479B (zh) 可穿戴设备、环境图像数据的处理系统、方法及可读介质
JP2022518779A (ja) 6自由度の姿勢測定値における半球の曖昧性を解決するための方法およびシステム
CN111420391A (zh) 一种头戴显示系统及其空间定位方法
US20230321536A1 (en) Controlling Movement of a Virtual Character in a Virtual Reality Environment
EP4185938A1 (de) Vorrichtungsverfolgung mit ankunftswinkeldaten
CN109634427B (zh) 基于头部追踪的ar眼镜控制系统及控制方法
US11845001B2 (en) Calibration system and method for handheld controller
US20230260210A1 (en) Computer, method, and computer-readable medium
EP4254139B1 (de) Reprojektionsverfahren zur erzeugung reprojizierter bilddaten, xr-projektionssystem und maschinenlernmodul
US20240112421A1 (en) System and method of object tracking for extended reality environment
WO2021177132A1 (ja) 情報処理装置、情報処理システム、情報処理方法及びプログラム
Chen et al. The Body Sensor Suit with Mixed Reality Interactive Games
Hutson et al. Using commodity accelerometers and gyroscopes to improve speed and accuracy of JanusVF
Mahadik et al. 3D VIRTUAL REALITY GAME

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

AX Request for extension of the european patent

Extension state: BA ME

17P Request for examination filed

Effective date: 20200228

RBV Designated contracting states (corrected)

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

17Q First examination report despatched

Effective date: 20200422

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION HAS BEEN WITHDRAWN

18W Application withdrawn

Effective date: 20200811