KR20140104539A - The infrared sensor is equipped with a 3-dimensional motion recognition input devices - Google Patents
The infrared sensor is equipped with a 3-dimensional motion recognition input devices Download PDFInfo
- Publication number
- KR20140104539A KR20140104539A KR1020130017107A KR20130017107A KR20140104539A KR 20140104539 A KR20140104539 A KR 20140104539A KR 1020130017107 A KR1020130017107 A KR 1020130017107A KR 20130017107 A KR20130017107 A KR 20130017107A KR 20140104539 A KR20140104539 A KR 20140104539A
- Authority
- KR
- South Korea
- Prior art keywords
- infrared sensor
- input device
- sensor
- terminal
- motion recognition
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0346—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/038—Control and interface arrangements therefor, e.g. drivers or device-embedded control circuitry
- G06F3/0383—Signal control means within the pointing device
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Position Input By Displaying (AREA)
Abstract
Description
The present invention relates to an input device capable of realizing various display functions and control functions by interfacing various data and measurement values for three-dimensional operation recognized and extracted through various sensors and AR marker functions with a 2D or 3D program (system) Development.
In recent years, 3D or 3D representation technology and control technology have been greatly emphasized, among which technologies related to recognition or expression using various sensors are most interested. In addition
The present invention intends to develop and invent a new 3D motion recognition system by combining sensor technology and AR marker recognition technology.
The present invention can recognize a motion of a hand in a three-dimensional space in real time, and can calculate and extract related data values such as a position, a slope, and a velocity of the motion in real time, The present invention relates to the development of an input device capable of presenting various programs or systems, techniques and control techniques to be realized in a web or a 2D format by enabling realization to be realized by realizing 3D images or images, We will design a new method for the mounting of sensors, calculation of spatial coordinates, and utilization of data values.
An infrared sensor, a PCB circuit, a wired / wireless communication, and the like are basically required for the cam pen and the pad to recognize the three-dimensional motion and to extract the data of the spatial coordinates, and the calculation method of the spatial coordinates through the infrared sensor is shown in FIG. The proposed method uses the trigonometric equation as shown in Fig. 3. As shown in Fig. 3, the motion of the user in the three-dimensional space is defined as motion -> recognition -> transfer -> processing -> output phase In order to determine the location through the infrared sensor, the spatial coordinates of the light emitting part of the cam pen and the receiving part of the pad are calculated, and the spatial coordinate value is obtained. Then, it is transmitted to the final output part through a separate process.
The three-dimensional input device, which is a final object of the present invention, is an apparatus that can recognize and process the position (spatial coordinates) in real time through interaction between a camping pen having an infrared sensor light emitting unit and a pad having an infrared sensor receiving unit, And 3D images and images can be expressed or controlled by a simple operation in a 3D space.
1 is a diagram showing a method of calculating spatial coordinates.
Figure 2 is a plan view of a cam pen and pad prospective design.
Figure 3 is a front view of the cam pan and pad projected schematic.
4 is a system flow diagram of a three-dimensional input device;
FIG. 4 is a summary of the three-dimensional motion recognition and processing procedure of the developed apparatus. When the intended operation of the user is presented within the use range of the first apparatus, the corresponding operation is recognized through the sensor or camera installed in the second apparatus The corresponding operation is transmitted to the system by the third device program, the corresponding value is processed by the fourth sensor or the marker recognition program, and the fifth corresponding value is transmitted through wired / wireless communication.
In order to realize the present invention among the basic systems of the present input device as described above, one light emitting portion of the infrared sensor is installed in the front portion of the cam pen as shown in Figs. 2 to 3, and four infrared sensor receiving portions And the distance between AD and BC, AB and CD in which the infrared sensor is installed is the same, and the triangle AZC and BZD are joint triangles as shown in FIG. 1, and the angle of the vertex B can be measured according to the trigonometric function SIN value.
At this time, the formula for calculating the angle is B ° = sin -1 (BZ / ZV), AD = BC, AB = CD, and Z is the position of the light emitting portion.
Further, it can be calculated as a three-way joint of? AZC and? BZD.
1: cam fan 2: camera and infrared receiver
3: cam fan button 4: USB terminal
5: additional function button 6: pad body
7: Jog marker 8: Marker pointer
9: Infrared pointer
Claims (2)
Wherein one device is provided with one light emitting part of the infrared sensor and the other device is provided with four receiving parts to realize mutual communication of the sensors.
The physical distance between the light emitting unit of the infrared sensor and the receiving unit of the three-dimensional motion recognition input device equipped with the infrared sensor can be measured by a trigonometric function and can be calculated and processed in real time. Wherein the infrared sensor is connected to the infrared sensor.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020130017107A KR20140104539A (en) | 2013-02-18 | 2013-02-18 | The infrared sensor is equipped with a 3-dimensional motion recognition input devices |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020130017107A KR20140104539A (en) | 2013-02-18 | 2013-02-18 | The infrared sensor is equipped with a 3-dimensional motion recognition input devices |
Publications (1)
Publication Number | Publication Date |
---|---|
KR20140104539A true KR20140104539A (en) | 2014-08-29 |
Family
ID=51748288
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
KR1020130017107A KR20140104539A (en) | 2013-02-18 | 2013-02-18 | The infrared sensor is equipped with a 3-dimensional motion recognition input devices |
Country Status (1)
Country | Link |
---|---|
KR (1) | KR20140104539A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20170078340A (en) | 2015-12-29 | 2017-07-07 | 동서대학교산학협력단 | User identification system for coordinate recognition of IR type input device |
CN112166401A (en) * | 2018-06-11 | 2021-01-01 | 三星电子株式会社 | Input device and electronic device including the same |
-
2013
- 2013-02-18 KR KR1020130017107A patent/KR20140104539A/en not_active Application Discontinuation
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20170078340A (en) | 2015-12-29 | 2017-07-07 | 동서대학교산학협력단 | User identification system for coordinate recognition of IR type input device |
CN112166401A (en) * | 2018-06-11 | 2021-01-01 | 三星电子株式会社 | Input device and electronic device including the same |
CN112166401B (en) * | 2018-06-11 | 2024-04-26 | 三星电子株式会社 | Input device and electronic device including the same |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP7336184B2 (en) | Systems, methods, and tools for spatially aligning virtual content with a physical environment in an augmented reality platform | |
CN108700939B (en) | System and method for augmented reality | |
CN109313502B (en) | Tap event location using selection device | |
US20140368539A1 (en) | Head wearable electronic device for augmented reality and method for generating augmented reality using the same | |
US20170140552A1 (en) | Apparatus and method for estimating hand position utilizing head mounted color depth camera, and bare hand interaction system using same | |
WO2014141504A1 (en) | Three-dimensional user interface device and three-dimensional operation processing method | |
US20150212583A1 (en) | Detection of user gestures | |
WO2012088285A3 (en) | Three-dimensional tracking of a user control device in a volume | |
WO2012111998A3 (en) | Virtual touch device without pointer | |
EP2846308A3 (en) | Pointing direction detecting device and its method, program and computer readable-medium | |
JP2013235373A5 (en) | ||
WO2012020696A8 (en) | Device for processing point cloud position data, system for processing point cloud position data, method for processing point cloud position data and program for processing point cloud position data | |
WO2004097612A3 (en) | A man-machine interface based on 3-d positions of the human body | |
KR101441882B1 (en) | method for controlling electronic devices by using virtural surface adjacent to display in virtual touch apparatus without pointer | |
CN102508578A (en) | Projection positioning device and method as well as interaction system and method | |
CN104620201A (en) | Apparatus for obtaining virtual 3d object information without requiring pointer | |
WO2012106370A3 (en) | Correlating areas on the physical object to areas on the phone screen | |
WO2017021902A1 (en) | System and method for gesture based measurement of virtual reality space | |
KR20160073025A (en) | Object generation apparatus and method of based augmented reality using actual measured | |
TWI688744B (en) | A measuring device and a measuring method for measuring three-dimensional coordinates of points on surface of an object | |
JP2004265222A (en) | Interface method, system, and program | |
KR20130082296A (en) | Apparatus and method for moving in virtual reality | |
TWI486815B (en) | Display device, system and method for controlling the display device | |
KR20140104539A (en) | The infrared sensor is equipped with a 3-dimensional motion recognition input devices | |
KR101530340B1 (en) | Motion sensing system for implementing hand position-posture information of user in a three-dimensional virtual space based on a combined motion tracker and ahrs system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WITN | Withdrawal due to no request for examination |