WO2014112782A1 - 트랙킹 시스템 및 이를 이용한 트랙킹 방법 - Google Patents
트랙킹 시스템 및 이를 이용한 트랙킹 방법 Download PDFInfo
- Publication number
- WO2014112782A1 WO2014112782A1 PCT/KR2014/000426 KR2014000426W WO2014112782A1 WO 2014112782 A1 WO2014112782 A1 WO 2014112782A1 KR 2014000426 W KR2014000426 W KR 2014000426W WO 2014112782 A1 WO2014112782 A1 WO 2014112782A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- markers
- lens array
- array unit
- lenses
- dimensional coordinates
- Prior art date
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/313—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor for introducing through surgical openings, e.g. laparoscopes
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/70—Manipulators specially adapted for use in surgery
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
- A61B2034/2046—Tracking techniques
- A61B2034/2055—Optical tracking systems
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
- A61B2034/2046—Tracking techniques
- A61B2034/2055—Optical tracking systems
- A61B2034/2057—Details of tracking cameras
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
- A61B2034/2046—Tracking techniques
- A61B2034/2065—Tracking using image or pattern recognition
Definitions
- the present invention relates to a tracking system and a tracking method using the same, and more particularly, to a surgical tracking system for detecting the spatial position information and the direction information of the target by tracking the coordinates of the markers attached to the target, such as surgical instruments of the affected area; It relates to a tracking method using the same.
- Surgical navigation as described above includes a tracking system that can accurately detect and detect the spatial position and direction of the object, such as the affected area or surgical instruments as described above.
- Such a tracking system is typically connected to markers attached to an object such as an affected part or a surgical tool, first and second imaging units for imaging light emitted by the markers, and the first and second imaging units.
- the space of the object is compared with the three-dimensional coordinates of the markers by comparing information of straight lines connecting previously stored markers with each other and angle information formed by a pair of neighboring straight lines. It includes a processor for calculating position and direction.
- the coordinates of the markers emitted from one marker and formed in the first imaging unit and the coordinates of the markers formed in the second imaging unit are the same.
- two detectors were necessary for the three-dimensional coordinates of each marker to be calculated by the processor through triangulation.
- the conventional general tracking system must include two imaging units for imaging light emitted from the respective markers at different positions, thereby increasing the manufacturing cost and increasing the overall size of the system, thereby limiting the surgical space. There was a problem to receive a lot.
- an object of the present invention is to provide a tracking system that can calculate the three-dimensional coordinates of each marker with only one imaging unit to reduce the manufacturing cost and to compact the equipment to minimize the constraint of the surgical space and It relates to a tracking method using the same.
- a tracking system includes a lens array unit having at least three markers attached to a target to emit light, at least two lenses passing light emitted from the markers at predetermined intervals, An imaging unit that receives light emitted from the markers and passes through each lens of the lens array unit, and forms an image corresponding to the number of lenses of the lens array unit per marker, and the lens per marker in the imaging unit.
- the 3D coordinates of the respective markers are calculated using the marker images formed by the number of lenses corresponding to the number of lenses of the array unit, and then the geometric information is compared between the 3D coordinates of the markers and previously stored neighboring markers.
- a processor for calculating the spatial position and orientation of the object.
- the markers may be active markers that emit light in themselves.
- it may further include at least one light source for emitting light from the lens array unit side to the markers, in this case the markers are passive markers for reflecting light emitted from the light source to the lens array unit side Can be.
- the imaging unit may be a camera that receives light emitted from the markers and passes through each lens of the lens array unit and forms at least two images corresponding to the number of lenses of the lens array unit per marker. Can be.
- the geometric information between the markers may be length information of straight lines connecting the neighboring markers and angle information formed by the pair of neighboring straight lines.
- a tracking method comprises the steps of emitting light from at least three markers attached to an object, and the light emitted from the markers passes through at least two lenses of the lens array unit to the imaging unit. Imaging each number of images corresponding to the number of lenses of the lens array unit, and using the marker images imaged by the number corresponding to the number of lenses of the lens array unit per marker in the imaging unit through the processor. Computing three-dimensional coordinates of the markers, Comparing the geometric information between the three-dimensional coordinates of the respective markers and the neighboring markers previously stored in the processor to calculate the spatial position and direction of the object Can be.
- the geometric information between the markers may be length information of straight lines connecting the neighboring markers and angle information formed by the pair of neighboring straight lines.
- the calculating of the three-dimensional coordinates of the markers may include calculating, by the processor, two-dimensional coordinates of the number of marker images corresponding to the number of lenses of the lens array unit per marker formed in the imaging unit. Computing the three-dimensional coordinates of the markers by the processor using the two-dimensional coordinates of the number of the marker image corresponding to the number of lenses of the lens array unit per marker.
- the emitting of the light may emit light generated by the markers to the lens array unit side.
- the step of emitting the light may reflect light emitted from at least one light source to the lens array unit through the markers.
- the spatial position and direction of the light source is pre-stored in the processor.
- the light emitted from the respective markers passes through the lens array unit having at least one pair of lenses, and the lenses per marker in the imaging unit.
- the three-dimensional coordinates of the markers can be calculated using triangulation using only one imaging unit to calculate the spatial position and orientation of the target object.
- FIG. 1 is a schematic diagram of a tracking system according to an embodiment of the present invention.
- FIG. 2 is an exemplary view in which markers are attached to an object
- 3 is an exemplary view for explaining a change in the position where the image of the marker image when the position of the marker is changed on the same optical path of the lens
- FIG. 4 is a block diagram illustrating a tracking method according to an embodiment of the present invention.
- FIG. 5 is a block diagram illustrating a process of calculating three-dimensional coordinates of markers.
- FIG. 6 is an exemplary diagram in which an image sensor of an imaging unit is virtually divided into coordinate systems of first and second marker images.
- FIG. 7 is a diagram for explaining a relationship between two-dimensional coordinates in an image and three-dimensional coordinates of an actual marker.
- first and second may be used to describe various components, but the components should not be limited by the terms. The terms are used only for the purpose of distinguishing one component from another.
- the first component may be referred to as the second component, and similarly, the second component may also be referred to as the first component.
- a tracking system and a tracking method using the same include attaching at least three markers to an object such as an affected part or a surgical tool, and then calculates three-dimensional coordinates of the markers and neighboring markers previously stored in a processor.
- a processor calculates the spatial position and direction of the target object such as the affected part or surgical instruments.
- FIG. 1 is a schematic diagram of a tracking system according to an embodiment of the present invention
- FIG. 2 is an exemplary diagram in which markers are attached to an object
- FIG. 3 is an image of a marker image formed when the position of the marker is changed on the same optical path of the lens. It is an illustration for demonstrating the change of the position to become.
- the tracking system 100 includes at least three markers 110, 111, 112, a lens array unit 120, and an imaging system.
- the unit 130 and the processor 140 may be included.
- the lens array unit 120 may be mounted in the imaging unit 130.
- the at least three markers 110, 111, 112 are attached to a target 200, such as an affected part or surgical tool.
- the at least three markers 110, 111, 112 are spaced apart from each other by the markers 110, 111, 112 that are adjacent to each other, and the markers 110, 111, which are adjacent to each other ( The pair of straight lines L1, L2, and L3 adjacent to each marker are virtually connected to each other so as to form a predetermined angle A1, A2, A3, and the like. Is attached to 200.
- the length information and the angles A1, A2, and A3 of a pair of neighboring straight lines connecting neighboring markers 110, 111, and 112 to each other are stored in the processor 140. (memory: 141) is already stored.
- the markers 110, 111, and 112 may be attached in a triangular form to the target 200, such as affected areas or surgical instruments, the three markers 110, 111, ( Length information of each of the straight lines L1, L2, and L3 constituting the sides of the triangle having the vertex 112 as a vertex, and a pair of adjacent straight lines connecting the markers 110, 111, and 112 to each other.
- the angle A1, A2, and A3 information may be stored in the memory 141 included in the processor 140.
- the markers 110, 111, and 112 may be active markers that emit light by themselves. As described above, when the markers 110, 111, and 112 are used as active markers, there is no need to use a separate light source.
- the markers 110, 111, 112 may be passive markers that reflect light emitted from at least one light source 150.
- the lens includes at least one light source 150 that emits light to the markers 110, 111, and 112. It may be disposed around the array unit 120. For example, a pair of light sources 150 may be disposed on both sides of the lens array unit 120.
- the spatial position and direction of the light source 150 is pre-stored in the memory 141 mounted in the processor 140.
- the lens array unit 120 is disposed at the front of the imaging unit 130.
- the lens array unit 120 is formed by arranging at least a pair of lenses 121 and 122 for passing light emitted from the markers 110, 111 and 112 at predetermined intervals.
- the lens array unit 120 may be formed by arranging the first lens 121 and the second lens 122 at predetermined intervals.
- the lens array unit 120 in which the first and second lenses 121 and 122 are arranged at predetermined intervals is illustrated, but the lens array unit 120 may be formed by arranging three or more lenses at predetermined intervals. It may be.
- the imaging unit 130 receives the light emitted from the markers 110, 111, 112 and passed through each lens 121, 122 of the lens array unit 120, and the lens per marker. An image corresponding to the number of lenses of the array unit 120 is imaged.
- the imaging unit 130 may include the markers 110 and 111.
- the light is emitted from the 112 and passes through the first and second lenses 121 and 122 to form a pair of marker images per marker.
- the imaging unit 130 receives the light emitted from the markers 110, 111, 112 and passed through each lens 121, 122 of the lens array unit 120. It may be a camera mounted with an image sensor 131 for forming an image corresponding to the number of lenses 121 and 122 of the lens array unit 120 per marker.
- the processor 140 uses the marker images formed by the number corresponding to the number of lenses 121 and 122 of the lens array unit 120 per marker, so that the respective markers 110, 111 and 112 are used. Calculating three-dimensional coordinates of the markers, and comparing the three-dimensional coordinates of the markers 110, 111, and 112 with geometric information between previously stored neighboring markers 110, 111, 112; Calculate the spatial position and direction of the object 200, such as the affected portion 110, 120, 130, surgical instruments.
- the memory 141 is mounted in the processor 140.
- the memory 141 mounted in the processor 140 includes geometric information between the neighboring markers 110, 111, and 112, that is, the neighboring markers 110, 111, 112. Angles A1 and A2 formed by the length information of the straight lines L1, L2, and L3 to be connected and a pair of neighboring straight lines to connect the markers 110, 111 and 112 that are adjacent to each other.
- A3) Information may be stored in advance.
- the memory 141 mounted in the processor 140 may have spatial positions and directions of the at least one pair of light sources 150. Can be stored.
- the tracking system 100 uses the lens array unit 120 in which at least one pair of lenses 121 and 122 are arranged at predetermined intervals. Only one imaging unit 130 is used by allowing light emitted from the (111) and 112 to pass through at least one pair of lenses 121 and 122 so that at least one pair of marker images per marker are imaged in the imaging unit. And there is an advantage that can calculate the three-dimensional coordinates of each marker.
- the positions of the markers 110, 111, and 112 are changed on the same optical axis AX of the lens 131, the image of the second lens 122 is imaged.
- the position of the sensor 133 is not changed, but the position of the image sensor 133 of the first lens 121 image is changed so that each of the markers 110, 111, and 112 may be used even when only one imaging unit is used.
- Three-dimensional coordinates can be calculated using trigonometry.
- FIG. 4 is a block diagram illustrating a tracking method according to an embodiment of the present invention
- FIG. 5 is a block diagram illustrating a process of calculating three-dimensional coordinates of markers
- FIG. 6 is an image sensor of an imaging unit.
- FIG. 7 is an exemplary diagram of virtual division into a coordinate system of first and second marker images
- FIG. 7 is a diagram for describing a relationship between two-dimensional coordinates in an image and three-dimensional coordinates of an actual marker.
- At least three markers attached to the object 200 Activate the 110, 111, 112 to emit light from the markers 110, 111, 112, or operate at least one light source 150 to operate the object 200 from the light source 150. At least three markers 110, 111, and 112 attached to the light are irradiated to emit light reflected by the markers 110, 111, and 112. (S110)
- the markers 110, 111, 112 when at least three active markers 110, 111, and 112 that emit light from the object 200 itself are attached, the markers 110, 111, 112 are replaced. Activation so that light is emitted from the markers 110, 111, 112.
- the at least one light source 150 is operated to operate the light source 150. Irradiates light from at least three passive markers 110, 111, 112 attached to the object 200 from the light so that the light is reflected and emitted by the passive markers 110, 111, 112. do.
- the first marker 110 may be used.
- the light emitted from the light passes through the first lens 121 and the second lens 122 through the first optical axis AX1 and the second optical axis AX2, respectively, and a pair of first marker images are formed on the imaging unit 130.
- the light emitted from the second marker 111 passes through the first lens 121 and the second lens 122 through the third optical axis AX3 and the fourth optical axis AX4, respectively, to form an imaging unit (
- a pair of second marker images are formed at 130, and the light emitted from the third marker 112 passes through the first lens 121 and the second through the fifth optical axis AX5 and the sixth optical axis AX6, respectively.
- a pair of third marker images are imaged through the lens 122 and formed in the imaging unit 130.
- each of the markers 110, 111 and 112 in the imaging unit 130 is used.
- a pair of marker images is formed per image.
- the imaging unit 130 When the number of marker images corresponding to the number of lenses 121 and 122 of the lens array unit 120 is formed in the imaging unit 130 for each marker 110, 111, 112, the imaging unit 130. Each of the markers 110 and 111 through the processor 140 using the marker image formed by the image corresponding to the number of lenses 121 and 122 of the lens array unit 120 per marker at 130. Calculate the three-dimensional coordinates of (112). (S130)
- the imaging unit may be formed for each coordinate system (by the first lens image FOV and the second lens image FOV).
- the camera calibration of 130 is performed (S132).
- the processor After performing the camera calibration of the imaging unit 130 for each coordinate system as described above, the processor uses the two-dimensional coordinates of the marker image formed by pairs for each of the markers 110, 111, and 112. The three-dimensional coordinates of the respective markers 110, 111, and 112 are calculated through 140 (S133).
- one side of the image sensor 133 is virtually divided into a field of view (FOV) of the first lens image, and the other side is referred to as an FOV of the second lens image, and a first portion of the image sensor 133 is formed.
- Two-dimensional coordinates of the lens image are displayed in a (U, V) coordinate system, and two-dimensional coordinates of the second lens image are denoted as (U ', V').
- the markers 110 in the image are displayed.
- the two-dimensional coordinates of the 111 and 112 and the three-dimensional coordinates of the markers 110 and 111 and 112 in the real space may be expressed by a relational expression as shown in Equation 1 below.
- m is the two-dimensional coordinates of the marker in the image
- M is the three-dimensional coordinates of the marker in real space
- a (R, t) is the camera matrix
- the three-dimensional coordinates of the actual markers 110, 111, 112 are X
- the relation of the two-dimensional coordinates (x R ) of the marker may be expressed as in Equation 2.
- P 1 is a camera matrix of an image passing through the first lens
- P 2 is a camera matrix of the image passing through the second lens
- P iT is the row vector of the matrix P.
- the respective markers After calculating the three-dimensional coordinates in the real space of the respective markers 110, 111 and 112 by the processor 140 as described above with reference to FIGS. 4 to 5, the respective markers ( Compare the three-dimensional coordinates in the real space of the 110 (111) 112 and the geometric information between the neighboring markers (110, 111, 112) previously stored in the processor 140 through the processor 140 By calculating the spatial position and direction of the object 200 to which the markers 110, 111, 112 are attached.
- the geometric information between the neighboring markers (110, 111, 112) is a straight line (L1) (L2) connecting the neighboring markers (110, 111, 112) as described above Length information of L3 and angle A1, A2, and A3 information formed by a pair of adjacent straight lines connecting the markers 110, 111, and 112.
- the light emitted from the respective markers (110, 111, 112) at least a pair of lenses (121, 122)
- a number of marker images are imaged.
- the lens array unit 120 having a pair of first and second lenses 121 and 122 when used, two images are formed in the imaging unit 130 through two optical axes per marker. Only the large imaging unit 130 may calculate the three-dimensional coordinates of the markers through trigonometry.
- the tracking system and the tracking method using the same can be used to determine the spatial position and direction of the markers 110, 111, 112 attached to the object 200 with only one imaging unit 130.
- the manufacturing cost of the tracking system can be reduced and the light weight can be achieved. Therefore, compared with the conventional tracking system, there is an advantage that the operation space is less restricted.
Abstract
Description
Claims (11)
- 목적물에 부착되어 광을 방출하는 적어도 3개의 마커;상기 마커들로부터 방출되는 광을 통과시키는 적어도 두 개의 렌즈가 소정 간격으로 배열된 렌즈 어레이 유닛;상기 마커들로부터 방출되어 상기 렌즈 어레이 유닛의 각 렌즈를 통과한 광을 받아 들여 마커 당 상기 렌즈 어레이 유닛의 렌즈 개수와 상응하는 수의 상을 결상시키는 결상 유닛; 및상기 결상 유닛에 마커 당 상기 렌즈 어레이 유닛의 렌즈 개수와 상응하는 수만큼 결상된 마커 영상을 이용하여 상기 각각의 마커들의 3차원 좌표를 산출한 후 상기 마커들의 3차원 좌표와 기 저장된 서로 이웃하는 마커들 간의 기하학적 정보를 비교하여 상기 목적물의 공간 위치와 방향을 산출하는 프로세서를 포함하는 트랙킹 시스템.
- 제 1 항에 있어서,상기 마커들은 자체에서 광을 방출하는 액티브 마커인 것을 특징으로 하는 트랙킹 시스템.
- 제 1 항에 있어서,상기 렌즈 어레이 유닛 측에서부터 상기 마커들로 광을 방출하는 적어도 하나의 광원을 더 포함하며,상기 마커들은 상기 광원으로부터 방출된 광을 상기 렌즈 어레이 유닛 측으로 반사시키는 패시브 마커인 것을 특징으로 하는 트랙킹 시스템.
- 제 1 항에 있어서,상기 결상 유닛은,상기 마커들로부터 방출되어 상기 렌즈 어레이 유닛의 각 렌즈를 통과한 광을 받아 들여 마커 당 상기 렌즈 어레이 유닛의 렌즈 개수와 상응하는 적어도 두 개의 상을 결상시키는 카메라인 것을 특징으로 하는 트랙킹 시스템.
- 제 1 항에 있어서,상기 마커들 간의 기하학적 정보는,상기 서로 이웃하는 마커들을 연결하는 직선들의 길이 정보와,상기 서로 이웃하는 한 쌍의 직선이 이루는 각도 정보인 것을 특징으로 하는 트랙킹 시스템.
- 목적물에 부착된 적어도 3개의 마커들로부터 광을 방출하는 단계;상기 마커들로부터 방출된 광이 렌즈 어레이 유닛의 적어도 두 개의 렌즈를 통과하여 결상 유닛에 상기 렌즈 어레이 유닛의 렌즈 개수와 상응하는 수의 상을 결상시키는 단계;상기 결상 유닛에 마커 당 상기 렌즈 어레이 유닛의 렌즈 개수와 상응하는 수만큼 결상된 마커 영상을 이용하여 프로세서를 통해 상기 각각의 마커들의 3차원 좌표를 산출하는 단계; 및상기 각각의 마커들의 3차원 좌표와 상기 프로세서에 기 저장된 서로 이웃하는 마커들 간의 기하학적 정보를 비교하여 상기 목적물의 공간 위치와 방향을 산출하는 단계를 포함하는 트랙킹 방법.
- 제 6 항에 있어서,상기 마커들 간의 기하학적 정보는,상기 서로 이웃하는 마커들을 연결하는 직선들의 길이 정보와,상기 서로 이웃하는 한 쌍의 직선이 이루는 각도 정보인 것을 특징으로 하는 트랙킹 방법.
- 제 6 항에 있어서,상기 마커들의 3차원 좌표를 산출하는 단계는,상기 프로세서를 통해 상기 결상 유닛에 결상된 마커 당 상기 렌즈 어레이 유닛의 렌즈 개수와 상응하는 수의 마커 영상의 2차원 좌표를 산출하는 단계; 및상기 마커당 상기 렌즈 어레이 유닛의 렌즈 개수와 상응하는 수의 마커 영상의 2차원 좌표를 이용하여 상기 프로세서를 통해 상기 마커들의 3차원 좌표를 산출하는 단계를 포함하는 트랙킹 방법.
- 제 6 항에 있어서,상기 광을 방출하는 단계는,상기 마커들이 자체적으로 발생시키는 광을 상기 렌즈 어레이 유닛 측으로 방출하는 것을 특징으로 하는 트랙킹 방법.
- 제 6 항에 있어서,상기 광을 방출하는 단계는,적어도 하나의 광원으로부터 방출되는 광을 상기 마커들을 통해 렌즈 어레이 유닛 측으로 반사시켜 방출하는 것을 특징으로 하는 트랙킹 방법.
- 제 10 항에 있어서,상기 광원의 공간 위치와 방향은 상기 프로세서에 기 저장되는 것을 특징으로 하는 트랙킹 방법.
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/372,307 US20160270860A1 (en) | 2013-01-18 | 2014-01-15 | Tracking system and tracking method using the same |
EP14740207.7A EP2946741A4 (en) | 2013-01-18 | 2014-01-15 | TRACKING SYSTEM AND TRACKING METHOD USING THE SAME |
CN201480004953.3A CN104936547A (zh) | 2013-01-18 | 2014-01-15 | 跟踪系统及利用上述跟踪系统的跟踪方法 |
JP2015553651A JP2016515837A (ja) | 2013-01-18 | 2014-01-15 | トラッキングシステム及びこれを用いたトラッキング方法 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020130005807A KR101371387B1 (ko) | 2013-01-18 | 2013-01-18 | 트랙킹 시스템 및 이를 이용한 트랙킹 방법 |
KR10-2013-0005807 | 2013-01-18 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2014112782A1 true WO2014112782A1 (ko) | 2014-07-24 |
Family
ID=50647855
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2014/000426 WO2014112782A1 (ko) | 2013-01-18 | 2014-01-15 | 트랙킹 시스템 및 이를 이용한 트랙킹 방법 |
Country Status (6)
Country | Link |
---|---|
US (1) | US20160270860A1 (ko) |
EP (1) | EP2946741A4 (ko) |
JP (1) | JP2016515837A (ko) |
KR (1) | KR101371387B1 (ko) |
CN (1) | CN104936547A (ko) |
WO (1) | WO2014112782A1 (ko) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105791800A (zh) * | 2014-12-29 | 2016-07-20 | 深圳超多维光电子有限公司 | 立体显示系统及立体显示方法 |
CN105812776A (zh) * | 2014-12-29 | 2016-07-27 | 广东省明医医疗慈善基金会 | 基于软镜的立体显示系统及方法 |
CN105812775A (zh) * | 2014-12-29 | 2016-07-27 | 广东省明医医疗慈善基金会 | 基于硬镜的立体显示系统及方法 |
CN105812774A (zh) * | 2014-12-29 | 2016-07-27 | 广东省明医医疗慈善基金会 | 基于插管镜的立体显示系统及方法 |
CN105812772A (zh) * | 2014-12-29 | 2016-07-27 | 深圳超多维光电子有限公司 | 医疗图像立体显示系统及方法 |
CN105809654A (zh) * | 2014-12-29 | 2016-07-27 | 深圳超多维光电子有限公司 | 目标对象跟踪方法、装置和立体显示设备及方法 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030209096A1 (en) * | 2001-01-30 | 2003-11-13 | Z-Kat, Inc. | Tool calibrator and tracker system |
US20050015005A1 (en) * | 2003-04-28 | 2005-01-20 | Kockro Ralf Alfons | Computer enhanced surgical navigation imaging system (camera probe) |
JP2007130398A (ja) * | 2005-11-14 | 2007-05-31 | Toshiba Corp | 光学式位置計測装置 |
US20070183041A1 (en) * | 2006-02-09 | 2007-08-09 | Northern Digital Inc. | Retroreflective marker-tracking systems |
KR20110118640A (ko) * | 2008-12-31 | 2011-10-31 | 인튜어티브 서지컬 오퍼레이션즈 인코포레이티드 | 기기 트래킹을 위한 컨피규레이션 마커 디자인 및 탐지 |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4396945A (en) * | 1981-08-19 | 1983-08-02 | Solid Photography Inc. | Method of sensing the position and orientation of elements in space |
US5923417A (en) * | 1997-09-26 | 1999-07-13 | Northern Digital Incorporated | System for determining the spatial position of a target |
US6061644A (en) * | 1997-12-05 | 2000-05-09 | Northern Digital Incorporated | System for determining the spatial position and orientation of a body |
US6279579B1 (en) | 1998-10-23 | 2001-08-28 | Varian Medical Systems, Inc. | Method and system for positioning patients for medical treatment procedures |
US20110015521A1 (en) * | 2003-03-27 | 2011-01-20 | Boulder Innovation Group, Inc. | Means of Tracking Movement of Bodies During Medical Treatment |
US8211094B2 (en) * | 2004-10-26 | 2012-07-03 | Brainlab Ag | Pre-calibrated reusable instrument |
KR100669250B1 (ko) | 2005-10-31 | 2007-01-16 | 한국전자통신연구원 | 인공표식 기반의 실시간 위치산출 시스템 및 방법 |
EP1872735B1 (de) * | 2006-06-23 | 2016-05-18 | Brainlab AG | Verfahren zum automatischen Identifizieren von Instrumenten bei der medizinischen Navigation |
KR101136743B1 (ko) | 2011-04-27 | 2012-04-19 | 목포대학교산학협력단 | 거리 및 각도측정 기능을 갖는 위치측정장치 |
KR101371384B1 (ko) * | 2013-01-10 | 2014-03-07 | 경북대학교 산학협력단 | 트랙킹 시스템 및 이를 이용한 트랙킹 방법 |
-
2013
- 2013-01-18 KR KR1020130005807A patent/KR101371387B1/ko active IP Right Grant
-
2014
- 2014-01-15 EP EP14740207.7A patent/EP2946741A4/en not_active Withdrawn
- 2014-01-15 WO PCT/KR2014/000426 patent/WO2014112782A1/ko active Application Filing
- 2014-01-15 JP JP2015553651A patent/JP2016515837A/ja active Pending
- 2014-01-15 US US14/372,307 patent/US20160270860A1/en not_active Abandoned
- 2014-01-15 CN CN201480004953.3A patent/CN104936547A/zh active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030209096A1 (en) * | 2001-01-30 | 2003-11-13 | Z-Kat, Inc. | Tool calibrator and tracker system |
US20050015005A1 (en) * | 2003-04-28 | 2005-01-20 | Kockro Ralf Alfons | Computer enhanced surgical navigation imaging system (camera probe) |
JP2007130398A (ja) * | 2005-11-14 | 2007-05-31 | Toshiba Corp | 光学式位置計測装置 |
US20070183041A1 (en) * | 2006-02-09 | 2007-08-09 | Northern Digital Inc. | Retroreflective marker-tracking systems |
KR20110118640A (ko) * | 2008-12-31 | 2011-10-31 | 인튜어티브 서지컬 오퍼레이션즈 인코포레이티드 | 기기 트래킹을 위한 컨피규레이션 마커 디자인 및 탐지 |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105791800A (zh) * | 2014-12-29 | 2016-07-20 | 深圳超多维光电子有限公司 | 立体显示系统及立体显示方法 |
CN105812776A (zh) * | 2014-12-29 | 2016-07-27 | 广东省明医医疗慈善基金会 | 基于软镜的立体显示系统及方法 |
CN105812775A (zh) * | 2014-12-29 | 2016-07-27 | 广东省明医医疗慈善基金会 | 基于硬镜的立体显示系统及方法 |
CN105812774A (zh) * | 2014-12-29 | 2016-07-27 | 广东省明医医疗慈善基金会 | 基于插管镜的立体显示系统及方法 |
CN105812772A (zh) * | 2014-12-29 | 2016-07-27 | 深圳超多维光电子有限公司 | 医疗图像立体显示系统及方法 |
CN105809654A (zh) * | 2014-12-29 | 2016-07-27 | 深圳超多维光电子有限公司 | 目标对象跟踪方法、装置和立体显示设备及方法 |
CN105791800B (zh) * | 2014-12-29 | 2019-09-10 | 深圳超多维科技有限公司 | 立体显示系统及立体显示方法 |
Also Published As
Publication number | Publication date |
---|---|
CN104936547A (zh) | 2015-09-23 |
KR101371387B1 (ko) | 2014-03-10 |
EP2946741A4 (en) | 2016-09-07 |
EP2946741A1 (en) | 2015-11-25 |
US20160270860A1 (en) | 2016-09-22 |
JP2016515837A (ja) | 2016-06-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2014112782A1 (ko) | 트랙킹 시스템 및 이를 이용한 트랙킹 방법 | |
WO2016024797A1 (ko) | 트랙킹 시스템 및 이를 이용한 트랙킹 방법 | |
WO2014109520A1 (ko) | 트랙킹 시스템 및 이를 이용한 트랙킹 방법 | |
US8885177B2 (en) | Medical wide field of view optical tracking system | |
WO2016093455A1 (ko) | 시뮬레이션 시스템에서의 오브젝트 자동 이동 방법 및 이를 적용한 시뮬레이션 시스템 | |
JP5704833B2 (ja) | 操作入力装置およびマニピュレータシステム | |
CN1241260A (zh) | 用于确定目标的空间位置的系统 | |
EP3265009A1 (en) | Redundant reciprocal tracking system | |
WO2018092944A1 (ko) | 경직 평가 장치, 방법 및 시스템 | |
CN106535806A (zh) | 来自多端口视角的手术场景的定量三维成像 | |
US20220175464A1 (en) | Tracker-Based Surgical Navigation | |
WO2014129760A1 (ko) | 트랙킹 시스템 및 이를 이용한 트랙킹 방법 | |
JP2016158911A5 (ko) | ||
US9002074B2 (en) | Facial validation sensor | |
WO2017195984A1 (ko) | 3차원 스캐닝 장치 및 방법 | |
CN207164367U (zh) | Ar眼镜及其追踪系统 | |
US20220011750A1 (en) | Information projection system, controller, and information projection method | |
US20230078919A1 (en) | Extended reality systems for visualizing and controlling operating room equipment | |
WO2020111389A1 (ko) | 사용자의 굴절력 이상 보정을 위한 다층 mla 구조, 디스플레이 패널 및 이미지 처리 방법 | |
WO2022164013A1 (ko) | 구강 스캐너 | |
WO2013162227A1 (ko) | 수술용 내비게이션 시스템 | |
WO2022145595A1 (ko) | 캘리브레이션 시스템 및 방법 | |
EP4354394A2 (en) | Camera tracking system for computer assisted surgery navigation | |
WO2019035546A1 (ko) | 3차원 자기 센서 기반의 손가락 모션 캡쳐 인터페이스 장치 | |
WO2022220383A1 (ko) | 엑스레이 이미지 내에서의 대상 병변의 크기 변화를 측정하는 방법 및 시스템 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 14372307 Country of ref document: US |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14740207 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2015553651 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
REEP | Request for entry into the european phase |
Ref document number: 2014740207 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2014740207 Country of ref document: EP |