CN104970754B - A kind of method controlling endoscope's optical fiber based on Kinect sensor gesture - Google Patents
A kind of method controlling endoscope's optical fiber based on Kinect sensor gesture Download PDFInfo
- Publication number
- CN104970754B CN104970754B CN201510358348.2A CN201510358348A CN104970754B CN 104970754 B CN104970754 B CN 104970754B CN 201510358348 A CN201510358348 A CN 201510358348A CN 104970754 B CN104970754 B CN 104970754B
- Authority
- CN
- China
- Prior art keywords
- endoscope
- optical fiber
- single chip
- chip computer
- kinect sensor
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00002—Operational features of endoscopes
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00064—Constructional details of the endoscope body
- A61B1/00071—Insertion part of the endoscope body
- A61B1/0008—Insertion part of the endoscope body characterised by distal tip features
- A61B1/00096—Optical elements
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00163—Optical arrangements
- A61B1/00165—Optical arrangements with light-conductive means, e.g. fibre optics
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/06—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor with illuminating arrangements
- A61B1/07—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor with illuminating arrangements using light-conductive means, e.g. optical fibres
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
Abstract
A kind of method controlling endoscope's optical fiber based on Kinect sensor gesture, controls end in endoscope and installs Microsoft's Kinect sensor, the change of detection human arm finger position;Endoscope controls end and includes data analysis central processing unit and control single chip computer, and central processing unit is for analyzing the data that Kinect sensor collects, and control single chip computer drives controller action;The data collected are analyzed processing by central processing unit, and by serial ports, the result of process is sent to control single chip computer as control command;After control single chip computer receives order, produce pulse, drive controller, controller action, control endoscope's optical fiber and move.The present invention passes through finger up/down/left/right movements, controls endoscope's optical fiber up/down/left/right movements so that the movement of endoscope's optical fiber becomes more intelligent, improves the accuracy controlling optical fiber.
Description
Technical field
The invention belongs to control technical field, a kind of control the side that endoscope optical fiber moves based on Kinect sensor gesture
Method.
Background technology
Along with the development of technology, modern society increasingly intellectuality and hommization.Gesture control, Voice command, face
The live and work that contactless operation mode is people such as identification brings great convenience.Break away from the bundle of remote controller, mouse etc.
Tiing up, people can realize man-machine interaction by the most comfortable mode.
Endoscope is the important tool of detection interior of articles defect, and operator can observe interior of articles situation by endoscope lens,
If needing to observe interior of articles defect details, needing mobile endoscope lens, endoscope lens is contained on endoscope's optical fiber, mirror
The movement of head is the drive that inner sight glass optical fiber moves.But at present, on market, the movement of endoscope's optical fiber, peeps in mostly passing through
Mirror controls end handle and is controlled, and the clumsy inconvenience of this type of method operation runs operation, leverages its convenience and interactivity.
Summary of the invention
In order to overcome above-mentioned the deficiencies in the prior art, the present invention provides a kind of and controls endoscope's optical fiber based on Kinect sensor gesture
Method.
Technical scheme is as follows:
A kind of method controlling endoscope's optical fiber based on Kinect sensor gesture, comprises the steps of
Step one: control end in endoscope and install the Kinect sensor that Microsoft provides, in real time detection human arm finger position
Change;Endoscope controls end and includes data analysis central processing unit and control single chip computer, and wherein central processing unit is used for analyzing
The data that Kinect sensor collects, control single chip computer drives controller action;
Step 2: endoscope control end central processing unit the data collected are analyzed process, using process result as
Control command is sent to control single chip computer by serial ports;
Step 3: after control single chip computer receives order, produces pulse, drives controller, controller action, controls endoscope's light
Fine mobile.
Kinect sensor of the present invention monitoring human arm finger position change, described position be changed to up/down/to the left
/ move right.
Kinect sensor of the present invention for being analyzed acquisition skeleton model by human body image information.
This method method makes the control of endoscope's optical fiber become more intelligent.Operator is peeped in no longer being carried out by clumsy handle
Mirror optical fiber moves, and only need to stand in the position within two meters, action finger, it is possible to control endoscope's optical fiber before Kinect sensor
Mobile, compared with conventional technology, convenient, quick, directly perceived.
Accompanying drawing explanation
Fig. 1 is control flow chart of the present invention;
Fig. 2 is apparatus of the present invention structural representation.
Detailed description of the invention
In order to make the purpose of the present invention and technical scheme and advantage clearer, below in conjunction with drawings and Examples, to this
Bright it is further elaborated.Should be appreciated that specific embodiment described herein, and need not only in order to explain the present invention
In limiting the present invention.
A kind of method controlling endoscope's optical fiber based on Kinect sensor gesture, comprises the steps of
Step one: control end in endoscope and install the Kinect sensor that Microsoft provides, in real time detection human arm finger position
Change;Endoscope controls end and includes data analysis central processing unit and control single chip computer, and wherein central processing unit is used for analyzing
The data that Kinect sensor collects, control single chip computer drives controller action;
Step 2: endoscope control end central processing unit the data collected are analyzed process, using process result as
Control command is sent to control single chip computer by serial ports;
Step 3: after control single chip computer receives order, produces pulse, drives controller, controller action, controls endoscope's light
Fine mobile.
Kinect sensor of the present invention monitoring human arm finger position change, described position be changed to up/down/to the left
/ move right.Kinect sensor for being analyzed acquisition skeleton model by human body image information.
Fig. 1 show this method schematic flow sheet, and Fig. 2 is apparatus of the present invention structural representation.Control above end 2 in endoscope
The Kinect sensor 1 that Microsoft provides, now, action within human arm finger 3 range sensor 2m, human body hands are installed
Arm finger 3 can upwards, downwards, to the left, paddling to the right.Kinect sensor 1 is analyzed and is obtained the bone of human body image information
Bone model, gathers action data.Endoscope controls the central processing unit in end 2 and is analyzed the data collected processing, will
The result processed is sent to endoscope as control command by serial ports and controls control single chip computer in end 2.Endoscope controls in end 2
Control single chip computer receive order after, produce pulse, drive controller, controller action, controller is peeped in action being delivered to
Mirror optical fiber 5, endoscope's optical fiber drives endoscope lens 4 upwards, downwards, to the left, to the right, thus reaches optimal viewing position.
Claims (1)
1. the method controlling endoscope's optical fiber based on Kinect sensor gesture, it is characterised in that comprise the steps of
Step one: control end in endoscope and install the Kinect sensor that Microsoft provides, detects human arm finger position in real time
Change;Endoscope controls end and includes data analysis central processing unit and control single chip computer, and wherein central processing unit is used for analyzing Kinect
The data that sensor acquisition arrives, control single chip computer drives controller action;
Step 2: endoscope control end central processing unit the data collected are analyzed process, using process result as
Control command is sent to control single chip computer by serial ports;
Step 3: after control single chip computer receives order, produces pulse, drives controller, controller action, controls endoscope's light
Fine mobile;
Kinect sensor monitoring human arm finger position change, described position be changed to up/down/left/right
Motion;Kinect sensor for being analyzed acquisition skeleton model by human body image information.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510358348.2A CN104970754B (en) | 2015-06-25 | 2015-06-25 | A kind of method controlling endoscope's optical fiber based on Kinect sensor gesture |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510358348.2A CN104970754B (en) | 2015-06-25 | 2015-06-25 | A kind of method controlling endoscope's optical fiber based on Kinect sensor gesture |
Publications (2)
Publication Number | Publication Date |
---|---|
CN104970754A CN104970754A (en) | 2015-10-14 |
CN104970754B true CN104970754B (en) | 2016-09-28 |
Family
ID=54268129
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510358348.2A Active CN104970754B (en) | 2015-06-25 | 2015-06-25 | A kind of method controlling endoscope's optical fiber based on Kinect sensor gesture |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN104970754B (en) |
Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2011060185A1 (en) * | 2009-11-13 | 2011-05-19 | Intuitive Surgical Operations, Inc. | Method and system for hand presence detection in a minimally invasive surgical system |
CN102647955A (en) * | 2009-11-13 | 2012-08-22 | 直观外科手术操作公司 | Method and apparatus for hand gesture control in a minimally invasive surgical system |
CN202512439U (en) * | 2012-02-28 | 2012-10-31 | 陶重犇 | Human-robot cooperation system with webcam and wearable sensor |
CN103491848A (en) * | 2011-12-26 | 2014-01-01 | 奥林巴斯医疗株式会社 | Medical endoscope system |
CN103729057A (en) * | 2013-12-18 | 2014-04-16 | 京东方科技集团股份有限公司 | Method and system for controlling postures of display device by using gestures |
CN103903011A (en) * | 2014-04-02 | 2014-07-02 | 重庆邮电大学 | Intelligent wheelchair gesture recognition control method based on image depth information |
CN104108097A (en) * | 2014-06-25 | 2014-10-22 | 陕西高华知本化工科技有限公司 | Feeding and discharging mechanical arm system based on gesture control |
CN203973551U (en) * | 2014-06-13 | 2014-12-03 | 济南翼菲自动化科技有限公司 | A kind of remote control robot of controlling by body gesture |
CN104284637A (en) * | 2012-05-18 | 2015-01-14 | 奥林巴斯株式会社 | Medical operation assistance device |
CN104335139A (en) * | 2012-05-10 | 2015-02-04 | 皇家飞利浦有限公司 | Gesture control |
CN104375504A (en) * | 2014-09-12 | 2015-02-25 | 中山大学 | Running accompanying robot and tracking control strategy and movement control method for running accompanying robot |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8880223B2 (en) * | 2012-07-16 | 2014-11-04 | Florida Institute for Human & Maching Cognition | Anthro-centric multisensory interface for sensory augmentation of telesurgery |
-
2015
- 2015-06-25 CN CN201510358348.2A patent/CN104970754B/en active Active
Patent Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2011060185A1 (en) * | 2009-11-13 | 2011-05-19 | Intuitive Surgical Operations, Inc. | Method and system for hand presence detection in a minimally invasive surgical system |
CN102647955A (en) * | 2009-11-13 | 2012-08-22 | 直观外科手术操作公司 | Method and apparatus for hand gesture control in a minimally invasive surgical system |
CN103491848A (en) * | 2011-12-26 | 2014-01-01 | 奥林巴斯医疗株式会社 | Medical endoscope system |
CN202512439U (en) * | 2012-02-28 | 2012-10-31 | 陶重犇 | Human-robot cooperation system with webcam and wearable sensor |
CN104335139A (en) * | 2012-05-10 | 2015-02-04 | 皇家飞利浦有限公司 | Gesture control |
CN104284637A (en) * | 2012-05-18 | 2015-01-14 | 奥林巴斯株式会社 | Medical operation assistance device |
CN103729057A (en) * | 2013-12-18 | 2014-04-16 | 京东方科技集团股份有限公司 | Method and system for controlling postures of display device by using gestures |
CN103903011A (en) * | 2014-04-02 | 2014-07-02 | 重庆邮电大学 | Intelligent wheelchair gesture recognition control method based on image depth information |
CN203973551U (en) * | 2014-06-13 | 2014-12-03 | 济南翼菲自动化科技有限公司 | A kind of remote control robot of controlling by body gesture |
CN104108097A (en) * | 2014-06-25 | 2014-10-22 | 陕西高华知本化工科技有限公司 | Feeding and discharging mechanical arm system based on gesture control |
CN104375504A (en) * | 2014-09-12 | 2015-02-25 | 中山大学 | Running accompanying robot and tracking control strategy and movement control method for running accompanying robot |
Also Published As
Publication number | Publication date |
---|---|
CN104970754A (en) | 2015-10-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR101977332B1 (en) | Table top system for intuitive guidance in augmented reality remote video communication environment | |
CN103529929B (en) | Gesture recognition system and glasses capable of recognizing gesture actions | |
CN205067206U (en) | Vickers | |
CN106426229B (en) | Virtual reality monitor control system and control method for inspection robot | |
CN103336581A (en) | Human eye movement characteristic design-based human-computer interaction method and system | |
JP2017530447A (en) | System and method for inputting a gesture in a 3D scene | |
US20140146360A1 (en) | Gesture control for printing presses | |
TW200723070A (en) | Optical navigation device and method of operating the same | |
CN103941864A (en) | Somatosensory controller based on human eye binocular visual angle | |
KR20170102991A (en) | Control systems and control methods | |
CN105500378B (en) | It is automatically performed the robot of game rhythm great master | |
CN103914130A (en) | Display device and method for adjusting observation distance of display device | |
EP2693323A3 (en) | Method and apparatus for virtual tour creation in mobile device | |
CN108305553A (en) | A kind of experiment device for teaching of view-based access control model assembly | |
CN101340562A (en) | Monitoring system and method | |
Zhang et al. | Robotic control of dynamic and static gesture recognition | |
CN110045904A (en) | Man-machine interactive system, method and the vehicle including the system | |
CN105242787A (en) | Display device and adjustment method thereof | |
CN104970754B (en) | A kind of method controlling endoscope's optical fiber based on Kinect sensor gesture | |
CN110825216A (en) | Method and system for man-machine interaction of driver during driving | |
CN105677040A (en) | Terminal control method, device and wearable device | |
CN104815804B (en) | Tank body tank switching detection means and detection method | |
CN104598138A (en) | Method and device for controlling electronic map | |
CN203482300U (en) | Image recognition device | |
WO2005015377A3 (en) | Method and system for gestural control of an apparatus |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant |