CN109064817A - Surgery simulation system based on CT Three-dimension Reconstruction Model - Google Patents
Surgery simulation system based on CT Three-dimension Reconstruction Model Download PDFInfo
- Publication number
- CN109064817A CN109064817A CN201810790286.6A CN201810790286A CN109064817A CN 109064817 A CN109064817 A CN 109064817A CN 201810790286 A CN201810790286 A CN 201810790286A CN 109064817 A CN109064817 A CN 109064817A
- Authority
- CN
- China
- Prior art keywords
- scene
- sub
- tracker
- level
- scalpel
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B9/00—Simulators for teaching or training purposes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Business, Economics & Management (AREA)
- Educational Administration (AREA)
- Educational Technology (AREA)
- Computer Graphics (AREA)
- Geometry (AREA)
- Software Systems (AREA)
- Processing Or Creating Images (AREA)
Abstract
Surgery simulation system provided by the invention based on CT Three-dimension Reconstruction Model, including scalpel, positioning gloves, VR glasses, scene database, selecting unit and analogue unit;Tracker is equipped on scalpel and positioning gloves;Scene database is stored with preset simulated scenario;Selective listing is equipped in selecting unit;Selective listing includes first category and second category;Selecting unit reads the location information of tracker, obtains the gesture motion of user according to the location information of tracker on positioning gloves, selects simulated scenario;Analogue unit is used for after entering simulated scenario, is started to carry out surgical simulation, is read the location information of tracker on positioning gloves and scalpel, switched over according to the location information of tracker to simulated scenario neutron scene;VR glasses are worn for user, the system, enable the physician to carry out surgical simulation, improve the experience sense that doctor comes personally.
Description
Technical field
The invention belongs to field of computer technology, and in particular to the surgery simulation system based on CT Three-dimension Reconstruction Model.
Background technique
The surgical procedure technical ability needs of culture surgeon are trained using a large amount of animals or human body entity.Animal is real
Body one side source is rare, on the other hand also can not true representations inside of human body tissue construction, carry out operation instruction using human body
Practice, in addition to the above-mentioned limited source the problem of other than, more there is ethics and jural limitation.Virtual reality (VR) is applied to
Operative training, solution is not only effective, but also has a extensive future, and trained surgeon is in the hand for receiving to be based on virtual reality (VR)
When art skills training, feeling of immersion on the spot in person is obtained by experiencing in conjunction with of both from vision and operation, but in mesh
The technical level of the surgery training system of preceding virtual reality (VR), can only be conceived to the visual experience of user, not to use
The operating experience of person, for trainee, operating experience not only affects its Cognitive Effects to operation method, more influences
Its gimmick intuition to operation method, therefore be directly related to trained effect, make the inefficient of operation.
Since body physiological institutional framework and the pathological anatomy construction of each individual are there are a degree of difference, in order to
The success rate of operation is improved, especially for complicated operation case, surgical field is high to the cry of preoperative simulation.
Summary of the invention
For the defects in the prior art, the present invention provides a kind of surgery simulation system based on CT Three-dimension Reconstruction Model,
It enables the physician to carry out surgical simulation, improves the experience sense that doctor comes personally.
In a first aspect, a kind of surgery simulation system based on CT Three-dimension Reconstruction Model, including image database, three-dimensional reconstruction
Unit and generation unit;
The image database is stored with CT images data;
The three-dimensional reconstruction unit carries out three-dimensional reconstruction for reading CT images data, obtains CT image;
The generation unit is used to generate simulated scenario according to CT image.
Second aspect, a kind of surgery simulation system based on CT Three-dimension Reconstruction Model, including scalpel, positioning gloves, VR
Glasses, scene database, selecting unit and analogue unit;
The scalpel includes blade and knife handle, and the lower end surface of blade forms blade, on the knife handle of the scalpel, blade
In the middle part of front end, blade and blade rear end is equipped with tracker;
The positioning gloves include the back of the hand portion and multiple fingerstall portions being arranged in the back of the hand portion for wrapping up the back of the hand, described every
Tracker is equipped in a fingerstall portion;
The scene database is stored with preset simulated scenario, and each simulated scenario is named using operation names, each
Simulated scenario includes multiple sub-scenes;
Selective listing is equipped in the selecting unit;Selective listing includes being classified according to human organ, with human body device
The first category of official's name, and be associated with first category, classified according to operation names, with the second of operation names name
Classification;
Selecting unit is also used to read the location information of tracker on positioning gloves, according to the position of tracker on positioning gloves
Confidence ceases to obtain the gesture motion of user;Selecting unit is also used to receive the selection instruction of user, into preset selection scene,
Scene is selected to show that first category in selective listing, cursor are located at the first row in selective listing;When the gesture motion recognized is
When upper sliding, cursor current location in selective listing moves up a row;When the gesture motion recognized is to glide, cursor is being selected
Current location line down in list;When the gesture motion recognized is left sliding, first category in selective listing, and light are shown
Mark the first row in selective listing;When the gesture motion recognized is right sliding, indicated by display and cursor current location
The associated second category of first category, and cursor is located at the first row in selective listing;When the gesture motion recognized is fist,
It reads and enters and the associated simulated scenario of second category indicated by cursor current location;
The analogue unit is used for after entering simulated scenario, is started to carry out surgical simulation, is read positioning gloves and hand
The location information of tracker on art knife switches over simulated scenario neutron scene according to the location information of tracker;
The VR glasses are worn for user, simulated scenario in viewing selecting unit selection scene and analogue unit.
Further, the sub-scene of the simulated scenario includes level-one sub-scene and multiple second level sub-scenes;The level-one
Sub-scene includes the body image in range of operation;The second level sub-scene includes the human body shadow after human body different parts are spaced
Picture;Second level sub-scene is equipped with trigger position.
Further, the location information according to tracker switches over simulated scenario neutron scene and specifically includes:
It is identified to obtain shift position of the scalpel in level-one sub-scene according to the location information of tracker;
When shift position is located in the trigger position of second level sub-scene, the second level sub-scene is switched to.
Further, the analogue unit, which is also used to work as, detects shift position and institute of the scalpel in level-one sub-scene
When thering is the trigger position of second level sub-scene to mismatch, error prompting information is shown.
Further, which further includes scoring unit;
Scoring unit is for scoring to the surgical simulation process of user.
Further, the scoring unit is also used to generate simulation report according to the surgical simulation process of user;Simulation report
It include that the operation diagram that screenshot obtains is carried out to current sub-scene when detecting that scalpel moves in level-one sub-scene in announcement
Piece.
Further, the location information according to tracker identifies to obtain mobile position of the scalpel in level-one sub-scene
It sets and specifically includes:
Effective moving area is equipped in level-one sub-scene;
When identifying that scalpel moves in effective moving area of level-one sub-scene according to the location information of tracker, obtain
Shift position of the scalpel in level-one sub-scene.
Further, the tracker is HTC VIVE Tracker tracker.
As shown from the above technical solution, the surgery simulation system provided by the invention based on CT Three-dimension Reconstruction Model, operation
Tracker on knife can obtain the location information and angle information of scalpel, determine the shift position of scalpel;Position gloves
On tracker can obtain the location information and angle information in each fingerstall portion, determine gesture and the position of user, realize hand
The positioning of art knife and user hand.The system combines VR technology, positions to the true hand of true scalpel, so that user puts on
VR glasses can accurately simulate operation while watching scene, improve the feeling of immersion of user, make user using when reach body and face
The effect in its border, doctor are able to carry out surgical simulation, improve the experience sense that doctor comes personally.
Detailed description of the invention
It, below will be to specific in order to illustrate more clearly of the specific embodiment of the invention or technical solution in the prior art
Embodiment or attached drawing needed to be used in the description of the prior art are briefly described.In all the appended drawings, similar element
Or part is generally identified by similar appended drawing reference.In attached drawing, each element or part might not be drawn according to actual ratio.
Fig. 1 is the module frame chart for the system that embodiment two provides.
Specific embodiment
It is described in detail below in conjunction with embodiment of the attached drawing to technical solution of the present invention.Following embodiment is only used for
Clearly illustrate technical solution of the present invention, therefore be only used as example, and cannot be used as a limitation and limit protection model of the invention
It encloses.It should be noted that unless otherwise indicated, technical term or scientific term used in this application are should be belonging to the present invention
The ordinary meaning that field technical staff is understood.
Embodiment one:
A kind of surgery simulation system based on CT Three-dimension Reconstruction Model, including image database, three-dimensional reconstruction unit and life
At unit;
The image database is stored with CT images data;
The three-dimensional reconstruction unit carries out three-dimensional reconstruction for reading CT images data, obtains CT image;
The generation unit is used to generate simulated scenario according to CT image.
Embodiment two:
A kind of surgery simulation system based on CT Three-dimension Reconstruction Model, referring to Fig. 1, including scalpel, positioning gloves, VR
Mirror, scene database, selecting unit and analogue unit;
The scalpel includes blade and knife handle, and the lower end surface of blade forms blade, on the knife handle of the scalpel, blade
In the middle part of front end, blade and blade rear end is equipped with tracker;
Specifically, the tracker is HTC VIVE Tracker tracker.3 tracker energy on scalpel are set
It is enough accurately located the blade of scalpel, then by location information synchronizing in the scalpel into virtual environment, realizes that scalpel exists
It is true and virtually in it is synchronous.
The positioning gloves include the back of the hand portion and multiple fingerstall portions being arranged in the back of the hand portion for wrapping up the back of the hand, described every
Tracker is equipped in a fingerstall portion;
Specifically, the tracker in each fingerstall portion on positioning gloves is set, the position of user's finger can be accurately located
Set, then by location information synchronizing in the hand into virtual environment, realize hand it is true and virtually in it is synchronous.
The scene database is stored with preset simulated scenario, and each simulated scenario is named using operation names, each
Simulated scenario includes multiple sub-scenes;
Specifically, each simulated scenario is constructed according to the factors such as the position of the organ of human body, size, arrangement in surgical procedure
It is formed.In surgical procedure, sub-scene is generated according to different body image, such as: it is constructed according to the image after A organ is separated
Sub-scene A constructs sub-scene B according to the image after B organ is separated.
Selective listing is equipped in the selecting unit;Selective listing includes being classified according to human organ, with human body device
The first category of official's name, and be associated with first category, classified according to operation names, with the second of operation names name
Classification;
Specifically, first category is named with human organ, such as heart, liver, stomach etc..Second category is operation names point
Class, such as operation relevant to heart have bypass surgery, congenital heart disease operation, valve replacement surgery etc..Then with
The associated second category of this first category of heart includes bypass surgery, congenital heart disease operation, valve replacement surgery etc.
Deng.
Selecting unit is also used to read the location information of tracker on positioning gloves, according to the position of tracker on positioning gloves
Confidence ceases to obtain the gesture motion of user;
Specifically, the movement that each finger is positioned according to the position of tracker on positioning gloves, obtains gesture motion.
Selecting unit is also used to receive the selection instruction of user, into preset selection scene, selects scene display selection
First category in list, cursor are located at the first row in selective listing;When the gesture motion recognized is upper sliding, cursor is being selected
Current location moves up a row in list;When the gesture motion recognized is to glide, cursor is in selective listing under current location
Move a line;When the gesture motion recognized is left sliding, first category in selective listing is shown, and cursor is located in selective listing
The first row;When the gesture motion recognized is right sliding, display and first category indicated by cursor current location associated the
Two classifications, and cursor is located at the first row in selective listing;When the gesture motion recognized is fist, read and entrance and cursor
The associated simulated scenario of second category indicated by current location;
Specifically, selection instruction, which is used to indicate, starts to select simulated scenario, and when starting selection, first which organ selection is
Operation, which first category belonged to.Specific operation, i.e. second category are selected under the first category again.Pass through two layers
Selective listing can help user quickly to choose simulated scenario.When sliding on user, cursor is moved to lastrow.Work as user
When downslide, cursor is moved to next line.When user is left sliding, returns to first category and selected.When user is right sliding, into light
The associated second category of current first category is marked to be selected.When the gesture motion for recognizing user is fist, into cursor
The simulated scenario of the second category currently indicated.
The analogue unit is used for after entering simulated scenario, is started to carry out surgical simulation, is read positioning gloves and hand
The location information of tracker on art knife switches over simulated scenario neutron scene according to the location information of tracker;
Specifically, such as after detecting that user separates A organ in virtual world, it is switched to sub-scene A.If separated
B organ is switched to sub-scene B.
The VR glasses are worn for user, simulated scenario in viewing selecting unit selection scene and analogue unit.
The surgery simulation system in use, user takes VR glasses, due to tracker to the hand of scalpel and user into
Row positioning, by scalpel and the hand of user in true environment scalpel and hand of the location information synchronizing into virtual environment
In, realize scalpel and the hand of user it is true and virtually in it is synchronous.In this way, after user takes VR glasses, it can be in void
Position or the movement of the hand of virtual scalpel and user are seen in near-ring border.In this way, user can execute true operation
The movement of scalpel, while it can also be seen that sub-scene when operation knife in virtual environment.The system combines VR skill
Art positions the true hand of true scalpel, allows user to put on VR glasses and accurately simulates while watching scene
Operation, improves the feeling of immersion of user, make user using when achieve the effect that on the spot in person, doctor is able to carry out surgical simulation, mentions
The experience sense that high doctor comes personally.
Embodiment three:
Embodiment three on the basis of example 2, increases the following contents:
The sub-scene of the simulated scenario includes level-one sub-scene and multiple second level sub-scenes;The level-one sub-scene includes
Body image in range of operation;The second level sub-scene includes the body image after human body different parts are spaced;Second level
Scene is equipped with trigger position.
Specifically, the instruction of level-one sub-scene does not start body image when performing the operation also, it can be seen that the complete shape of human body.Touching
Hair position is for triggering the switching of second level sub-scene.Such as according to separating the building sub-scene A of the image after A organ, according to separating B device
Image after official constructs sub-scene B.
After building sub-scene, the location information according to tracker switches over tool to simulated scenario neutron scene
Body includes:
It is identified to obtain shift position of the scalpel in level-one sub-scene according to the location information of tracker;
When shift position is located in the trigger position of second level sub-scene, the second level sub-scene is switched to.
Specifically, shift position of the scalpel in level-one sub-scene is for indicating simulated environment menisectomy knife in human body
Cutting position illustrate that the cutting position of scalpel just cuts to accurate position when shift position is located in trigger position
It sets, switches to the sub-scene after the position is spaced at this time.
The present embodiment also provides error prompting function.The analogue unit, which is also used to work as, detects scalpel in level-one subfield
When the trigger position of shift position and all second level sub-scenes in scape mismatches, error prompting information is shown.
Specifically, if the trigger position of the shift position of simulated environment menisectomy knife and all second level sub-scenes not
Together, then it is assumed that the cutting position of user is wrong, progress error prompting.
Further, which further includes scoring unit;
Scoring unit is for scoring to the surgical simulation process of user.
Specifically, if mistake is fewer during the surgical simulation of user, score is higher.Use is assessed by scoring score
The surgical simulation process at family.Score range can be 0-10 points.
Further, the scoring unit is also used to generate simulation report according to the surgical simulation process of user;Simulation report
It include that the operation diagram that screenshot obtains is carried out to current sub-scene when detecting that scalpel moves in level-one sub-scene in announcement
Piece.
After system provided in this embodiment is in addition to scoring to operation simulation process, simulation report is also generated.Simulation report
Include screenshot when scalpel moves every time in level-one sub-scene in announcement, user is facilitated to look back hand during entire surgical simulation
The shift position of art knife, summarizes to faulty operation.
Further, the location information according to tracker identifies to obtain mobile position of the scalpel in level-one sub-scene
It sets and specifically includes:
Effective moving area is equipped in level-one sub-scene;
When identifying that scalpel moves in effective moving area of level-one sub-scene according to the location information of tracker, obtain
Shift position of the scalpel in level-one sub-scene.
Specifically, effective moving area is specifically limited according to different operations.In order to avoid by all scalpels in level-one
The mobile misjudgement of sub-scene is effective movement, and each level-one sub-scene is equipped with effective moving area, and only scalpel is one
When moving in effective moving area of grade sub-scene, it is just considered effective movement, just identifies scalpel in level-one sub-scene
Shift position.
System provided by the embodiment of the present invention, to briefly describe, embodiment part does not refer to place, can refer to embodiment
Corresponding contents in two.
Finally, it should be noted that the above embodiments are only used to illustrate the technical solution of the present invention., rather than its limitations;To the greatest extent
Pipe present invention has been described in detail with reference to the aforementioned embodiments, those skilled in the art should understand that: its according to
So be possible to modify the technical solutions described in the foregoing embodiments, or to some or all of the technical features into
Row equivalent replacement;And these are modified or replaceed, various embodiments of the present invention technology that it does not separate the essence of the corresponding technical solution
The range of scheme should all cover within the scope of the claims and the description of the invention.
Claims (9)
1. a kind of surgery simulation system based on CT Three-dimension Reconstruction Model, which is characterized in that including image database, three-dimensional reconstruction
Unit and generation unit;
The image database is stored with CT images data;
The three-dimensional reconstruction unit carries out three-dimensional reconstruction for reading CT images data, obtains CT image;
The generation unit is used to generate simulated scenario according to CT image.
2. a kind of surgery simulation system based on CT Three-dimension Reconstruction Model, which is characterized in that including scalpel, positioning gloves, VR
Glasses, scene database, selecting unit and analogue unit;
The scalpel includes blade and knife handle, and the lower end surface of blade forms blade, on the knife handle of the scalpel, before blade
In the middle part of end, blade and blade rear end is equipped with tracker;
The positioning gloves include the back of the hand portion and multiple fingerstall portions being arranged in the back of the hand portion for wrapping up the back of the hand, each finger
Set is equipped with tracker in portion;
The scene database is stored with preset simulated scenario, and each simulated scenario is named using operation names, each simulation
Scene includes multiple sub-scenes;
Selective listing is equipped in the selecting unit;Selective listing includes being classified according to human organ, being ordered with human organ
The first category of name, and the second class for being associated with, being classified according to operation names, being named with operation names with first category
Not;
Selecting unit is also used to read the location information of tracker on positioning gloves, is believed according to the position of tracker on positioning gloves
Breath obtains the gesture motion of user;Selecting unit is also used to receive the selection instruction of user, into preset selection scene, selection
Scene shows that first category in selective listing, cursor are located at the first row in selective listing;When the gesture motion recognized is upper cunning
When, cursor current location in selective listing moves up a row;When the gesture motion recognized is to glide, cursor is in selective listing
Middle current location line down;When the gesture motion recognized is left sliding, first category in selective listing, and cursor position are shown
The first row in selective listing;When the gesture motion recognized is right sliding, first indicated by display and cursor current location
The second category of category associations, and cursor is located at the first row in selective listing;When the gesture motion recognized is fist, read
And enter and the associated simulated scenario of second category indicated by cursor current location;
The analogue unit is used for after entering simulated scenario, is started to carry out surgical simulation, is read positioning gloves and scalpel
The location information of upper tracker switches over simulated scenario neutron scene according to the location information of tracker;
The VR glasses are worn for user, simulated scenario in viewing selecting unit selection scene and analogue unit.
3. the surgery simulation system according to claim 2 based on CT Three-dimension Reconstruction Model, which is characterized in that the simulation yard
The sub-scene of scape includes level-one sub-scene and multiple second level sub-scenes;The level-one sub-scene includes the human body shadow in range of operation
Picture;The second level sub-scene includes the body image after human body different parts are spaced;Second level sub-scene is equipped with trigger position.
4. the surgery simulation system according to claim 3 based on CT Three-dimension Reconstruction Model, which is characterized in that the basis chases after
The location information of track device switches over simulated scenario neutron scene and specifically includes:
It is identified to obtain shift position of the scalpel in level-one sub-scene according to the location information of tracker;
When shift position is located in the trigger position of second level sub-scene, the second level sub-scene is switched to.
5. the surgery simulation system according to claim 4 based on CT Three-dimension Reconstruction Model, which is characterized in that the simulation is single
Member is also used to when not detecting the trigger positions of shift position of the scalpel in level-one sub-scene and all second level sub-scenes not
Timing shows error prompting information.
6. the surgery simulation system according to claim 4 based on CT Three-dimension Reconstruction Model, which is characterized in that the system is also wrapped
Include scoring unit;
Scoring unit is for scoring to the surgical simulation process of user.
7. the surgery simulation system according to claim 6 based on CT Three-dimension Reconstruction Model, which is characterized in that the judge paper
Member is also used to generate simulation report according to the surgical simulation process of user;It include working as to detect scalpel in level-one in simulation report
When moving in sub-scene, the operation picture that screenshot obtains is carried out to current sub-scene.
8. the surgery simulation system according to claim 4 based on CT Three-dimension Reconstruction Model, which is characterized in that the basis chases after
The location information of track device, which identifies to obtain shift position of the scalpel in level-one sub-scene, to be specifically included:
Effective moving area is equipped in level-one sub-scene;
When identifying that scalpel moves in effective moving area of level-one sub-scene according to the location information of tracker, performed the operation
Shift position of the knife in level-one sub-scene.
9. the surgery simulation system according to claim 2 based on CT Three-dimension Reconstruction Model, which is characterized in that the tracker
For HTC VIVE Tracker tracker.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810790286.6A CN109064817A (en) | 2018-07-18 | 2018-07-18 | Surgery simulation system based on CT Three-dimension Reconstruction Model |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810790286.6A CN109064817A (en) | 2018-07-18 | 2018-07-18 | Surgery simulation system based on CT Three-dimension Reconstruction Model |
Publications (1)
Publication Number | Publication Date |
---|---|
CN109064817A true CN109064817A (en) | 2018-12-21 |
Family
ID=64817191
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810790286.6A Pending CN109064817A (en) | 2018-07-18 | 2018-07-18 | Surgery simulation system based on CT Three-dimension Reconstruction Model |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109064817A (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110349149A (en) * | 2019-07-12 | 2019-10-18 | 广东省人民医院(广东省医学科学院) | Congenital heart disease categorizing system and method based on deep neural network and form similarity |
CN112509410A (en) * | 2020-12-08 | 2021-03-16 | 中日友好医院(中日友好临床医学研究所) | Virtual reality-based auxiliary teaching system for hip arthroscopy operation |
CN115273583A (en) * | 2022-05-16 | 2022-11-01 | 华中科技大学同济医学院附属协和医院 | Multi-person interactive orthopedics clinical teaching method based on mixed reality |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8786613B2 (en) * | 2006-04-08 | 2014-07-22 | Alan Millman | Method and system for interactive simulation of materials and models |
CN104463965A (en) * | 2014-12-17 | 2015-03-25 | 中国科学院自动化研究所 | Training scene simulation system and method for minimally invasive cardiovascular interventional operation |
CN105632310A (en) * | 2016-01-25 | 2016-06-01 | 新乡医学院 | Human anatomy teaching system |
CN106652624A (en) * | 2016-10-12 | 2017-05-10 | 大连文森特软件科技有限公司 | Medical surgery simulation system based on VR technology and transient noise removal technology |
CN106708260A (en) * | 2016-11-30 | 2017-05-24 | 宇龙计算机通信科技(深圳)有限公司 | Generation method and device for virtual reality surgery scene |
CN107067856A (en) * | 2016-12-31 | 2017-08-18 | 歌尔科技有限公司 | A kind of medical simulation training system and method |
CN107066180A (en) * | 2017-03-13 | 2017-08-18 | 深圳国泰安教育技术股份有限公司 | A kind of task creating method and its system based on VR operations |
CN107168530A (en) * | 2017-04-26 | 2017-09-15 | 腾讯科技(深圳)有限公司 | Object processing method and device in virtual scene |
CN206594626U (en) * | 2017-03-20 | 2017-10-27 | 李佳亮 | The wearable gloves of VR |
-
2018
- 2018-07-18 CN CN201810790286.6A patent/CN109064817A/en active Pending
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8786613B2 (en) * | 2006-04-08 | 2014-07-22 | Alan Millman | Method and system for interactive simulation of materials and models |
CN104463965A (en) * | 2014-12-17 | 2015-03-25 | 中国科学院自动化研究所 | Training scene simulation system and method for minimally invasive cardiovascular interventional operation |
CN105632310A (en) * | 2016-01-25 | 2016-06-01 | 新乡医学院 | Human anatomy teaching system |
CN106652624A (en) * | 2016-10-12 | 2017-05-10 | 大连文森特软件科技有限公司 | Medical surgery simulation system based on VR technology and transient noise removal technology |
CN106708260A (en) * | 2016-11-30 | 2017-05-24 | 宇龙计算机通信科技(深圳)有限公司 | Generation method and device for virtual reality surgery scene |
CN107067856A (en) * | 2016-12-31 | 2017-08-18 | 歌尔科技有限公司 | A kind of medical simulation training system and method |
CN107066180A (en) * | 2017-03-13 | 2017-08-18 | 深圳国泰安教育技术股份有限公司 | A kind of task creating method and its system based on VR operations |
CN206594626U (en) * | 2017-03-20 | 2017-10-27 | 李佳亮 | The wearable gloves of VR |
CN107168530A (en) * | 2017-04-26 | 2017-09-15 | 腾讯科技(深圳)有限公司 | Object processing method and device in virtual scene |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110349149A (en) * | 2019-07-12 | 2019-10-18 | 广东省人民医院(广东省医学科学院) | Congenital heart disease categorizing system and method based on deep neural network and form similarity |
CN110349149B (en) * | 2019-07-12 | 2022-06-24 | 广东省人民医院(广东省医学科学院) | Advanced heart disease classification system and method based on deep neural network and morphological similarity |
CN112509410A (en) * | 2020-12-08 | 2021-03-16 | 中日友好医院(中日友好临床医学研究所) | Virtual reality-based auxiliary teaching system for hip arthroscopy operation |
CN115273583A (en) * | 2022-05-16 | 2022-11-01 | 华中科技大学同济医学院附属协和医院 | Multi-person interactive orthopedics clinical teaching method based on mixed reality |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
AU2019289084B2 (en) | Virtual guidance for orthopedic surgical procedures | |
CN110390851B (en) | Augmented reality training system | |
KR102477325B1 (en) | Multi-panel graphical user interface for robotic surgical systems | |
CN107067856B (en) | Medical simulation training system and method | |
CN106251751B (en) | A kind of simulated medical surgery analogue system based on VR technology | |
Birlo et al. | Utility of optical see-through head mounted displays in augmented reality-assisted surgery: A systematic review | |
Sewell et al. | Providing metrics and performance feedback in a surgical simulator | |
US20200020171A1 (en) | Systems and methods for mixed reality medical training | |
US20190183591A1 (en) | Multi-panel graphical user interface for a robotic surgical system | |
JP2016524262A (en) | 3D user interface | |
CN1973780B (en) | System and method for facilitating surgical | |
CN109064817A (en) | Surgery simulation system based on CT Three-dimension Reconstruction Model | |
Horn et al. | Visual search and coordination changes in response to video and point-light demonstrations without KR | |
CN107847289A (en) | The morphology operation of reality enhancing | |
Lin et al. | A first-person mentee second-person mentor AR interface for surgical telementoring | |
CN104883975A (en) | Real-time scene-modeling combining 3d ultrasound and 2d x-ray imagery | |
JP2015506726A (en) | Universal microsurgery simulator | |
CN106980383A (en) | A kind of dummy model methods of exhibiting, module and the virtual human body anatomical model display systems based on the module | |
Long et al. | Integrating artificial intelligence and augmented reality in robotic surgery: An initial dvrk study using a surgical education scenario | |
CN108961907A (en) | Virtual micro- ophthalmologic operation training method and system | |
CN104978872A (en) | Surgery demonstration method, surgery demonstration device and surgery demonstration system | |
CN108922300A (en) | Surgical simulation 3D system based on digitized humans | |
Xu et al. | Doctoral colloquium—a snapshot of the future: Virtual and augmented reality training for radiology | |
Ricca et al. | Comparing touch-based and head-tracking navigation techniques in a virtual reality biopsy simulator | |
Coles | Investigating augmented reality visio-haptic techniques for medical training |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20181221 |
|
RJ01 | Rejection of invention patent application after publication |