CN110246216A - Spine model generation method, spine model generate system and terminal - Google Patents

Spine model generation method, spine model generate system and terminal Download PDF

Info

Publication number
CN110246216A
CN110246216A CN201910450859.5A CN201910450859A CN110246216A CN 110246216 A CN110246216 A CN 110246216A CN 201910450859 A CN201910450859 A CN 201910450859A CN 110246216 A CN110246216 A CN 110246216A
Authority
CN
China
Prior art keywords
image
spine
lesion
model
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910450859.5A
Other languages
Chinese (zh)
Other versions
CN110246216B (en
Inventor
苗燕茹
孙宇
胡颖
李世博
谭志强
齐晓志
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Institute of Advanced Technology of CAS
Original Assignee
Shenzhen Institute of Advanced Technology of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Institute of Advanced Technology of CAS filed Critical Shenzhen Institute of Advanced Technology of CAS
Priority to CN201910450859.5A priority Critical patent/CN110246216B/en
Publication of CN110246216A publication Critical patent/CN110246216A/en
Application granted granted Critical
Publication of CN110246216B publication Critical patent/CN110246216B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10081Computed x-ray tomography [CT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30008Bone
    • G06T2207/30012Spine; Backbone

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computer Graphics (AREA)
  • Geometry (AREA)
  • Software Systems (AREA)
  • Apparatus For Radiation Diagnosis (AREA)
  • Image Analysis (AREA)

Abstract

The application is suitable for engineering in medicine technical field, provides a kind of spine model generation method, spine model generates system and terminal, and wherein method includes: according to non-lesion spine image, and training generates confrontation network model;Network model is fought by the generation after training, image repair is carried out to lesion site in lesion spine image, obtains targets spine image, wherein lesion site described in the targets spine image is repaired as non-lesion spine form;Three-dimensional reconstruction is carried out to the targets spine image, obtains corresponding targets spine model, promotes the precision of spinal prostheses production, the activity limitation after reducing Using prosthesis patient's body.

Description

Spine model generation method, spine model generate system and terminal
Technical field
The application belongs to engineering in medicine technical field more particularly to a kind of spine model generation method, spine model generate System and terminal.
Background technique
Full backbone resection (TES) refers to complete resection one or more spinal segments, including centrum, interverbebral disc, vertebral arch Root, vertebral plate, spinous process, ligament and the complete resection of operation together of the attached soft tissue of other backbones.The operation be at present in the world The minimum modus operandi of generally acknowledged treatment tumor of spine recurrence rate.
But with the completion of operation, the backbone of human body, which is influenced continuity by operation, to be interrupted.To realize spinal stability Reconstruction, the method currently mainly used be by 3D printing technique manufacturing artificial centrum, and by the Using prosthesis produced suffer from In person's body, human vertebra function is realized to substitute original backbone position.Therefore the design and production of prosthese are for patients It is extremely important.
And in the prior art, the preoperative cast of implantation material needed for generally requiring production dozens of 3D printing, those arts The foundation of preceding model is checked on the determining clinician for needing to be rich in experience and is just able to achieve.And it is limited to the influence of the prior art, There are biggish errors for the prosthese and ideal prosthetic produced, and cause patient's motor function after being implanted into prosthese limited, postoperative ridge The problems such as vertebra is inadequate in the stability of axial-rotation, low with the bond strength of upper hypocentrum occurs.
Summary of the invention
In view of this, the embodiment of the present application provides a kind of spine model generation method, spine model generates system and end End, to solve the preoperative cast of implantation material needed for needing to make dozens of 3D printing in the prior art, and the vertebra produced There is biggish error in prosthese and ideal prosthetic.
The first aspect of the embodiment of the present application provides a kind of spine model generation method, comprising:
According to non-lesion spine image, training generates confrontation network model;
Network model is fought by the generation after training, image is carried out to lesion site in lesion spine image It repairs, obtains targets spine image, wherein lesion site described in the targets spine image is repaired as non-lesion ridge Cylindricality state;
Three-dimensional reconstruction is carried out to the targets spine image, obtains corresponding targets spine model.
The second aspect of the embodiment of the present application provides a kind of spine model generation system, comprising:
Model training module, for according to non-lesion spine image, training to generate confrontation network model;
Image repair module, for fighting network model by the generation after training, to sick in lesion spine image Become happening part and carry out image repair, obtains targets spine image, wherein lesion site quilt described in the target image Repairing is non-lesion spine form;
Model building module obtains corresponding targets spine mould for carrying out three-dimensional reconstruction to the targets spine image Type.
The third aspect of the embodiment of the present application provides a kind of terminal, including memory, processor and is stored in described In memory and the computer program that can run on the processor, the processor are realized when executing the computer program The step of method as described in relation to the first aspect.
The fourth aspect of the embodiment of the present application provides a kind of computer readable storage medium, the computer-readable storage Media storage has computer program, and the step of method as described in relation to the first aspect is realized when the computer program is executed by processor Suddenly.
The 5th aspect of the application provides a kind of computer program product, and the computer program product includes computer Program is realized when the computer program is executed by one or more processors such as the step of above-mentioned first aspect the method.
Therefore in the embodiment of the present application, by being based on non-lesion spine image, training generates confrontation network model, Network model is fought using the generation after training, image repair is carried out to lesion site in lesion spine image, by lesion Happening part reparation is non-lesion spine form, obtains targets spine image, carries out three-dimensional reconstruction to targets spine image, obtains Final targets spine model, the process pass through using confrontation network is generated, restore to lesion spine image, with as far as possible A vertebral model ideally is obtained, the reconstruction of as a normal as possible, complete centrum is realized, avoids needing at present The problem of making the three-dimensional preoperative cast of dozens of implantation material promotes the precision of spinal prostheses production, reduces Using prosthesis and suffers from Activity limitation after in person's body.
Detailed description of the invention
It in order to more clearly explain the technical solutions in the embodiments of the present application, below will be to embodiment or description of the prior art Needed in attached drawing be briefly described, it should be apparent that, the accompanying drawings in the following description is only some of the application Embodiment for those of ordinary skill in the art without any creative labor, can also be according to these Attached drawing obtains other attached drawings.
Fig. 1 is a kind of flow chart one of spine model generation method provided by the embodiments of the present application;
Fig. 2 is a kind of flowchart 2 of spine model generation method provided by the embodiments of the present application;
Fig. 3 is the backbone mask image in the embodiment of the present application;
Fig. 4 is the targets spine image after the reparation in the embodiment of the present application;
Fig. 5 is to carry out the two of the spine regions that semantic segmentation obtains based on V-Net neural network in the embodiment of the present application It is worth image;
Fig. 6 is the structure chart that a kind of spine model provided by the embodiments of the present application generates system;
Fig. 7 is a kind of structure chart of terminal provided by the embodiments of the present application.
Specific embodiment
In being described below, for illustration and not for limitation, the tool of such as particular system structure, technology etc is proposed Body details, so as to provide a thorough understanding of the present application embodiment.However, it will be clear to one skilled in the art that there is no these specific The application also may be implemented in the other embodiments of details.In other situations, it omits to well-known system, device, electricity The detailed description of road and method, so as not to obscure the description of the present application with unnecessary details.
It should be appreciated that ought use in this specification and in the appended claims, term " includes " instruction is described special Sign, entirety, step, operation, the presence of element and/or component, but be not precluded one or more of the other feature, entirety, step, Operation, the presence or addition of element, component and/or its set.
It is also understood that mesh of the term used in this present specification merely for the sake of description specific embodiment And be not intended to limit the application.As present specification and it is used in the attached claims, unless on Other situations are hereafter clearly indicated, otherwise " one " of singular, "one" and "the" are intended to include plural form.
It will be further appreciated that the term "and/or" used in present specification and the appended claims is Refer to any combination and all possible combinations of one or more of associated item listed, and including these combinations.
As used in this specification and in the appended claims, term " if " can be according to context quilt Be construed to " when ... " or " once " or " in response to determination " or " in response to detecting ".Similarly, phrase " if it is determined that " or " if detecting [described condition or event] " can be interpreted to mean according to context " once it is determined that " or " in response to true It is fixed " or " once detecting [described condition or event] " or " in response to detecting [described condition or event] ".
In the specific implementation, terminal described in the embodiment of the present application is including but not limited to such as with touch sensitive surface The mobile phone, laptop computer or tablet computer of (for example, touch-screen display and/or touch tablet) etc it is other just Portable device.It is to be further understood that in certain embodiments, the equipment is not portable communication device, but there is touching Touch the desktop computer of sensing surface (for example, touch-screen display and/or touch tablet).
In following discussion, the terminal including display and touch sensitive surface is described.It is, however, to be understood that It is that terminal may include one or more of the other physical user-interface device of such as physical keyboard, mouse and/or control-rod.
Terminal supports various application programs, such as one of the following or multiple: drawing application program, demonstration application journey Sequence, word-processing application, website create application program, disk imprinting application program, spreadsheet applications, game application Program, telephony application, videoconference application, email application, instant messaging applications, exercise Support application program, photo management application program, digital camera application program, digital camera application program, web-browsing application Program, digital music player application and/or video frequency player application program.
The various application programs that can be executed at the terminal can be used such as touch sensitive surface at least one is public Physical user-interface device.It can adjust and/or change among applications and/or in corresponding application programs and touch sensitive table The corresponding information shown in the one or more functions and terminal in face.In this way, the public physical structure of terminal is (for example, touch Sensing surface) it can support the various application programs with user interface intuitive and transparent for a user.
It should be understood that the size of the serial number of each step is not meant that the order of the execution order in the present embodiment, each process Execution sequence should be determined by its function and internal logic, and the implementation process without coping with the embodiment of the present application constitutes any restriction.
In order to illustrate technical solution described herein, the following is a description of specific embodiments.
It is a kind of flow chart one of spine model generation method provided by the embodiments of the present application referring to Fig. 1, Fig. 1.Such as Fig. 1 institute Show, a kind of spine model generation method, method includes the following steps:
Step 101, according to non-lesion spine image, training generates confrontation network model.
The non-lesion spine image is based on image obtained from the backbone that lesion does not occur.Non-lesion spine image tool Body is the bianry image of a spine regions, or can be based on the non-lesion spine image, and auxiliary, which is realized, generates confrontation network mould The other types image of type training, for example, gray level image.
Specifically, generation confrontation network model includes two network models: generating network model and differentiates network model. Wherein, the network that network is a production is generated, it receives a random noise, generates image by this noise, defeated Image G out;Differentiate that network is used to judge whether true a picture is, its input x is a picture, and output D (x) indicates X is that the probability of true picture means that x is a true picture if it is 1, and exporting is 0, means that and is unlikely to be true Real picture.During model training, the target for generating network is just to try to generate true picture and go to cheat to differentiate network, And differentiate that the target of network is just to try to determine and generate picture and true image that network generates, such G and D constitute one A dynamic " gambling process ", final equalization point, that is, Nash equilibrium point.
Generation confrontation network model is exactly by generating network and differentiation for being applied in deep learning neural network The continuous game of network generates, training is completed so that generating the distribution of e-learning to data if using picture Afterwards, image true to nature can be generated from one section of random number by generating network.
Step 102, network model is fought by the generation after training, to lesion site in lesion spine image Image repair is carried out, targets spine image is obtained.
Wherein, lesion site described in the target image is repaired as non-lesion spine form.
Specifically, it is generally the case that in lesion spine image, lesion site can be influenced to make by sick cells such as tumours Script natural spine area information in spine image is lost, such as backbone profile is imperfect, spine regions abnormal deformation etc.. Those image-regions need to carry out image repair, to obtain the targets spine image under natural spine state.
In this step, network model is fought by the generation after the training of non-lesion spine image, by lesion spine image It is repaired, is repaired targets spine image for non-lesion spine form to obtain lesion site, with can be to lesion Spine image is restored, and the image after recovery is applied to correspond to the three-dimension modeling at backbone position after full backbone resection In the process, to realize manufacturing for ideal spinal prostheses, the dependence to human factor is reduced, promotes the essence of spinal prostheses production Accuracy, the malaise symptoms after reducing Using prosthesis patient's body.
When specific implementation, how the realization process of image repair is utilized it is to be understood that dig a hole on piece image This hole completion is allowed human eye that can not identify the part of completion by other information.This problem is easy to the mankind, But but seem addition difficult for computer, secondly how this problem first utilizes other either with or without the solution uniquely determined Information, there are also whether the result for how judging completion true enough.
In the embodiment of the present application, as an optional embodiment, wherein fight net by the generation after training Network model carries out image repair to lesion site in lesion spine image, obtains targets spine image, comprising:
Based on the lesion spine image and preset two-value mask code matrix, generates and covered in lesion site addition The backbone mask image of code.
It is repaired based on the backbone mask image and backbone referring to image, network mould is fought by the generation after training Type generates the targets spine image F '=M ⊙ F+ (1-M) ⊙ G (z) after repairing*
Wherein, F is the lesion spine image;M is the preset two-value mask code matrix, in the two-value mask code matrix Element it is corresponding with the pixel of the lesion spine image;M ⊙ F is the backbone mask image;G(z)*It is repaired for the backbone The multiple symbol generated during model training referring to image for the generation confrontation network model referring to image, the backbone reparation Setting is closed referring to the image of condition;⊙ is the operation that matrix element is multiplied with pixel point value in image.
The process, firstly, defining several symbols for image repair.M indicates the mask code matrix of a two-value, therein Numerical value only has 0 or 1;The region for needing to repair in 0 one picture of expression, 1 indicates the part to be retained in same picture.
Then, we define how to repair lesion spine image F by given mask, and specific practice is exactly to allow Metzler matrix In element be multiplied with the corresponding pixel point value in F, the method that this matrix is multiplied with respective pixel point value is called Harbin-to-Dalian horse Product is indicated with M ⊙ F, and then on the basis of lesion spine image, obtains being added to mask in lesion site wherein Backbone mask image, as shown in Figure 3.Next, we are from confrontation network model is generated during model training, generator is raw At result images in find one meet setting referring to condition picture G (z)*, as enterprising in backbone mask image basis Backbone when row image repair is repaired referring to image.
We can define targets spine image F '=M ⊙ F+ (1-M) ⊙ G (z) after repairing*
During before, we fight network model (specially WGAN-GP network according to trained generation is obtained Model), the result i.e. targets spine image using the backbone mask image M ⊙ F after addition mask as its input, after being repaired F ', as shown in Figure 4.
Wherein, this meets setting referring to the picture G (z) of condition*Network model is fought in model training mistake for aforementioned generation In the image generated in journey between image F an as far as possible similar optimal figure of the corresponding pixel points in the distribution of value, i.e., Similarity exceeds an optimal figure of threshold value.
Image G (z)*In contain spine regions corresponding with lesion site in lesion spine image, and remove the ridge Other regions except columnar region;(1-M)⊙G(z)*It has then found out image G (z)*In other regions add mask, retain with The backbone of the corresponding spine regions part of lesion site is repaired referring to mask image in lesion spine image.M ⊙ F is represented Backbone mask image and (1-M) ⊙ G (z)*The backbone of representative is repaired to be superimposed referring to mask image pixel, after obtaining the reparation Targets spine image F '.
It further, is realization to optimal figure G (z)*Determination, it is described based on described as an optional embodiment Backbone mask image and backbone are repaired referring to image, are fought network model by the generation after training, are generated after repairing Before the targets spine image, further includes:
Obtain the output image for generating confrontation network model and generating during model training;
According to the output image, according to loss function L=Lcontextual+λLperceptual, obtained from the output image The backbone is taken to repair referring to image;
Wherein, λ is hyper parameter;
Lcontextual=| | M ⊙ G (z)-M ⊙ F | |1, G (z) is the output image;
Lperceptual=log (1-D (G (z))), D (G (z)) be training after the generation confrontation network model in differentiate Differentiation end value of the device to the output image;
The backbone reparation is referring to the output that image is when the value of the loss function meets sets requirement Image.
Specifically, this, which meets sets requirement specifically, may is that the variation trend of the value of the loss function meets sets requirement, The variation trend of for example, value of the loss function is the trend that approach variation is carried out towards a target value, is specially in court The variation trend of the trend for carrying out approach variation to 0, the i.e. value of the loss function is downward trend.Or this meets sets requirement It specifically may is that the value of the loss function is in setting range, for example, close in 0 numberical range.
From formula F '=M ⊙ F+ (1-M) ⊙ G (z), and generation confrontation network model is combined to generate during model training Output image G (z) the fact that have multiple situation, it will be seen that most importantly to the targets spine image after reparation Section 2 generating portion, that is, need to realize good image repair loss zone algorithm.Find out the optimum image G in G (z) (z)*
In order to realize the purpose, it would be desirable to two important informations: contextual information and perception information, the two information Acquisition mainly by loss function realize, loss function is smaller, and the result of image repair is better.
To guarantee the identical contextual information of input picture, picture F and pixel corresponding in G (z) are similar as far as possible, because This needs to make punishment, loss function to the G (z) for generating dissimilar pixel are as follows:
Lcontextual=| | M ⊙ G (z)-M ⊙ F | |1
The most ideal situation is that all pixels value of F and G (z) be it is identical, loss function value be 0.
In order to obtain the reparation picture of high quality, it would be desirable to allow arbiter that there is the ability for correctly differentiating true picture, Loss function are as follows:
Lperceptual=log (1-D (G (z)));
Therefore final loss function are as follows:
L=Lcontextual+λLperceptual
Wherein, λ is set in advance, is a hyper parameter;D(G(z))∈[0,1].
In view of the loss in the reality and image information controlled range in image procossing, when specific implementation, backbone reparation Referring to the output image that image is when the value of the loss function is in setting range;The setting range be close to 0 range.
Step 103, three-dimensional reconstruction is carried out to the targets spine image, obtains corresponding targets spine model.
The targets spine image is a bianry image.Finally, the targets spine image after reparation is subjected to three-dimensional reconstruction, Normal spine model is obtained, which can perform the operation by 3D printing technique for TES.
In the embodiment of the present application, by being based on non-lesion spine image, training generates confrontation network model, after training Generation fight network model, in lesion spine image lesion site carry out image repair, lesion site is repaired It is again non-lesion spine form, obtains targets spine image, three-dimensional reconstruction is carried out to targets spine image, obtains final target Spine model, the process pass through using confrontation network is generated, restore to lesion spine image, to obtain a reason as far as possible Think the vertebral model under state, realizes the reconstruction of as a normal as possible, complete centrum, avoid needing to make dozens of at present The problem of three-dimensional preoperative cast of implantation material, promotes the precision of spinal prostheses production, after reducing Using prosthesis patient's body Activity limitation.
The different embodiments of spine model generation method are additionally provided in the embodiment of the present application.
Referring to fig. 2, Fig. 2 is a kind of flowchart 2 of spine model generation method provided by the embodiments of the present application.Such as Fig. 2 institute Show, a kind of spine model generation method, method includes the following steps:
Step 201, semantic segmentation is carried out to the computer tomography CT image of non-lesion backbone, obtains spine regions First bianry image.
Firstly, carrying out semantic segmentation to non-lesion spine CT image by neural network, acquisition can clearly distinguish spinal area The binary map in domain and non-spinal region.Specifically, spine regions gray value is 255 in first bianry image, non-spinal region Gray value is 0.
Specifically, semantic segmentation is carried out to the computer tomography CT image of non-lesion backbone, can be based on V-Net The spine CT image segmentation that neural network carries out.
V-Net is a typical network structure end to end, aiming at the problem that be three-dimensional image segmentation.Here, we It is used to carry out two dimensional image segmentation.When carrying out non-lesion spine CT image segmentation using V-Net neural network, by non-disease The two-dimensional image information for becoming spine CT image is inputted, and depth information is ignored.
Specifically, in V-Net neural network structure here, left side is the process of a down-sampling, for capturing up and down Literary information, point 5 groups of convolution operations carry out, and similar pooling is carried out once after every group of convolution operation and is operated, is specifically exactly to walk The convolution kernel for the 2*2 that width is 2 carries out deconvolution, is original 1/2 by image down.By four groups of operations by size be 512* 512 image becomes the image that size is 32*32.Right side in structure is upper sampling process, is come in conjunction with each layer information of down-sampling Detailed information is restored, using 4 groups of deconvolution, picture is extended to original 2 times by up-sampling every time, then by corresponding characteristic pattern It is replicated and the result of left side down-sampling is linked, the figure of 512*512*16 size is obtained after upper sampling process, most Port number is reduced to 1 with the convolution kernel of a 1*1 afterwards.Do not have to change the size of image when making convolution during down-sampling, this The benefit that sample is done is omitted the work of cutting because the size of respective layer be it is the same, convolution operation can be retained in this way and mentioned The all information got.First bianry image of the spine regions obtained after segmentation is as shown in Figure 5.
Step 202, it is based on first bianry image, the training generation fights network model.
The bianry image of the natural spine CT image of acquisition is put into generation confrontation network to be trained, with trained life Image repair is carried out to tumor of spine happening part at confrontation network model, obtains the natural spine bianry image at the position.
Specifically, when obtaining the bianry image of non-lesion spine CT image, it is possible that spine regions phase in image Than the situation too small in background area, therefore first bianry image can also be advanced optimized, by the first bianry image into Row is cut, and the image after cutting is amplified, first bianry image after being optimized, and then training generates confrontation network Model.It specifically can be, the image that initial size is 512*512 cut and scaled, 64*64 is become;By treated Image is put into WGAN-GP and is trained, and finally obtains training pattern, training pattern meeting output model training result, i.e., aforementioned reality It applies and generates the image that confrontation network model generates during model training in mode.
Step 201 and step 202 realize that, according to non-lesion spine image, training generates the implementation of confrontation network model Journey.
Step 203, semantic segmentation is carried out to the CT image of lesion backbone, obtains the second bianry image of spine regions.
In the step, it is accomplished that the processing to the CT image of lesion vertebra, generates the second bianry image of spine regions. This based on CT image carry out semantic segmentation obtain spine regions bianry image treatment process with reference can be made to step 201 in non- The CT image of lesion backbone carries out semantic segmentation, obtains the treatment process of the bianry image of spine regions, details are not described herein again.
Step 204, determine that second bianry image is the lesion spine image.
When fighting network model progress image repair by the generation after training, the image of reparation need to be a binary map Therefore picture in this step, by semantic segmentation and handles obtained the second bianry image of spine regions and is determined as lesion backbone Image is prepared for subsequent image reparation.
Step 205, network model is fought by the generation after training, to lesion site in lesion spine image Image repair is carried out, targets spine image is obtained.
The realization process of the step and the realization process of the step 102 in aforementioned embodiments are identical, and details are not described herein again.
Step 206, three-dimensional reconstruction is carried out to the targets spine image by marching cubes algorithm, obtains corresponding institute State targets spine model.
Here, three-dimensional reconstruction is carried out to the tumor of spine bianry image after reparation using marching cubes algorithm, obtained most Whole targets spine model, technology are easy to accomplish.
In the embodiment of the present application, by being based on non-lesion spine image, training generates confrontation network model, after training Generation fight network model, in lesion spine image lesion site carry out image repair, lesion site is repaired It is again non-lesion spine form, obtains targets spine image, three-dimensional reconstruction is carried out to targets spine image, obtains final target Spine model, the process pass through using confrontation network is generated, restore to lesion spine image, to obtain a reason as far as possible Think the vertebral model under state, realizes the reconstruction of as a normal as possible, complete centrum, avoid needing to make dozens of at present The problem of three-dimensional preoperative cast of implantation material, promotes the precision of spinal prostheses production, after reducing Using prosthesis patient's body Activity limitation.
It is the structure chart that Fig. 6 is a kind of spine model generation system provided by the embodiments of the present application, in order to just referring to Fig. 6 In explanation, part relevant to the embodiment of the present application is illustrated only.
It includes: model training module 301, image repair module 302 and model foundation that the spine model, which generates system 300, Module 303.
Model training module 301, for according to non-lesion spine image, training to generate confrontation network model;
Image repair module 302, for fighting network model by the generation after training, in lesion spine image Lesion site carries out image repair, obtains targets spine image, wherein lesion site described in the target image It is repaired as non-lesion spine form;
Model building module 303 obtains corresponding targets spine for carrying out three-dimensional reconstruction to the targets spine image Model.
Wherein, the model training module 301 is specifically used for:
Semantic segmentation is carried out to the computer tomography CT image of non-lesion backbone, obtains the first two-value of spine regions Image;
Based on first bianry image, the training generation fights network model.
Described image repair module is specifically used for:
Based on the lesion spine image and preset two-value mask code matrix, generates and covered in lesion site addition The backbone mask image of code;
It is repaired based on the backbone mask image and backbone referring to image, network mould is fought by the generation after training Type generates the targets spine image F '=M ⊙ F+ (1-M) ⊙ G (z) after repairing*
Wherein, F is the lesion spine image;M is the preset two-value mask code matrix, in the two-value mask code matrix Element it is corresponding with the pixel of the lesion spine image;M ⊙ F is the backbone mask image;G(z)*It is repaired for the backbone The multiple symbol generated during model training referring to image for the generation confrontation network model referring to image, the backbone reparation Setting is closed referring to the image of condition;⊙ is the operation that matrix element is multiplied with pixel point value in image.
Described image repair module is also used to:
Obtain the output image for generating confrontation network model and generating during model training;
According to the output image, according to loss function L=Lcontextual+λLperceptual, obtained from the output image The backbone is taken to repair referring to image;
Wherein, λ is hyper parameter;
Lcontextual=| | M ⊙ G (z)-M ⊙ F | |1, G (z) is the output image;
Lperceptual=log (1-D (G (z))), D (G (z)) be training after the generation confrontation network model in differentiate Differentiation end value of the device to the output image;
The backbone reparation is referring to the output that image is when the value of the loss function meets sets requirement Image.
Spine model generates system further include:
Module is obtained, semantic segmentation is carried out for the CT image to lesion backbone, obtains the second binary map of spine regions Picture;
Determining module, for determining that second bianry image is the lesion spine image.
The model building module 303 is specifically used for:
Three-dimensional reconstruction is carried out to the targets spine image by marching cubes algorithm, obtains the corresponding target ridge Column model.
Spine model provided by the embodiments of the present application, which generates system, can be realized the implementation of above-mentioned spine model generation method Each process of example, and identical technical effect can be reached, to avoid repeating, which is not described herein again.
Fig. 7 is a kind of structure chart of terminal provided by the embodiments of the present application.As shown in Fig. 7, the terminal 4 of the embodiment is wrapped It includes: processor 40, memory 41 and being stored in the computer that can be run in the memory 41 and on the processor 40 Program 42.
Illustratively, the computer program 42 can be divided into one or more module/units, it is one or Multiple module/units are stored in the memory 41, and are executed by the processor 40, to complete the application.Described one A or multiple module/units can be the series of computation machine program instruction section that can complete specific function, which is used for Implementation procedure of the computer program 42 in the terminal 4 is described.For example, the computer program 42 can be divided into Model training module, image repair module, model building module, acquisition module and determining module, each module concrete function are as follows:
Model training module, for according to non-lesion spine image, training to generate confrontation network model;
Image repair module, for fighting network model by the generation after training, to sick in lesion spine image Become happening part and carry out image repair, obtains targets spine image, wherein lesion site quilt described in the target image Repairing is non-lesion spine form;
Model building module obtains corresponding targets spine mould for carrying out three-dimensional reconstruction to the targets spine image Type.
Wherein, the model training module is specifically used for:
Semantic segmentation is carried out to the computer tomography CT image of non-lesion backbone, obtains the first two-value of spine regions Image;
Based on first bianry image, the training generation fights network model.
Described image repair module is specifically used for:
Based on the lesion spine image and preset two-value mask code matrix, generates and covered in lesion site addition The backbone mask image of code;
It is repaired based on the backbone mask image and backbone referring to image, network mould is fought by the generation after training Type generates the targets spine image F '=M ⊙ F+ (1-M) ⊙ G (z) after repairing*
Wherein, F is the lesion spine image;M is the preset two-value mask code matrix, in the two-value mask code matrix Element it is corresponding with the pixel of the lesion spine image;M ⊙ F is the backbone mask image;G(z)*It is repaired for the backbone The multiple symbol generated during model training referring to image for the generation confrontation network model referring to image, the backbone reparation Setting is closed referring to the image of condition;⊙ is the operation that matrix element is multiplied with pixel point value in image.
Described image repair module is also used to:
Obtain the output image for generating confrontation network model and generating during model training;
According to the output image, according to loss function L=Lcontextual+λLperceptual, obtained from the output image The backbone is taken to repair referring to image;
Wherein, λ is hyper parameter;
Lcontextual=| | M ⊙ G (z)-M ⊙ F | |1, G (z) is the output image;
Lperceptual=log (1-D (G (z))), D (G (z)) be training after the generation confrontation network model in differentiate Differentiation end value of the device to the output image;
The backbone reparation is referring to the output that image is when the value of the loss function meets sets requirement Image.
Module is obtained, semantic segmentation is carried out for the CT image to lesion backbone, obtains the second binary map of spine regions Picture;
Determining module, for determining that second bianry image is the lesion spine image.
The model building module is specifically used for:
Three-dimensional reconstruction is carried out to the targets spine image by marching cubes algorithm, obtains the corresponding target ridge Column model.
The terminal 4 can be desktop PC, notebook, palm PC and cloud server etc. and calculate equipment.Institute Stating terminal 4 may include, but be not limited only to, processor 40, memory 41.It will be understood by those skilled in the art that Fig. 7 is only eventually The example at end 4, the not restriction of structure paired terminal 4 may include than illustrating more or fewer components, or the certain portions of combination Part or different components, such as the terminal can also include input-output equipment, network access equipment, bus etc..
Alleged processor 40 can be central processing unit (Central Processing Unit, CPU), can also be Other general processors, digital signal processor (Digital Signal Processor, DSP), specific integrated circuit (Application Specific Integrated Circuit, ASIC), ready-made programmable gate array (Field- Programmable Gate Array, FPGA) either other programmable logic device, discrete gate or transistor logic, Discrete hardware components etc..General processor can be microprocessor or the processor is also possible to any conventional processor Deng.
The memory 41 can be the internal storage unit of the terminal 4, such as the hard disk or memory of terminal 4.It is described Memory 41 is also possible to the External memory equipment of the terminal 4, such as the plug-in type hard disk being equipped in the terminal 4, intelligence Storage card (Smart Media Card, SMC), secure digital (Secure Digital, SD) card, flash card (Flash Card) Deng.Further, the memory 41 can also both include the internal storage unit of the terminal 4 or set including external storage It is standby.The memory 41 is for other programs and data needed for storing the computer program and the terminal.It is described to deposit Reservoir 41 can be also used for temporarily storing the data that has exported or will export.
It is apparent to those skilled in the art that for convenience of description and succinctly, only with above-mentioned each function Can unit, module division progress for example, in practical application, can according to need and by above-mentioned function distribution by different Functional unit, module are completed, i.e., the internal structure of described device is divided into different functional unit or module, more than completing The all or part of function of description.Each functional unit in embodiment, module can integrate in one processing unit, can also To be that each unit physically exists alone, can also be integrated in one unit with two or more units, it is above-mentioned integrated Unit both can take the form of hardware realization, can also realize in the form of software functional units.In addition, each function list Member, the specific name of module are also only for convenience of distinguishing each other, the protection scope being not intended to limit this application.Above system The specific work process of middle unit, module, can refer to corresponding processes in the foregoing method embodiment, and details are not described herein.
In the above-described embodiments, it all emphasizes particularly on different fields to the description of each embodiment, is not described in detail or remembers in some embodiment The part of load may refer to the associated description of other embodiments.
Those of ordinary skill in the art may be aware that list described in conjunction with the examples disclosed in the embodiments of the present disclosure Member and algorithm steps can be realized with the combination of electronic hardware or computer software and electronic hardware.These functions are actually It is implemented in hardware or software, the specific application and design constraint depending on technical solution.Professional technician Each specific application can be used different methods to achieve the described function, but this realization is it is not considered that exceed Scope of the present application.
In embodiment provided herein, it should be understood that disclosed terminal and method can pass through others Mode is realized.For example, terminal embodiment described above is only schematical, for example, the division of the module or unit, Only a kind of logical function partition, there may be another division manner in actual implementation, such as multiple units or components can be with In conjunction with or be desirably integrated into another system, or some features can be ignored or not executed.Another point, it is shown or discussed Mutual coupling or direct-coupling or communication connection can be through some interfaces, the INDIRECT COUPLING of device or unit or Communication connection can be electrical property, mechanical or other forms.
The unit as illustrated by the separation member may or may not be physically separated, aobvious as unit The component shown may or may not be physical unit, it can and it is in one place, or may be distributed over multiple In network unit.It can select some or all of unit therein according to the actual needs to realize the mesh of this embodiment scheme 's.
It, can also be in addition, each functional unit in each embodiment of the application can integrate in one processing unit It is that each unit physically exists alone, can also be integrated in one unit with two or more units.Above-mentioned integrated list Member both can take the form of hardware realization, can also realize in the form of software functional units.
If the integrated module/unit be realized in the form of SFU software functional unit and as independent product sale or In use, can store in a computer readable storage medium.Based on this understanding, the application realizes above-mentioned implementation All or part of the process in example method, can also instruct relevant hardware to complete, the meter by computer program Calculation machine program can be stored in a computer readable storage medium, the computer program when being executed by processor, it can be achieved that on The step of stating each embodiment of the method.Wherein, the computer program includes computer program code, the computer program generation Code can be source code form, object identification code form, executable file or certain intermediate forms etc..The computer-readable medium It may include: any entity or device, recording medium, USB flash disk, mobile hard disk, magnetic that can carry the computer program code Dish, CD, computer storage, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), electric carrier signal, telecommunication signal and software distribution medium etc..It should be noted that described The content that computer-readable medium includes can carry out increasing appropriate according to the requirement made laws in jurisdiction with patent practice Subtract, such as does not include electric carrier signal and electricity according to legislation and patent practice, computer-readable medium in certain jurisdictions Believe signal.
Embodiment described above is only to illustrate the technical solution of the application, rather than its limitations;Although referring to aforementioned reality Example is applied the application is described in detail, those skilled in the art should understand that: it still can be to aforementioned each Technical solution documented by embodiment is modified or equivalent replacement of some of the technical features;And these are modified Or replacement, the spirit and scope of each embodiment technical solution of the application that it does not separate the essence of the corresponding technical solution should all Comprising within the scope of protection of this application.

Claims (10)

1. a kind of spine model generation method characterized by comprising
According to non-lesion spine image, training generates confrontation network model;
Network model is fought by the generation after training, image is carried out to lesion site in lesion spine image and is repaired It is multiple, obtain targets spine image, wherein lesion site described in the targets spine image is repaired as non-lesion backbone Form;
Three-dimensional reconstruction is carried out to the targets spine image, obtains corresponding targets spine model.
2. spine model generation method according to claim 1, which is characterized in that it is described according to non-lesion spine image, Training generates confrontation network model, comprising:
Semantic segmentation is carried out to the computer tomography CT image of non-lesion backbone, obtains the first binary map of spine regions Picture;
Based on first bianry image, the training generation fights network model.
3. spine model generation method according to claim 1, which is characterized in that the generation by after training Network model is fought, image repair is carried out to lesion site in lesion spine image, obtains targets spine image, comprising:
Based on the lesion spine image and preset two-value mask code matrix, generate in lesion site addition mask Backbone mask image;
It is repaired based on the backbone mask image and backbone referring to image, network model is fought by the generation after training, Generate the targets spine image F '=M ⊙ F+ (1-M) ⊙ G (z) after repairing*
Wherein, F is the lesion spine image;M is the preset two-value mask code matrix, the member in the two-value mask code matrix It is plain corresponding with the pixel of the lesion spine image;M ⊙ F is the backbone mask image;G(z)*For backbone reparation ginseng According to image, the backbone reparation is that described generate sets confrontation meeting of generating during model training of network model referring to image The fixed image referring to condition;⊙ is the operation that matrix element is multiplied with pixel point value in image.
4. spine model generation method according to claim 3, which is characterized in that described to be based on the backbone mask image And backbone is repaired referring to image, is fought network model by the generation after training, is generated the targets spine after repairing Before image, further includes:
Obtain the output image for generating confrontation network model and generating during model training;
According to the output image, according to loss function L=Lcontextual+λLperceptual, institute is obtained from the output image Backbone is stated to repair referring to image;
Wherein, λ is hyper parameter;
Lcontextual=| | M ⊙ G (z)-M ⊙ F | |1, G (z) is the output image;
Lperceptual=log (1-D (G (z))), D (G (z)) are arbiter in the generation confrontation network model after training to institute State the differentiation end value of output image;
The backbone reparation is referring to the output image that image is when the value of the loss function meets sets requirement.
5. spine model generation method according to claim 1, which is characterized in that the generation by after training Network model is fought, image repair is carried out to lesion site in lesion spine image, before obtaining targets spine image, also Include:
Semantic segmentation is carried out to the CT image of lesion backbone, obtains the second bianry image of spine regions;
Determine that second bianry image is the lesion spine image.
6. spine model generation method according to claim 1, which is characterized in that it is described to the targets spine image into Row three-dimensional reconstruction obtains corresponding targets spine model, comprising:
Three-dimensional reconstruction is carried out to the targets spine image by marching cubes algorithm, obtains the corresponding targets spine mould Type.
7. a kind of spine model generates system characterized by comprising
Model training module, for according to non-lesion spine image, training to generate confrontation network model;
Image repair module sends out lesion in lesion spine image for fighting network model by the generation after training Raw position carries out image repair, obtains targets spine image, wherein lesion site described in the target image is repaired For non-lesion spine form;
Model building module obtains corresponding targets spine model for carrying out three-dimensional reconstruction to the targets spine image.
8. spine model according to claim 7 generates system, which is characterized in that the model training module is specifically used In:
Semantic segmentation is carried out to the computer tomography CT image of non-lesion backbone, obtains the first binary map of spine regions Picture;
Based on first bianry image, the training generation fights network model.
9. a kind of terminal, including memory, processor and storage can be run in the memory and on the processor Computer program, which is characterized in that the processor is realized when executing the computer program as claim 1 to 6 is any The step of item the method.
10. a kind of computer readable storage medium, the computer-readable recording medium storage has computer program, and feature exists In when the computer program is executed by processor the step of any one of such as claim 1 to 6 of realization the method.
CN201910450859.5A 2019-05-28 2019-05-28 Spine model generation method, spine model generation system and terminal Active CN110246216B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910450859.5A CN110246216B (en) 2019-05-28 2019-05-28 Spine model generation method, spine model generation system and terminal

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910450859.5A CN110246216B (en) 2019-05-28 2019-05-28 Spine model generation method, spine model generation system and terminal

Publications (2)

Publication Number Publication Date
CN110246216A true CN110246216A (en) 2019-09-17
CN110246216B CN110246216B (en) 2023-02-28

Family

ID=67885234

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910450859.5A Active CN110246216B (en) 2019-05-28 2019-05-28 Spine model generation method, spine model generation system and terminal

Country Status (1)

Country Link
CN (1) CN110246216B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111127412A (en) * 2019-12-18 2020-05-08 广东工业大学 Pathological image recognition device based on generation countermeasure network
CN111274990A (en) * 2020-02-11 2020-06-12 广东同天投资管理有限公司 Computer device and storage medium for classifying spinal morphology
CN111353995A (en) * 2020-03-31 2020-06-30 成都信息工程大学 Cervical single cell image data generation method based on generation countermeasure network
CN112734740A (en) * 2021-01-18 2021-04-30 推想医疗科技股份有限公司 Method for training target detection model, method for detecting target and device thereof

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105723423A (en) * 2013-10-30 2016-06-29 皇家飞利浦有限公司 Volumetric image data visualization
CN107945140A (en) * 2017-12-20 2018-04-20 中国科学院深圳先进技术研究院 A kind of image repair method, device and equipment
CN107977971A (en) * 2017-11-09 2018-05-01 哈尔滨理工大学 The method of vertebra positioning based on convolutional neural networks
CN108038860A (en) * 2017-11-30 2018-05-15 杭州电子科技大学 Spine segmentation method based on the full convolutional neural networks of 3D
CN108460830A (en) * 2018-05-09 2018-08-28 厦门美图之家科技有限公司 Image repair method, device and image processing equipment
CN108765340A (en) * 2018-05-29 2018-11-06 Oppo(重庆)智能科技有限公司 Fuzzy image processing method, apparatus and terminal device
CN108765417A (en) * 2018-06-15 2018-11-06 西安邮电大学 It is a kind of that system and method is generated based on the femur X-ray film of deep learning and digital reconstruction irradiation image
CN109242865A (en) * 2018-09-26 2019-01-18 上海联影智能医疗科技有限公司 Medical image auto-partition system, method, apparatus and storage medium based on multichannel chromatogram
US20190103190A1 (en) * 2017-09-29 2019-04-04 K2M, Inc. Systems and methods for simulating spine and skeletal system pathologies
US20190105009A1 (en) * 2017-10-10 2019-04-11 Holo Surgical Inc. Automated segmentation of three dimensional bony structure images
WO2019079778A1 (en) * 2017-10-20 2019-04-25 Nuvasive, Inc. Intervertebral disc modeling

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105723423A (en) * 2013-10-30 2016-06-29 皇家飞利浦有限公司 Volumetric image data visualization
US20190103190A1 (en) * 2017-09-29 2019-04-04 K2M, Inc. Systems and methods for simulating spine and skeletal system pathologies
US20190105009A1 (en) * 2017-10-10 2019-04-11 Holo Surgical Inc. Automated segmentation of three dimensional bony structure images
WO2019079778A1 (en) * 2017-10-20 2019-04-25 Nuvasive, Inc. Intervertebral disc modeling
CN107977971A (en) * 2017-11-09 2018-05-01 哈尔滨理工大学 The method of vertebra positioning based on convolutional neural networks
CN108038860A (en) * 2017-11-30 2018-05-15 杭州电子科技大学 Spine segmentation method based on the full convolutional neural networks of 3D
CN107945140A (en) * 2017-12-20 2018-04-20 中国科学院深圳先进技术研究院 A kind of image repair method, device and equipment
CN108460830A (en) * 2018-05-09 2018-08-28 厦门美图之家科技有限公司 Image repair method, device and image processing equipment
CN108765340A (en) * 2018-05-29 2018-11-06 Oppo(重庆)智能科技有限公司 Fuzzy image processing method, apparatus and terminal device
CN108765417A (en) * 2018-06-15 2018-11-06 西安邮电大学 It is a kind of that system and method is generated based on the femur X-ray film of deep learning and digital reconstruction irradiation image
CN109242865A (en) * 2018-09-26 2019-01-18 上海联影智能医疗科技有限公司 Medical image auto-partition system, method, apparatus and storage medium based on multichannel chromatogram

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
刘忠利等: "基于深度学习的脊柱CT图像分割", 《计算机应用与软件》 *
陈锟等: "生成对抗网络在医学图像处理中的应用", 《生命科学仪器》 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111127412A (en) * 2019-12-18 2020-05-08 广东工业大学 Pathological image recognition device based on generation countermeasure network
CN111127412B (en) * 2019-12-18 2021-05-28 广东工业大学 Pathological image recognition device based on generation countermeasure network
CN111274990A (en) * 2020-02-11 2020-06-12 广东同天投资管理有限公司 Computer device and storage medium for classifying spinal morphology
CN111274990B (en) * 2020-02-11 2024-01-12 广东同天科技产业发展有限公司 Computer device and storage medium for classifying spinal morphology
CN111353995A (en) * 2020-03-31 2020-06-30 成都信息工程大学 Cervical single cell image data generation method based on generation countermeasure network
CN111353995B (en) * 2020-03-31 2023-03-28 成都信息工程大学 Cervical single cell image data generation method based on generation countermeasure network
CN112734740A (en) * 2021-01-18 2021-04-30 推想医疗科技股份有限公司 Method for training target detection model, method for detecting target and device thereof
CN112734740B (en) * 2021-01-18 2024-02-23 推想医疗科技股份有限公司 Method for training target detection model, target detection method and device

Also Published As

Publication number Publication date
CN110246216B (en) 2023-02-28

Similar Documents

Publication Publication Date Title
CN110246216A (en) Spine model generation method, spine model generate system and terminal
Ismael et al. An enhanced deep learning approach for brain cancer MRI images classification using residual networks
CN108257134A (en) Nasopharyngeal Carcinoma Lesions automatic division method and system based on deep learning
CN109966662A (en) A kind of method and system for verifying radiotherapy dosage
Tomatis et al. Late rectal bleeding after 3D-CRT for prostate cancer: development of a neural-network-based predictive model
Auricchio et al. Patient‐specific finite element analysis of carotid artery stenting: a focus on vessel modeling
Veneziano et al. Surface smoothing, decimation, and their effects on 3D biological specimens
CN110070540A (en) Image generating method, device, computer equipment and storage medium
CN108765340A (en) Fuzzy image processing method, apparatus and terminal device
CN113096137B (en) Adaptive segmentation method and system for OCT (optical coherence tomography) retinal image field
CN109785399B (en) Synthetic lesion image generation method, device, equipment and readable storage medium
Shiranthika et al. Supervised optimal chemotherapy regimen based on offline reinforcement learning
Qiu et al. Learnable ophthalmology sam
Amini et al. Augmented reality mastectomy surgical planning prototype using the HoloLens template for healthcare technology letters
CN110796064B (en) Human muscle image establishing method and device, storage medium and electronic equipment
CN114998374A (en) Image segmentation method, device and equipment based on position prior and storage medium
Jahan et al. Automated diagnosis of pneumonia from classification of chest x-ray im ages using efficientnet
Jayasinghe et al. Patient Reported Outcome Measures After Breast Augmentation–Using the BREAST-Q IS
CN115985513B (en) Data processing method, device and equipment based on multiple groups of chemical cancer typing
CN116612849A (en) Medical treatment scheme recommendation method, device, equipment and storage medium
CN115760813A (en) Screw channel generation method, device, equipment, medium and program product
Jerg et al. Diffuse domain method for needle insertion simulations
Huang et al. Development of a deep learning-based tool to assist wound classification
Damayanti et al. Lung cancer classification using convolutional neural network and DenseNet
US20060264713A1 (en) Disease and therapy dissemination representation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant