CN109671053A - A kind of gastric cancer image identification system, device and its application - Google Patents

A kind of gastric cancer image identification system, device and its application Download PDF

Info

Publication number
CN109671053A
CN109671053A CN201811371788.1A CN201811371788A CN109671053A CN 109671053 A CN109671053 A CN 109671053A CN 201811371788 A CN201811371788 A CN 201811371788A CN 109671053 A CN109671053 A CN 109671053A
Authority
CN
China
Prior art keywords
image
frame
lesion
layer
candidate region
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201811371788.1A
Other languages
Chinese (zh)
Inventor
朱圣韬
张澍田
闵力
陈蕾
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Friendship Hospital
Original Assignee
Beijing Friendship Hospital
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Friendship Hospital filed Critical Beijing Friendship Hospital
Priority to CN201811371788.1A priority Critical patent/CN109671053A/en
Priority to CN202210461368.2A priority patent/CN115345819A/en
Publication of CN109671053A publication Critical patent/CN109671053A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10068Endoscopic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30092Stomach; Gastric

Landscapes

  • Engineering & Computer Science (AREA)
  • Quality & Reliability (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The present invention relates to a kind of gastric cancer image identification system, device and its application, which includes data input module, data preprocessing module, image recognition model construction module and lesion identification module;Self training may be implemented in the system simultaneously, to accurately identify the diseased region in gastric cancer image.

Description

A kind of gastric cancer image identification system, device and its application
Technical field
The invention belongs to medical domain, the technology that pathological image identification is realized using image identification system is more particularly related to Field.
Background technique
Although the disease incidence of gastric cancer was gradually reduced from 1975, still there are within 2012 nearly 1,000,000 new cases (total 951000, the 6.8% of Zhan Suoyou cancer morbidity), make the fifth-largest most common malignant tumour in the world.Wherein, have Case more than 70% appears in developing country, and has half to occur East Asia (mainly in China).In terms of the death rate, stomach Cancer is the big cancer cause of the death in third place in the world (totally 723000 death accounts for the 8.8% of general mortality rate).
The prognosis of gastric cancer largely depends on its disagreement.Some researches show that 5 years survival rates of stomach morning cancer almost to surpass 90% is crossed, and the survival rate of advanced gastric carcinoma is lower than 20%.So high risk suffer from cancer crowd early detection and rule with Examining is the most effective means for reducing incidence gastric cancer rate, improving survival.
Since the mistaken diagnosis of common white light endoscopic diagnosis gastric cancer (especially superficial flat and depressed lesion), rate of missed diagnosis are quite high, respectively Kind endoscopic diagnosis technology is come into being.But the superb operation skill not only needed using these endoscopic assistances, it is also necessary to can The economic support of sight.Therefore, be badly in need of researching and developing a kind of discovery, Diagnosis of Gastric morning cancer and precancerous lesion it is simple and easy to get, economical and practical simultaneously And safe and reliable diagnostic techniques.
Summary of the invention
Inventor, in order to reduce various problems brought by artificial endoscopic diagnosis, utilizes machine in long-term medical practice Device learning art by repeatedly developing, optimizes and training obtains the system that can be used for diagnosing gastric cancer repeatedly, is aided with system and tight The optical sieving of lattice and pretreatment further improve trained efficiency.Diagnostic system of the invention can accurately be known very much Cancerous lesion position in other pathological image (such as gastroscope picture and realtime graphic), it is special that discrimination even has been over internal medicine Family doctor.
The first aspect of the invention provides a kind of gastric cancer image identification system comprising:
A, data input module, for inputting the image comprising gastric cancer diseased region, described image is preferably endoscope figure Picture;
B, data preprocessing module, for receiving the image from data input module, and accurate frame selects the lesion of gastric cancer Position, the part in frame choosing is defined as positive sample, and the part outside frame choosing is defined as negative sample, and exports diseased region Coordinate information and/or lesion type information;It is preferred that the module also carries out desensitization process to image in advance before frame choosing, go Except sufferer personal information;
Preferably, the frame choosing can generate a rectangle frame or square-shaped frame comprising lesions position;The coordinate letter Breath is preferably the coordinate information of the point in the upper left corner and lower right corner of the rectangle frame or square-shaped frame;
It is also preferred that frame selects position to be determined by following methods: 2n scope doctors carry out frame choosing in a manner of " back-to-back ", i.e., 2n people is randomly divided into n group, 2 people/group, while all images are randomly divided into n parts, and is randomly assigned to each group doctor and carries out frame Choosing;When frame choosing after the completion of, compare every group of two doctors frame choosing as a result, and to frame between two doctors select the consistency of result into Row assessment, final determination block selects position, wherein natural number of the n between 1-100, for example, 1,2,3,4,5,6,7,8,9,10,20, 30,40,50,60,70,80,90 or 100;
It is further preferred that described as follows to the standard that frame selects the consistency of result to be assessed between two doctors:
For each lesion picture, the frame for comparing every group of two doctors selects the overlapping area of result, if every group two The 50% of the area that the union that the area (i.e. intersection) that doctor distinguishes the position lap of frame choosing is greater than the rwo is covered, then Think that the frame of two doctors selects judging result consistency good, and by the corresponding diagonal line coordinates of above-mentioned intersection, i.e., the upper left corner and The coordinate of the point in the lower right corner saves as the final positioning of target lesion;
If the 50% of the area that the union that the area (i.e. intersection) of lap is less than the rwo is covered, then it is assumed that two The frame of doctor selects judging result difference larger, and such lesion picture is individually picked out, and the 2n of work is selected by all participation frames Position doctor discusses the final position for determining target lesion jointly;
C, image recognition model construction module can be received through data preprocessing module treated image, for constructing And training image recognition model neural network based, the neural network is preferably convolutional neural networks;
D, lesion identification module is known for image to be checked to be input to the image recognition model after training, and based on image With the presence or absence of the position of lesion and/or lesion in the output result judgement image to be checked of other model.
In one embodiment, described image identification model building module includes feature extractor, candidate region generation Device and target marker, in which:
The feature extractor is used to carry out feature extraction to the image from data preprocessing module to obtain feature Figure, it is preferred that the feature extraction is carried out by convolution operation;
The candidate region generator is used to generate several candidate regions based on the characteristic pattern;
The target marker calculates the classification score of the candidate region, and the score indicates that the region belongs to the sun The probability of property sample and/or the negative sample;Target marker can propose to adjust to the bezel locations in each region simultaneously Value so that the bezel locations for each region are adjusted, and then accurately determines lesions position;Preferably, the classification score With loss function (Loss function) has been used in the training of adjusted value;
It is also preferred that using the gradient descent method based on mini-batch, i.e., being instructed to each when carrying out described trained Practice picture and generates the mini-batch comprising multiple positive and negative candidate regions;The then random sampling from every picture 256 candidate regions then calculate corresponding mini- until the ratio of positive candidate region and negative candidate region is close to 1:1 The loss function of batch;If the quantity of positive candidate region is less than 128 in a picture, go to fill out with feminine gender candidate region Mend this mini-batch;
It is further preferred that 0.001 is set by the learning rate of preceding 50000 mini-batch, 50000 by after The learning rate of mini-batch is set as 0.0001;Momentum term is preferably arranged to 0.9, and weight decaying is preferably arranged to 0.0005.
In another embodiment, wherein the feature extractor can be to the arbitrary dimension and/or resolution ratio of input Image carry out feature extraction, described image can be original image size and/or resolution ratio, be also possible to change size and/or point The image inputted after resolution obtains the characteristic pattern of multidimensional (such as 256 dimensions or 512 dimensions);
Specifically, the feature extractor includes X convolutional layer and Y sample level, wherein i-th (i is between 1-X) volume Lamination includes a QiIt is a having a size of m*m*piConvolution kernel, wherein m*m indicate convolution kernel length and wide pixel value, piEqual to upper The convolution nuclear volume Q of one convolutional layeri-1, in i-th of convolutional layer, convolution kernel is with step-length L to the data (example from upper level Such as original image, (i-1)-th convolutional layer or sample level) carry out convolution operation;Each sample level includes 1 mobile with step-length 2L, Size is the convolution kernel of 2L*2L, carries out convolution operation to the image of convolutional layer input;Wherein, it is carried out by feature extractor special After sign is extracted, the final characteristic pattern for obtaining Qx dimension;
Wherein X is between 1-20, for example, 1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19 Or 20;Y is between 1-10, such as 1,2,3,4,5,6,7,8,9 or 10;M between 2-10, such as 2,3,4,5,6,7,8,9 or 10;P between 1-1024, Q between 1-1024, the numerical value of p or Q respectively such as 1,2,3,4,5,6,7,8,9,10,11,12, 13,14,15,16,32,64,128,256,512 or 1024.
In another embodiment, wherein sliding window is arranged in the characteristic pattern in the candidate region generator, The size of sliding window is n × n, such as 3 × 3;Slide sliding window along characteristic pattern, simultaneously for every where sliding window There are corresponding relationships for corresponding position in one position, central point and original image, and in original image centered on the corresponding position It is middle to generate the k candidate regions with different scale and length-width ratio;Wherein, if k candidate region has x kind (such as 3 Kind) different scale and length-width ratio, then k=x2(such as k=9).
In another embodiment, the target marker includes middle layer again, classify layer and frame recurrence layer, wherein Middle layer is used to map sliding window operation and is formed by the data of candidate region, be multidimensional (such as 256 dimensions or 512 dimensions) to Amount;
Classification layer and frame return layer and connect respectively with middle layer, and classification layer is for determining that the object candidate area is prospect (i.e. positive sample) or background (i.e. negative sample), frame return the x coordinate and y seat that layer is used to generate candidate region central point Mark and the wide w and high h of candidate region.
The second aspect of the invention provides a kind of identification device of gastric cancer image, including is stored with diagnosing gastric cancer figure The storage unit of picture, image preprocessing program and trainable image recognition program, it is also preferable to include arithmetic elements and display Unit;
Described device, which can utilize the image recognition program of the image comprising gastric cancer lesion to be trained, (preferably has supervision Training), so that the image recognition program after training be enable to identify gastric cancer diseased region in image to be checked;
Preferably, the image to be checked is scope photo or real-time imaging.
In one embodiment, wherein described image preprocessor accurate frame choosing in the diagnosing gastric cancer image The diseased region of gastric cancer, the part in frame choosing is defined as positive sample, and the part outside frame choosing is defined as negative sample, and exports The location coordinate information and/or lesion type information of lesion;It is preferred that also carrying out desensitization process before frame choosing to image in advance, going Except sufferer personal information;
Preferably, the frame choosing can generate a rectangle frame or square-shaped frame comprising lesions position;The coordinate letter Breath is preferably the coordinate information of the point in the upper left corner and the lower right corner;
It is also preferred that frame selects position to be determined by following method: 2n scope doctors carry out frame choosing in a manner of " back-to-back ", i.e., 2n people is randomly divided into n group, 2 people/group, while all images are randomly divided into n parts, and is randomly assigned to each group doctor and carries out frame Choosing;When frame choosing after the completion of, compare every group of two doctors frame choosing as a result, and to frame between two doctors select the consistency of result into Row assessment, final determination block selects position, wherein natural number of the n between 1-100, for example, 1,2,3,4,5,6,7,8,9,10,20, 30,40,50,60,70,80,90 or 100;
It is further preferred that described as follows to the standard that frame selects the consistency of result to be assessed between two doctors:
For each lesion image, the frame for comparing every group of 2 doctors selects the overlapping area of result, if every group of two doctors The 50% of the area that the union that the area (i.e. intersection) that teacher distinguishes the position lap of frame choosing is greater than the rwo is covered, then recognize It selects judging result consistency good for the frame of 2 doctors, and the corresponding diagonal line coordinates of above-mentioned intersection is saved as into target lesion Final positioning;
If the 50% of the area that the union that the area (i.e. intersection) of lap is less than the rwo is covered, then it is assumed that 2 The frame of doctor selects judging result difference larger, and such lesion picture is individually picked out, and the 2n of work is selected by all participation frames Position doctor discusses the final position for determining target lesion jointly.
In another embodiment, described image recognizer is trainable image recognition journey neural network based Sequence, the neural network are preferably convolutional neural networks;Preferably, described image recognizer includes feature extractor, candidate Region generator and target marker, in which:
The feature extractor is used to carry out feature extraction to image to obtain characteristic pattern, it is preferred that the feature mentions It takes and is carried out by convolution operation;
The candidate region generator is used to generate several candidate regions based on the characteristic pattern;
The target marker calculates the classification score of the candidate region, and the score indicates that the region belongs to the sun The probability of property sample and/or the negative sample;Target marker can propose to adjust to the bezel locations in each region simultaneously Value, so that the bezel locations for each region are adjusted, to accurately determine lesions position;Preferably, the classification score With loss function (Loss function) has been used in the training of adjusted value;
In another embodiment, wherein when carrying out described trained, declined using the gradient based on mini-batch Method generates the mini-batch comprising multiple positive and negative candidate regions to each Zhang Xunlian picture.Then from every 256 candidate regions of random sampling are until the ratio of positive candidate region and negative candidate region is close to 1:1 in picture, then Calculate the loss function of corresponding mini-batch.If the quantity of positive candidate region is less than 128 in a picture, with yin Property candidate region is gone to fill up this mini-batch;
Preferably, 0.001 is set by the learning rate of preceding 50000 mini-batch, 50000 mini-batch by after Learning rate be set as 0.0001;Momentum term is preferably arranged to 0.9, and weight decaying is preferably arranged to 0.0005.
In another embodiment, wherein the feature extractor can be to the arbitrary dimension and/or resolution ratio of input Image carry out feature extraction, described image can be original image size and/or resolution ratio, be also possible to change size and/or point The image inputted after resolution obtains the characteristic pattern of multidimensional (such as 256 dimensions or 512 dimensions);
Specifically, the feature extractor includes X convolutional layer and Y sample level, wherein i-th (i is between 1-X) volume Lamination includes a QiIt is a having a size of m*m*piConvolution kernel, wherein m*m indicate convolution kernel length and wide pixel value, piEqual to upper The convolution nuclear volume Q of one convolutional layeri-1, in i-th of convolutional layer, convolution kernel is with step-length L to the data (example from upper level Such as original image, (i-1)-th convolutional layer or sample level) carry out convolution operation;Each sample level includes 1 mobile with step-length 2L, Size is the convolution kernel of 2L*2L, carries out convolution operation to the image of convolutional layer input;Wherein, it is carried out by feature extractor special After sign is extracted, the final characteristic pattern for obtaining Qx dimension;
Wherein X is between 1-20, for example, 1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19 Or 20;Y is between 1-10, such as 1,2,3,4,5,6,7,8,9 or 10;M between 2-10, such as 2,3,4,5,6,7,8,9 or 10;P between 1-1024, Q between 1-1024, the numerical value of p or Q respectively such as 1,2,3,4,5,6,7,8,9,10,11,12, 13,14,15,16,32,64,128,256,512 or 1024.
In another embodiment, wherein sliding window is arranged in the characteristic pattern in the candidate region generator, The size of sliding window is n × n, such as 3 × 3;Slide sliding window along characteristic pattern, simultaneously for every where sliding window There are corresponding relationships for corresponding position in one position, central point and original image, and in original image centered on the corresponding position It is middle to generate the k candidate regions with different scale and length-width ratio;Wherein, if k candidate region has x kind (such as 3 Kind) different scale and length-width ratio, then k=x2(such as k=9).
In another embodiment, the target marker includes middle layer again, classify layer and frame recurrence layer, wherein Middle layer is used to map sliding window operation and is formed by the data of candidate region, be multidimensional (such as 256 dimensions or 512 dimensions) to Amount;
Classification layer and frame return layer and connect respectively with middle layer, and classification layer is for determining that the object candidate area is prospect (i.e. positive sample) or background (i.e. negative sample), frame return the x coordinate and y seat that layer is used to generate candidate region central point Mark and the wide w and high h of candidate region.
The third aspect of the invention provides the system of first aspect of the present invention or the device of the second aspect in stomach Purposes in the prediction and diagnosis of cancer and/or gastric precancerous lesion.
The fourth aspect of the invention provides the system of first aspect of the present invention or the device of the second aspect in stomach Purposes in cancer image or gastric cancer image in the identification of diseased region.
The fifth aspect of the invention provides the system of first aspect of the present invention or the device of the second aspect in stomach Purposes in the real-time diagnosis of cancer and/or gastric precancerous lesion.
The sixth aspect of the invention provides the system of first aspect of the present invention or the device of the second aspect in stomach Purposes in cancer image or gastric cancer image in the real-time identification of diseased region.
By inventor it is long-term grope to find, due to gastric cancer diseased region there is own characteristics, i.e. diseased region not It is enough significant, it is not clear enough with perienchyma boundary, therefore the difficulty of image recognition model training compares conventional task (as known Not Sheng Huo in object) difficulty it is bigger, slightly will lead to training accidentally and be difficult to restrain so as to cause failure.And in the present invention In, inventor is improved its training method and (such as is accurately defined by frame choosing by image recognition model neural network based Target lesion position in training image, improves identification precision of image recognition model etc.), to obtain a kind of internal The identifying system (and/or device) of the intelligent and high-efficiency identification of gastric cancer lesion in mirror picture, discrimination are higher than common scope Doctor.Use the strengthened real-time diagnosis system of machine learning, additionally it is possible to carry out to Alimentary tract disease and its position and probability real When monitor and identification reduces misdiagnosis rate, provide for diagnosing gastric cancer so as to significant increase general doctor to the recall rate of gastric cancer Safe and reliable technology.
Detailed description of the invention
Fig. 1 includes the endoscopic images of gastric cancer lesions position
Fig. 2 frame selects process schematic
The lesions position for the gastric cancer that image identification system Fig. 3 of the invention is identified.
Specific embodiment
Unless otherwise stated, term used in the disclosure has generally containing for one skilled in the art's understanding Justice.Here is the meaning of some terms in this disclosure, if having inconsistent with other definition, is subject to defined below.
Definition
Term " gastric cancer ", refers to the malignant tumour derived from gastric epithelial cell, includes early carcinoma of stomach and advanced gastric carcinoma.
Term " module " refers to that the function set that can be realized certain effects, the module can be executed only by computer, It can also be completed together by manually performing, or by computer and manually.
Obtain lesion data
The key effect for obtaining lesion data step is to obtain the sample material for deep learning.
In one embodiment, the acquisition process can specifically include the step of acquisition and primary dcreening operation.
" acquisition " refers to that searching for acquisition in all scope databases according to the standard of " being diagnosed as gastric cancer " owns All endoscopic diagnosis images of patient with gastric cancer, such as it is diagnosed as figure all in the affiliated file of patient of " gastric cancer " Piece, the i.e. picture of certain patient all storages in entire endoscopic procedures, therefore be also possible that other than target site lesion Gastrocopy picture, such as the patient is diagnosed as benign ulcer, polyp etc., but further comprises food in its file under one's name The picture of each position storage in the checking processes such as pipe, stomach bottom, body of stomach, duodenum.
" primary dcreening operation " is the step of screening to the pathological image of the patients with gastric cancer collected, specifically can be by passing through Scope doctor abundant is tested to describe according to the related content in case " seen in endoscopy " in combination " pathological diagnosis " to carry out 's.Since the picture for deep learning network must be that quality is clear, feature is accurate, learning difficulty increasing otherwise will lead to Or recognition result inaccuracy.Therefore the module and/or step of lesion data primary dcreening operation can be there are clear gastric cancer lesions positions Picture picked out from a set of inspection picture.
Importantly, primary dcreening operation can be in " pathological diagnosis " to atrophy position in conjunction with histopathological findings after patient's biopsy Description is accurately positioned lesion, combines picture clarity, shooting angle, magnification level etc., select those clarity as far as possible Height, magnification level is moderate, can get a glimpse of the endoscopic image of lesion overall picture.
By primary dcreening operation, it can guarantee that the picture for inputting training set is the image comprising determining diseased region of high quality, To improve typing training data set feature accuracy, so that smart network preferably can therefrom conclude, summarize The characteristics of image of atrophic lesion out improves accuracy rate of diagnosis.
Lesion data pretreatment
The process that accurate frame selects the lesions position of gastric cancer is completed in the pretreatment, and the part in frame choosing is defined as sun Property sample, and the part outside frame choosing is defined as negative sample, and exports the location coordinate information and lesion type information of lesion.
In one embodiment, lesion data pretreatment is real by " image preprocessing program " institute in whole or in part Existing.
Term " image preprocessing program " refers to the frame choosing that can be realized objective area in image, to indicate target The program of area type and range.
In one embodiment, image preprocessing program can also carry out desensitization process to image, and removal sufferer is personal Information.
In one embodiment, image preprocessing program is one and is able to carry out using what computer programming language was write The software of aforementioned function.
In another embodiment, image preprocessing program is the software for being able to carry out frame and selecting function.
In a specific embodiment, executing frame selects the software of function that picture to be processed can be imported to software, and The picture is shown in operation interface, and implementing frame selection operation personnel (such as doctor) at this time only need to be at the quasi- target lesion position outlined Mouse is dragged along from (or other diagonal directions) direction of upper left to bottom right, to form the square for covering target lesion Shape frame or square-shaped frame, and generate from the background simultaneously and store the accurate coordinate in the rectangle frame upper left corner and the lower right corner with uniquely fixed Position.
In order to guarantee to pre-process the accuracy of (or frame choosing), the present invention, which is further enhanced, selects quality control to frame, this And method/system of the invention can obtain an important guarantee of bigger accuracy, concrete mode is as follows: selection 2n (such as 6,8,10 etc.) scope doctor carries out frame choosing in a manner of " back-to-back ", i.e. 2n people is randomly divided into n group, 2 people/group, Simultaneously the training image after all screenings is also divided into n parts at random, and is randomly assigned to each group doctor and carries out frame choosing;When frame selects After the completion, compare every group of 2 doctors frame choosing as a result, and select the consistency of result to assess frame between two doctors, most Whole determination block selects position.
In one embodiment, the evaluation criterion of consistency are as follows: for same lesion picture, compare every group 2 The frame of doctor selects result namely comparison to the overlapping area of rectangle frame determined by angular coordinate, if we provide the overlapping of two rectangle frames The 50% of the area that the union that partial area (i.e. intersection) is greater than the rwo is covered, then it is assumed that the frame of 2 doctors selects judgement As a result consistency is good, and the corresponding diagonal line coordinates of above-mentioned intersection is saved as the final positioning of target lesion.On the contrary, if The 50% of the area that the union that the area (i.e. intersection) of two rectangle frame laps is less than the rwo is covered, then it is assumed that 2 doctors The frame of teacher selects judging result difference larger, then such lesion picture will individually be picked out by software backstage, later period collection In select the doctor of work discuss the final position of determination target lesion jointly by all participation frames.
Image recognition model
Term " image recognition model " refers to the algorithm based on the building of the principle of machine learning and/or deep learning, can also To be referred to as " trainable image recognition model " or " image recognition program ".
In one embodiment, which is a kind of neural network, and the neural network is preferably convolutional Neural net Network;In another embodiment, the neural network is based on LeNet-5, RCNN, SPP, Fast-RCNN and/or Faster- The convolutional neural networks of RCNN framework;Wherein faster-RCNN can regard the combination of Fast-RCNN and RPN as, in a reality It applies in scheme, is based on faster-RCNN network.
Image recognition program includes at least following level: original image feature extraction layer, candidate region selected layer and target identification Layer, can training parameter by the adjustment of preset algorithm.
Term " original image feature extraction layer " is to refer to pass through mathematical computations to training image to what is inputted to multidimensional Degree extracts the level or level combination of original image information.The layer can actually indicate the combination of multiple and different functional layers.
In one embodiment, original image feature extraction layer can be based on ZF or VGG16 network.
Term " convolutional layer ", refers in original image feature extraction layer, is responsible for original input picture or passes through sample level Image information that treated carries out convolution operation, to extract the network layer of information.The convolution operation is indeed through one The convolution kernel (such as 3*3) of a particular size is slided on the image of input with certain step-length (such as 1 pixel), in convolution kernel The pixel on picture is multiplied with the respective weights of convolution kernel in mobile process, finally by all product additions obtain one it is defeated It realizes out.In image procossing, often image is expressed as the vector of pixel, therefore a secondary digital picture is considered as one The discrete function of a two-dimensional space, such as it is expressed as f (x, y), it is assumed that have for two-dimensional convolution handling function C (u, v), then can produce Raw output image g (x, y)=f (x, y) * C (u, v), may be implemented using convolution to image Fuzzy Processing and information extraction.
Term " training " refers to by inputting largely by the samples that manually mark, to trainable image recognition program into The self-regulated repeatedly of row parameter realizes the diseased region in identification gastric cancer image to realize expected purpose.
In one embodiment, the present invention is based on faster-rcnn networks, and are arrived in step s 4 using following end The training method at end:
(1) parameter of network (RPN) is generated using the model initialization object candidate area of the pre-training on ImageNet, And the network is finely adjusted;
(2) the model initialization Fast R-CNN network parameter for equally using the pre-training on ImageNet, followed by (1) the region proposal that RPN network extracts in is trained;
(3) the Fast R-CNN network of (2) is used to reinitialize RPN, fixed convolutional layer finely tunes RPN network, wherein only Cls and/or reg layers of RPN in fine tuning;
(4) convolutional layer for fixing Fast R-CNN in (2), uses the region proposal couple that RPN is extracted in (3) Fast R-CNN network is finely adjusted, wherein only finely tuning the full articulamentum of Fast R-CNN.
Term " candidate region selected layer ": refer to by algorithm realization select on the original image specific region for point Class identification and frame return level or level combination, it is similar with original image feature extraction layer, the layer can also indicate it is multiple not The combination of same layer.
Candidate region selected layer is directly connected to for original input layer in one embodiment.
In one embodiment, candidate region selected layer and the last layer of original image feature extraction layer are directly connected to.
In one embodiment, " candidate region selected layer " can be based on RPN.
Term " target identification layer " term " sample level " can sometimes be called pond layer, and operation is similar to convolutional layer, Only the convolution kernel of sample level is only to take maximum value, average value of corresponding position etc. (maximum pond, average pond).
Term " characteristic pattern ", is also feature map, refers to and carries out convolution to original image image by original image feature extraction layer The high-dimensional multichannel image of the small area obtained after operation leads to as an example, characteristic pattern can be 256 that scale is 51*39 Road image.
Term " sliding window " refers to the window of the small size (such as 2*2,3*3) generated on characteristic pattern, along characteristic pattern Each position it is mobile, although characteristic pattern size is also and less, since the data that characteristic pattern has already passed through multilayer are extracted (such as convolution), therefore the bigger visual field can be realized using lesser sliding window on characteristic pattern.
Term " candidate region ", is referred to as candidate window, object candidate area, reference box, bounding Box can also be replaced also in this context with anchor or anchor box.
In one embodiment, it is positioned first by sliding window to a position of characteristic pattern, it is raw for the position At the rectangular or square window of k different area different proportion, such as 9, and it is anchored to the center of the position, therefore also cry Anchor or anchor box, and the relationship based on the center of each sliding window and original image in characteristic pattern are done, is formed candidate Region, the candidate region substantially may be considered original corresponding to the sliding window (3*3) moved on the last layer convolutional layer Graph region range.
In one embodiment of the invention, k=9 includes the following steps: when generating candidate region
(1) 9 kinds of anchor box are generated first, in accordance with different area and length-width ratio, the anchor box not with characteristic pattern or The size of person's original input picture changes;
(2) for every input picture, according to image size calculate each sliding window corresponding to original image central point;
(3) mapping relations of sliding window position and original image position are established based on above-mentioned calculating.
Term " middle layer " refers to after forming object candidate area using sliding window, characteristic pattern is further mapped to one In the vector of multidimensional (such as 256 dimensions or 512 dimensions), this layer can be considered as to a new level, be referred to as in the present invention as centre Layer.Link sort layer and window return layer after middle layer.
Term " classification layer " (cls_score), exports a branch connecting with middle layer, which can export 2k A score respectively corresponds two scores of k object candidate area, one of them is prospect (i.e. positive sample) score, and one It is background (i.e. negative sample) score, this score may determine that the object candidate area is real target or background.Cause , for each sliding window position, layer of classifying output can belong to prospect (i.e. positive sample from high-dimensional (such as 256 dimensions) feature for this This) and background (i.e. negative sample) probability.
Specifically, in one embodiment, when candidate region and any (authentic specimen side ground-truth box Boundary, that is, boundary of the object for needing to identify in original image) IOU (hand over and compare) to be greater than 0.7 be to be considered the positive Sample or positive label, when the IOU of candidate region and any ground-truth box are less than 0.3, then it is assumed that be that background is (i.e. negative Property sample), to be assigned with class label to each anchor.Wherein IOU from mathematics containing above represent candidate region with The degree of overlapping of ground-truth box, calculation method are as follows:
IOU=(A ∩ B)/(A ∪ B)
Classification layer can export k+1 dimension group p, indicate the probability for belonging to k class and background.To each RoI (Region of Interesting discrete type probability distribution) is exported, p is then calculated by the full articulamentum of k+1 class using softmax.Mathematical table Up to as follows:
P=(p0, p1..., pk)
Term " window recurrence layer " (bbox_pred) exports another branch connecting with middle layer, simultaneously with classification layer Column.The layer can export on each position, and 9 anchor, which correspond to window, should translate the parameter of scaling.Respectively correspond k Object candidate area, each object candidate area have 4 bezel locations adjusted values, this 4 bezel locations adjusted values refer to mesh Mark the x in the upper left corner of candidate regionaCoordinate, yaThe high h of coordinate and object candidate areaaWith wide waAdjusted value.The work of the branch With being finely adjusted to object candidate area position, keep the final result position more accurate.
Window, which returns layer, can export the displacement of bounding box recurrence, export 4*K dimension group t, indicate to be belonging respectively to When k class, it should translate the parameter of scaling.Mathematical expression is as follows:
K indicates the index of classification,Refer to the translation relative to object proposal Scale invariant, Refer to the Gao Yukuan in log space relative to object proposal.
In one embodiment, the present invention is realized by loss function (Loss function) to classification layer and window Training while returning layer, the function be by classification loss (i.e. classification layer softmax loss) and Regression loss (i.e. L1loss) is by certain weight proportion composition.:
Calculate calibration result and prediction result that softmax loss needs candidate region to correspond to ground truth;It calculates Regression loss needs three group informations:
(1) predicting candidate regional center position coordinates x, y and width high w, h;
(2) each of 9 anchor point reference boxes in candidate region periphery center position coordinate xa, yaAnd width High wa, ha
(3) frame (ground truth) corresponding center position coordinate x*, y* and width high w*, h* are really demarcated.
It calculates regression loss and total Loss mode is as follows:
tx=(x-xa)/wa, ty=(y-ya)/ha,
tw=log (w/wa), th=log (h/ha),
Wherein, piThe probability of target is predicted as anchor.
There are two numerical value,Be negative label equal to 0,It is positive label equal to 1.
tiIndicate the vector set of 4 parametrization coordinates of the candidate region of prediction.
Indicate the coordinate vector of the corresponding ground truth bounding box of postive anchor.
In one embodiment, in the training of loss function, using the gradient descent method based on mini-batch, i.e., One mini-batch comprising multiple positive and negative candidate regions (anchor) is generated to each Zhang Xunlian picture.Then from 256 anchor of random sampling are until the ratio of positive anchor and feminine gender anchor are close to 1:1, then calculating in every picture The loss function (Loss function) of corresponding mini-batch.If the quantity of positive anchor is less than 128 in a picture It is a, then it goes to fill up this mini-batch with feminine gender anchor.
In a specific embodiment, set 0.001 for the learning rate of preceding 50000 mini-batch, will after The learning rate of 50000 mini-batch is set as 0.0001;Momentum term is preferably arranged to 0.9, and weight decaying is preferably arranged to 0.0005。
After above-mentioned training, by the scope picture of the target lesion for identification of the deep learning network after training.One In a embodiment, classification scoring is set to 0.85, i.e. deep learning network validation lesion probability is more than 85% lesion It can be labeled out, so that the picture is judged as the positive;On the contrary, if not detecting suspicious diseased region in a picture Domain, then this picture is just judged as feminine gender.
Embodiment
1. exempting informed consent statement:
(1) the scope figure that this research obtains in previous clinic diagnosis merely with Gastroenterology dept., Beijing Friendship Hospital endoscope center Piece and relevant clinical data carry out retrospective observational study, will not make to conditions of patients, treatment, prognosis even life security At any influence;
(2) all data collection tasks are individually completed by one people of principal investigator, and after the completion of image data acquisition, stood All pictures are carried out to erase personal information processing using special software, it is ensured that in subsequent doctor screening, frame choosing and artificial In intelligence programming expert typing training, debugging and test process, the leakage of patients' privacy information is not caused;
(3) in Gastroenterology dept.'s endoscope center electronic health record inquiry system, and not set " contact method " or " home address " etc. Entry can show, i.e. the contact details of the system not typing patient, therefore this research can not trace back to and be included in patient's signature and know Letter of consent.
2. pathological image acquires
Inclusion criteria:
(1) it is terminated in the receiving of Beijing Friendship Hospital's digestive endoscopy center from January 1st, 2013 on June 10th, 2017 The trouble of spectroscopy (including electronic gastroscope, electronic colonoscope, endoscopic ultrasonography, electron stain scope, magnifying endoscope and Endoscopy) Person;
(2) under mirror diagnose " gastric cancer " (including and do not distinguish early carcinoma of stomach and advanced gastric carcinoma) patient;
Exclusion criteria:
(1) malignant tumor of digestive tract involves that position is extensive or indefinite person;
(2) pancreatobiliary malignant tumour person is only suffered from;
(3) merge other systems malignant tumour person simultaneously;
(2) scope picture is unintelligible and/or the undesirable person of shooting angle.
3, experiment flow and result
(1) data acquire: being found out from Gastroenterology dept., Beijing Friendship Hospital endoscope center electronic medical record system by researcher Receive between on June 10,1 day to 2017 January in 2013 endoscopy (including electronic gastroscope, electronic colonoscope, ultrasound in Mirror, electron stain scope, magnifying endoscope and Endoscopy), and be diagnosed as " gastric cancer " under mirror and (including and do not distinguish early carcinoma of stomach And advanced gastric carcinoma) patient scope picture and relevant clinical data;
(2) it erases personal information: all pictures being carried out erasing personal information processing immediately after the completion of acquisition.
(3) picture screens: treated that picture is finished to all, has filtered out clear pathological examination and has been confirmed as stomach Scope picture corresponding to the case of cancer, and according to biopsy pathology position, finishing screen is selected in each case comprising target lesion The few picture of the clear of position, background interference amounts to 3774;
(4) construct test data set: test totally 100, picture includes " gastric cancer " (early carcinoma of stomach of pathological examination confirmation With advanced gastric carcinoma) 50, then " the Non-cancerous disease of the another random acquisition in the database stomach that there is pathological examination to confirm Change " (including gastric benign ulcer, polyp, mesenchymoma, lipoma, ectopic pancreas) scope picture 50 is opened.Concrete operations include:
50 are randomly selected from all gastric cancer pictures that step (3) filters out first;
" nonneoplastic lesion " for the stomach that another random acquisition in the database has pathological examination to confirm again
(including gastric benign ulcer, polyp, mesenchymoma, lipoma, ectopic pancreas) scope picture 50 is opened, and immediately to above-mentioned 50 pictures carry out erasing personal information processing;
(5) it constructs training dataset: from the gastric cancer picture that step (3) filters out, excluding random selection in step (4) and use In the picture of building test data set, remaining 3724 are used for deep learning network training, thus composing training data set;
(6) frame selects target lesion: 6 scope doctors are randomly divided into 3 groups in a manner of " back-to-back ", by 6 people, 2 people/group;Institute Training picture after having screening is divided into 3 parts at random, and is randomly assigned to each group doctor and carries out frame choosing.Lesion frame selects the reality of step It applies based on the software voluntarily write, the software can show this in operation interface after picture to be processed capable of being imported software Picture, doctor need to drag mouse along from the direction of upper left to bottom right at the quasi- target lesion position outlined at this time, to be formed One covers the rectangle frame of target lesion, and generates and store the accurate coordinate in the rectangle frame upper left corner and the lower right corner from the background simultaneously Uniquely to position.
After the completion of frame choosing, the frame choosing of every group of 2 doctors is compared as a result, comparing diagonal sit for same lesion picture The overlapping area of rectangle frame determined by marking, after test, if the area of two rectangle frame lap of final determination (is handed over Collection) it is greater than the 50% of the area that the rwo union is covered, then it is assumed that and the frame of 2 doctors selects judging result consistency good, and And the corresponding diagonal line coordinates of above-mentioned intersection is saved as into the final positioning of target lesion.If opposite two rectangle frame laps The 50% of the area that the union that area (i.e. intersection) is less than the rwo is covered, then it is assumed that the frame of 2 doctors selects judging result phase Difference is larger, then such lesion picture will individually be picked out by software backstage (or handmarking), later period concentration by All frames that participate in select the doctor of work to discuss the final position for determining target lesion jointly.
(7) typing training: the picture typing that the framed choosing of above-mentioned institute is completed is based in faster-rcnn convolutional neural networks It is trained, and tests two kinds of network structures of ZF and VGG16;Training is by the way of end-to-end;
Wherein ZF network connects layer and a softmax classification output layer, VGG16 network tool with 5 convolutional layers, 3 entirely Have 13 convolutional layers, 3 connect layer and softmax classification output layer entirely, under the frame of Faster-RCNN, ZF and VGG16 model is the basic CNN for extracting training image feature.
When training, using the gradient descent method based on mini-batch, i.e., generating one to each Zhang Xunlian picture includes The mini-batch of multiple positive and negative candidate regions (anchor).Then random sampling 256 from every picture Anchor then calculates the loss of corresponding mini-batch until the ratio of positive anchor and feminine gender anchor are close to 1:1 Function (Loss function).If the quantity of positive anchor is less than 128 in a picture, go to fill out with feminine gender anchor Mend this mini-batch.
0.001 is set by the learning rate of preceding 50000 mini-batch, by the study of rear 50000 mini-batch Rate is set as 0.0001;Momentum term is preferably arranged to 0.9, and weight decaying is preferably arranged to 0.0005.
The loss function (Loss Function) used in training is as follows:
In above formula, i represents the index of anchor in each batch, piRepresent whether anchor is target (Object) Probability;pi* be the true tag of the anchor: when anchor is that then label is 1 to Object, on the contrary then label is 0.tiIt is one 4 Dimensional vector respectively indicates the parametrization coordinate of bounding box, and ti* it then indicates in bounding box regression forecasting Bounding box parametrization coordinate label.
(8) test and result statistics: using test data set (including 50 gastric cancers and 50 stomaches " nonneoplastic lesion " Picture), artificial intelligence system, Gastroenterology dept. doctor of different years are tested respectively, compares, evaluate the two in terms of diagnosis The indexs such as sensibility, specificity, accuracy rate, consistency, and carry out statistical analysis.In test, by the depth after training Practise network for identification the scope picture of target lesion when classification scoring be set as 0.85, i.e. deep learning network validation lesion Probability is more than that 85% lesion just can be labeled out, so that the picture is judged as the positive;On the contrary, if not having in a picture Have and detect suspicious lesion region, then this picture is just judged as feminine gender.
As a result as follows:
Based on the platform of national digestive disease Clinical Research Center, share under gastric cancer scope in pathological changes diagnosis test, 89 The sensibility that name participates in doctor's totality is fluctuated in 48%~100% range, median 88%, average sensitivity 87%; In 10%~98% range, (its median 78%, average specificity is 74%), accuracy rate is then fluctuated in 51% for specificity fluctuation ~91% range (its median 82%, Average Accuracy 80%).And the identification sensibility of deep learning network model diagnosis It is 90%, specificity is 50%, accuracy rate 70%.Therefore in terms of the diagnosing gastric cancer based on gastroscope picture, artificial intelligence is quick Perception is higher than overall physician level, but specificity is relatively low relative to Median levels, and accuracy rate is also slightly less than the middle position of doctor Number is horizontal, it is contemplated that deep learning network model diagnostic model has splendid stability in identification, and difference is cured Teacher has greatly fluctuation and unstability in terms of specificity, accuracy rate, therefore still can using artificial intelligent recognition lesion Enough effective doctor's individual difference brings that excludes diagnose deviation, thus have a good application prospect.
Wherein, sensibility is also referred to as susceptibility (sensitivity, SEN), also known as true positive rate (true positive Rate, TPR), i.e., actual diseased is diagnosed the percentage that standard is correctly diagnosed again.
Specificity, also referred to as specificity (specificity, SPE), also known as true negative rate (true negative rate, TNR), the ability that Screen test determines non-patient is reflected.
Accuracy rate=the individual sum correctly identified/individual sum identified.
The foregoing is merely illustrative of the preferred embodiments of the present invention, is not intended to limit the scope of the present invention.It is all Any modification, equivalent replacement, improvement and so within the spirit and principles in the present invention, are all contained in protection scope of the present invention It is interior.

Claims (16)

1. a kind of gastric cancer image identification system comprising:
A, data input module, for inputting the image comprising gastric cancer diseased region, described image is preferably endoscopic images;
B, data preprocessing module, for receiving the image from data input module, and accurate frame selects the diseased region of gastric cancer, Part in frame choosing is defined as positive sample, and the part outside frame choosing is defined as negative sample, and exports the seat of diseased region Mark information and/or lesion type information;It is preferred that the module also carries out desensitization process, removal disease to image in advance before frame choosing Suffer from personal information;
Preferably, the frame choosing can generate a rectangle frame or square-shaped frame comprising diseased region;The coordinate information is excellent It is selected as the coordinate information in the upper left corner of the rectangle frame or square-shaped frame and the point in the lower right corner;
It is also preferred that the position of frame choosing is determined by following methods: 2n scope doctors carry out frame choosing in a manner of " back-to-back ", i.e., will 2n people is randomly divided into n group, 2 people/group, while all images are randomly divided into n parts, and is randomly assigned to each group doctor and carries out frame Choosing;When frame choosing after the completion of, compare every group of two doctors frame choosing as a result, and to frame between two doctors select the consistency of result into Row assessment, final determination block selects position, wherein natural number of the n between 1-100, for example, 1,2,3,4,5,6,7,8,9,10,20, 30,40,50,60,70,80,90 or 100;
It is further preferred that described as follows to the standard that frame selects the consistency of result to be assessed between two doctors:
For each lesion picture, the frame for comparing every group of two doctors selects the overlapping area of result, if every group of two doctors The 50% of the area that the union that the area (i.e. intersection) of the position lap of frame choosing is greater than the rwo respectively is covered, then it is assumed that The frame of two doctors selects judging result consistency good, and by the corresponding diagonal line coordinates of above-mentioned intersection, the i.e. upper left corner and bottom right The coordinate of the point at angle saves as the final positioning of target lesion;
If the 50% of the area that the union that the area (i.e. intersection) of lap is less than the rwo is covered, then it is assumed that two doctors Frame select judging result difference it is larger, such lesion picture is individually picked out, and selects the position 2n of work to cure by all participation frames Shi Gongtong discusses the final position for determining target lesion;
C, image recognition model construction module can be received through data preprocessing module treated image, for constructing and instructing Practice image recognition model neural network based, the neural network is preferably convolutional neural networks;
D, lesion identification module for image to be checked to be input to the image recognition model after training, and is based on image recognition mould With the presence or absence of the position of lesion and/or lesion in the output result judgement image to be checked of type.
2. system according to claim 1, it includes feature extractor, candidate region that described image identification model, which constructs module, Generator and target marker, in which:
The feature extractor is used to carry out feature extraction to the image from data preprocessing module to obtain characteristic pattern, excellent Choosing, the feature extraction is carried out by convolution operation;
The candidate region generator is used to generate several candidate regions based on the characteristic pattern;
The target marker calculates the classification score of the candidate region, and the score indicates that the region belongs to the positive sample The probability of this and/or the negative sample;Target marker can propose adjusted value to the bezel locations in each region simultaneously, from And it is adjusted for the bezel locations in each region, and then accurately determine lesions position;Preferably, the classification score and adjustment Loss function (Loss function) has been used in the training of value;
It is also preferred that using the gradient descent method based on mini-batch, i.e., scheming to each Zhang Xunlian when carrying out described trained Piece generates the mini-batch comprising multiple positive and negative candidate regions;The then random sampling 256 from every picture A candidate region then calculates corresponding mini- until the ratio of positive candidate region and negative candidate region is close to 1:1 The loss function of batch;If the quantity of positive candidate region is less than 128 in a picture, go to fill out with feminine gender candidate region Mend this mini-batch;
It is further preferred that 0.001 is set by the learning rate of preceding 50000 mini-batch, 50000 mini- by after The learning rate of batch is set as 0.0001;Momentum term is preferably arranged to 0.9, and weight decaying is preferably arranged to 0.0005.
3. system according to claim 2, wherein the feature extractor being capable of arbitrary dimension to input and/or resolution The image of rate carries out feature extraction, and described image can be original image size and/or resolution ratio, be also possible to change size and/or The image inputted after resolution ratio obtains the characteristic pattern of multidimensional (such as 256 dimensions or 512 dimensions);
Specifically, the feature extractor includes X convolutional layer and Y sample level, wherein i-th of (i is between 1-X) convolutional layer Include a QiIt is a having a size of m*m*piConvolution kernel, wherein m*m indicate convolution kernel length and wide pixel value, piEqual to upper one The convolution nuclear volume Q of convolutional layeri-1, in i-th of convolutional layer, convolution kernel is (such as former to the data from upper level with step-length L Figure, (i-1)-th convolutional layer or sample level) carry out convolution operation;Each sample level includes 1 mobile with step-length 2L, size For the convolution kernel of 2L*2L, convolution operation is carried out to the image of convolutional layer input;Wherein, feature is carried out by feature extractor to mention After taking, the final characteristic pattern for obtaining Qx dimension;
Wherein X is between 1-20, for example, 1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19 or 20; Y is between 1-10, such as 1,2,3,4,5,6,7,8,9 or 10;M is between 2-10, such as 2,3,4,5,6,7,8,9 or 10;P exists Between 1-1024, Q between 1-1024, the numerical value of p or Q respectively such as 1,2,3,4,5,6,7,8,9,10,11,12,13,14, 15,16,32,64,128,256,512 or 1024.
4. system according to claim 2 or 3, wherein sliding is arranged in the candidate region generator in the characteristic pattern Window, the size of sliding window are n × n, such as 3 × 3;Slide sliding window along characteristic pattern, simultaneously for sliding window institute Each position, there are corresponding relationships for the corresponding position in central point and original image, and centered on the corresponding position The k candidate regions with different scale and length-width ratio are generated in original image;Wherein, if k candidate region has x kind (such as 3 kinds) different scale and length-width ratio, then k=x2(such as k=9).
5. the system according to any one of claim 2-4, the target marker include middle layer again, classify layer and side Frame returns layer, and wherein middle layer is used to map sliding window operation and is formed by the data of candidate region, is multidimensional (such as 256 Dimension or 512 dimension) vector;
Classification layer and frame return layer connect respectively with middle layer, classification layer be used for determine the object candidate area be prospect (i.e. Positive sample) or background (i.e. negative sample), frame return layer be used to generate candidate region central point x coordinate and y-coordinate, And the wide w and high h of candidate region.
6. a kind of identification device of gastric cancer image, including be stored with diagnosing gastric cancer image, image preprocessing program and can train Image recognition program storage unit, it is also preferable to include arithmetic elements and display unit;
Described device, which can utilize the image recognition program of the image comprising gastric cancer lesion to be trained, (preferably has supervision to instruct Practice), so that the image recognition program after training be enable to identify gastric cancer diseased region in image to be checked;
Preferably, the image to be checked is scope photo or real-time imaging.
7. device according to claim 6, wherein described image preprocessor is smart in the diagnosing gastric cancer image True frame selects the diseased region of gastric cancer, and the part in frame choosing is defined as positive sample, and the part outside frame choosing is defined as negative sample, And export the location coordinate information and/or lesion type information of lesion;It is preferred that also being carried out at desensitization to image in advance before frame choosing Reason removes sufferer personal information;
Preferably, the frame choosing can generate a rectangle frame or square-shaped frame comprising lesions position;The coordinate information is excellent It is selected as the coordinate information of the point in the upper left corner and the lower right corner;
It is also preferred that frame selects position to be determined by following method: 2n scope doctors carry out frame choosing in a manner of " back-to-back ", i.e., by 2n People is randomly divided into n group, 2 people/group, while all images are randomly divided into n parts, and is randomly assigned to each group doctor and carries out frame choosing; After the completion of frame choosing, the frame for comparing every group of two doctors is selected as a result, and selecting the consistency of result to carry out frame between two doctors Assessment, final determination block selects position, wherein natural number of the n between 1-100, for example, 1,2,3,4,5,6,7,8,9,10,20, 30,40,50,60,70,80,90 or 100;
It is further preferred that described as follows to the standard that frame selects the consistency of result to be assessed between two doctors:
For each lesion image, the frame for comparing every group of 2 doctors selects the overlapping area of result, if every group of two doctors point The 50% of the area that the union that the area (i.e. intersection) of the position lap of other frame choosing is greater than the rwo is covered, then it is assumed that 2 The frame of position doctor selects judging result consistency good, and it is final that the corresponding diagonal line coordinates of above-mentioned intersection saved as target lesion Positioning;
If the 50% of the area that the union that the area (i.e. intersection) of lap is less than the rwo is covered, then it is assumed that 2 doctors Frame select judging result difference it is larger, such lesion picture is individually picked out, and selects the position 2n of work to cure by all participation frames Shi Gongtong discusses the final position for determining target lesion.
8. device according to claim 6 or 7, described image recognizer is trainable image neural network based Recognizer, the neural network are preferably convolutional neural networks;Preferably, described image recognizer includes feature extraction Device, candidate region generator and target marker, in which:
The feature extractor is used to carry out feature extraction to image to obtain characteristic pattern, it is preferred that the feature extraction is logical Cross convolution operation progress;
The candidate region generator is used to generate several candidate regions based on the characteristic pattern;
The target marker calculates the classification score of the candidate region, and the score indicates that the region belongs to the positive sample The probability of this and/or the negative sample;Target marker can propose adjusted value to the bezel locations in each region simultaneously, from And be adjusted for the bezel locations in each region, to accurately determine lesions position;Preferably, the classification score and adjustment Loss function (Loss function) has been used in the training of value.
9. the device according to any one of claim 6 to 8, wherein when carrying out described trained, using based on mini- The gradient descent method of batch generates the mini- comprising multiple positive and negative candidate regions to each Zhang Xunlian picture batch.Then ratio of 256 candidate regions of random sampling until positive candidate region and negative candidate region from every picture Example then calculates the loss function of corresponding mini-batch close to 1:1.If the quantity of positive candidate region is few in a picture In 128, then go to fill up this mini-batch with feminine gender candidate region;
Preferably, 0.001 is set by the learning rate of preceding 50000 mini-batch, by rear 50000 mini-batch Habit rate is set as 0.0001;Momentum term is preferably arranged to 0.9, and weight decaying is preferably arranged to 0.0005.
10. device according to claim 8 or claim 9, wherein the feature extractor can to the arbitrary dimension of input and/or The image of resolution ratio carries out feature extraction, and described image can be original image size and/or resolution ratio, is also possible to change size And/or the image inputted after resolution ratio, obtain the characteristic pattern of multidimensional (such as 256 dimensions or 512 dimensions);
Specifically, the feature extractor includes X convolutional layer and Y sample level, wherein i-th of (i is between 1-X) convolutional layer Include a QiIt is a having a size of m*m*piConvolution kernel, wherein m*m indicate convolution kernel length and wide pixel value, piEqual to upper one The convolution nuclear volume Q of convolutional layeri-1, in i-th of convolutional layer, convolution kernel is (such as former to the data from upper level with step-length L Figure, (i-1)-th convolutional layer or sample level) carry out convolution operation;Each sample level includes 1 mobile with step-length 2L, size For the convolution kernel of 2L*2L, convolution operation is carried out to the image of convolutional layer input;Wherein, feature is carried out by feature extractor to mention After taking, the final characteristic pattern for obtaining Qx dimension;
Wherein X is between 1-20, for example, 1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19 or 20; Y is between 1-10, such as 1,2,3,4,5,6,7,8,9 or 10;M is between 2-10, such as 2,3,4,5,6,7,8,9 or 10;P exists Between 1-1024, Q between 1-1024, the numerical value of p or Q respectively such as 1,2,3,4,5,6,7,8,9,10,11,12,13,14, 15,16,32,64,128,256,512 or 1024.
11. the device according to any one of claim 8 to 10, wherein the candidate region generator is in the characteristic pattern Middle setting sliding window, the size of sliding window are n × n, such as 3 × 3;Slide sliding window along characteristic pattern, simultaneously for There are corresponding relationships for corresponding position in each position where sliding window, central point and original image, and with described corresponding The k candidate regions with different scale and length-width ratio are generated centered on position in original image;Wherein, if k candidate regions Domain has x kind (such as 3 kinds) different scale and length-width ratio, then k=x2(such as k=9).
12. the device according to any one of claim 8 to 11, the target marker includes middle layer again, layer of classifying Return layer with frame, wherein middle layer is used to map sliding window operation and is formed by the data of candidate region, be a multidimensional (such as 256 dimension or 512 dimension) vector;
Classification layer and frame return layer connect respectively with middle layer, classification layer be used for determine the object candidate area be prospect (i.e. Positive sample) or background (i.e. negative sample), frame return layer be used to generate candidate region central point x coordinate and y-coordinate, And the wide w and high h of candidate region.
13. system according to any one of claim 1 to 5 or the described in any item devices of claim 6 to 12 are in stomach Purposes in the prediction and diagnosis of cancer and/or gastric precancerous lesion.
14. system according to any one of claim 1 to 5 or the described in any item devices of claim 6 to 12 are in stomach Purposes in cancer image or gastric cancer image in the identification of diseased region.
15. system according to any one of claim 1 to 5 or the described in any item devices of claim 6 to 12 are in stomach Purposes in the real-time diagnosis of cancer and/or gastric precancerous lesion.
16. system according to any one of claim 1 to 5 or the described in any item devices of claim 6 to 12 are in stomach Purposes in cancer image or gastric cancer image in the real-time identification of diseased region.
CN201811371788.1A 2018-11-15 2018-11-15 A kind of gastric cancer image identification system, device and its application Pending CN109671053A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201811371788.1A CN109671053A (en) 2018-11-15 2018-11-15 A kind of gastric cancer image identification system, device and its application
CN202210461368.2A CN115345819A (en) 2018-11-15 2018-11-15 Gastric cancer image recognition system, device and application thereof

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811371788.1A CN109671053A (en) 2018-11-15 2018-11-15 A kind of gastric cancer image identification system, device and its application

Related Child Applications (1)

Application Number Title Priority Date Filing Date
CN202210461368.2A Division CN115345819A (en) 2018-11-15 2018-11-15 Gastric cancer image recognition system, device and application thereof

Publications (1)

Publication Number Publication Date
CN109671053A true CN109671053A (en) 2019-04-23

Family

ID=66141781

Family Applications (2)

Application Number Title Priority Date Filing Date
CN202210461368.2A Pending CN115345819A (en) 2018-11-15 2018-11-15 Gastric cancer image recognition system, device and application thereof
CN201811371788.1A Pending CN109671053A (en) 2018-11-15 2018-11-15 A kind of gastric cancer image identification system, device and its application

Family Applications Before (1)

Application Number Title Priority Date Filing Date
CN202210461368.2A Pending CN115345819A (en) 2018-11-15 2018-11-15 Gastric cancer image recognition system, device and application thereof

Country Status (1)

Country Link
CN (2) CN115345819A (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110176002A (en) * 2019-06-05 2019-08-27 深圳大学 A kind of the lesion detection method and terminal device of radioscopic image
CN110490262A (en) * 2019-08-22 2019-11-22 京东方科技集团股份有限公司 Image processing model generation method, image processing method, device and electronic equipment
CN110866897A (en) * 2019-10-30 2020-03-06 上海联影智能医疗科技有限公司 Image detection method and computer readable storage medium
CN110880177A (en) * 2019-11-26 2020-03-13 北京推想科技有限公司 Image identification method and device
CN110974306A (en) * 2019-12-17 2020-04-10 山东大学齐鲁医院 System for discernment and location pancreas neuroendocrine tumour under ultrasonic endoscope
WO2020215807A1 (en) * 2019-04-25 2020-10-29 天津御锦人工智能医疗科技有限公司 Deep-learning-based method for improving colonoscope adenomatous polyp detection rate
CN112568864A (en) * 2020-12-03 2021-03-30 牡丹江医学院 Uterine cavity operation monitoring system
CN113642679A (en) * 2021-10-13 2021-11-12 山东凤和凰城市科技有限公司 Multi-type data identification method
CN114463246A (en) * 2020-11-06 2022-05-10 广达电脑股份有限公司 Circle selection system and circle selection method
CN115517682A (en) * 2022-11-25 2022-12-27 四川大学华西医院 Cognitive dysfunction prediction system based on gastrointestinal electric signals and construction method
TWI825643B (en) * 2022-03-30 2023-12-11 緯創資通股份有限公司 Medical auxiliary information generation method and medical auxiliary information generation system

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017101142A1 (en) * 2015-12-17 2017-06-22 安宁 Medical image labelling method and system
CN107368859A (en) * 2017-07-18 2017-11-21 北京华信佳音医疗科技发展有限责任公司 Training method, verification method and the lesion pattern recognition device of lesion identification model
CN107563123A (en) * 2017-09-27 2018-01-09 百度在线网络技术(北京)有限公司 Method and apparatus for marking medical image
CN108665456A (en) * 2018-05-15 2018-10-16 广州尚医网信息技术有限公司 The method and system that breast ultrasound focal area based on artificial intelligence marks in real time
CN108734694A (en) * 2018-04-09 2018-11-02 华南农业大学 Thyroid tumors ultrasonoscopy automatic identifying method based on faster r-cnn

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017101142A1 (en) * 2015-12-17 2017-06-22 安宁 Medical image labelling method and system
CN107368859A (en) * 2017-07-18 2017-11-21 北京华信佳音医疗科技发展有限责任公司 Training method, verification method and the lesion pattern recognition device of lesion identification model
CN107563123A (en) * 2017-09-27 2018-01-09 百度在线网络技术(北京)有限公司 Method and apparatus for marking medical image
CN108734694A (en) * 2018-04-09 2018-11-02 华南农业大学 Thyroid tumors ultrasonoscopy automatic identifying method based on faster r-cnn
CN108665456A (en) * 2018-05-15 2018-10-16 广州尚医网信息技术有限公司 The method and system that breast ultrasound focal area based on artificial intelligence marks in real time

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020215807A1 (en) * 2019-04-25 2020-10-29 天津御锦人工智能医疗科技有限公司 Deep-learning-based method for improving colonoscope adenomatous polyp detection rate
CN110176002A (en) * 2019-06-05 2019-08-27 深圳大学 A kind of the lesion detection method and terminal device of radioscopic image
CN110176002B (en) * 2019-06-05 2022-04-01 深圳大学 Focus detection method of X-ray image and terminal device
CN110490262B (en) * 2019-08-22 2022-06-03 京东方科技集团股份有限公司 Image processing model generation method, image processing device and electronic equipment
CN110490262A (en) * 2019-08-22 2019-11-22 京东方科技集团股份有限公司 Image processing model generation method, image processing method, device and electronic equipment
US11887303B2 (en) 2019-08-22 2024-01-30 Beijing Boe Technology Development Co., Ltd. Image processing model generation method, image processing method and device, and electronic device
CN110866897A (en) * 2019-10-30 2020-03-06 上海联影智能医疗科技有限公司 Image detection method and computer readable storage medium
CN110880177A (en) * 2019-11-26 2020-03-13 北京推想科技有限公司 Image identification method and device
CN110974306A (en) * 2019-12-17 2020-04-10 山东大学齐鲁医院 System for discernment and location pancreas neuroendocrine tumour under ultrasonic endoscope
US12014813B2 (en) 2020-11-06 2024-06-18 Quanta Computer Inc. Contouring system
CN114463246A (en) * 2020-11-06 2022-05-10 广达电脑股份有限公司 Circle selection system and circle selection method
CN112568864B (en) * 2020-12-03 2022-03-18 牡丹江医学院 Uterine cavity operation monitoring system
CN112568864A (en) * 2020-12-03 2021-03-30 牡丹江医学院 Uterine cavity operation monitoring system
CN113642679A (en) * 2021-10-13 2021-11-12 山东凤和凰城市科技有限公司 Multi-type data identification method
TWI825643B (en) * 2022-03-30 2023-12-11 緯創資通股份有限公司 Medical auxiliary information generation method and medical auxiliary information generation system
CN115517682A (en) * 2022-11-25 2022-12-27 四川大学华西医院 Cognitive dysfunction prediction system based on gastrointestinal electric signals and construction method
CN115517682B (en) * 2022-11-25 2023-01-31 四川大学华西医院 Cognitive dysfunction prediction system based on gastrointestinal electric signals and construction method

Also Published As

Publication number Publication date
CN115345819A (en) 2022-11-15

Similar Documents

Publication Publication Date Title
CN109584218A (en) A kind of construction method of gastric cancer image recognition model and its application
CN109671053A (en) A kind of gastric cancer image identification system, device and its application
CN109544526A (en) A kind of atrophic gastritis image identification system, device and method
CN109523535A (en) A kind of preprocess method of lesion image
Li et al. A comprehensive review of computer-aided whole-slide image analysis: from datasets to feature extraction, segmentation, classification and detection approaches
Silva-Rodríguez et al. Going deeper through the Gleason scoring scale: An automatic end-to-end system for histology prostate grading and cribriform pattern detection
Adegun et al. Deep learning-based system for automatic melanoma detection
Zhou et al. A comprehensive review for breast histopathology image analysis using classical and deep neural networks
US11257209B2 (en) Cancer risk stratification based on histopathological tissue slide analysis
JP6000345B2 (en) Analysis method of biological samples by spectral imaging
CN109670530A (en) A kind of construction method of atrophic gastritis image recognition model and its application
Du et al. Identification of COPD from multi-view snapshots of 3D lung airway tree via deep CNN
Turkmen et al. Classification of laryngeal disorders based on shape and vascular defects of vocal folds
Fan et al. Effect of image noise on the classification of skin lesions using deep convolutional neural networks
Hu et al. A state-of-the-art survey of artificial neural networks for whole-slide image analysis: from popular convolutional neural networks to potential visual transformers
CN106096654A (en) A kind of cell atypia automatic grading method tactful based on degree of depth study and combination
CN110097974A (en) A kind of nasopharyngeal carcinoma far-end transfer forecasting system based on deep learning algorithm
Wang et al. Study on automatic detection and classification of breast nodule using deep convolutional neural network system
CN101714153A (en) Visual perception based interactive mammography image searth method
CN114600155A (en) Weakly supervised multitask learning for cell detection and segmentation
CN112508884A (en) Comprehensive detection device and method for cancerous region
Lee et al. Model architecture and tile size selection for convolutional neural network training for non-small cell lung cancer detection on whole slide images
WO2023249552A1 (en) System and method for classification of basal cell carcinoma based on confocal microscopy
CN115861598A (en) Polyp detection method based on convolutional neural network
Ahmad et al. Colorectal cancer image classification using image pre-processing and multilayer Perceptron

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20190423