CN111476292B - Small sample element learning training method for medical image classification processing artificial intelligence - Google Patents

Small sample element learning training method for medical image classification processing artificial intelligence Download PDF

Info

Publication number
CN111476292B
CN111476292B CN202010262936.7A CN202010262936A CN111476292B CN 111476292 B CN111476292 B CN 111476292B CN 202010262936 A CN202010262936 A CN 202010262936A CN 111476292 B CN111476292 B CN 111476292B
Authority
CN
China
Prior art keywords
layer
learning
sample
scale
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010262936.7A
Other languages
Chinese (zh)
Other versions
CN111476292A (en
Inventor
李功杰
马潞娜
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Quanjingdekang Medical Imaging Diagnosis Center Co ltd
Original Assignee
Beijing Quanjingdekang Medical Imaging Diagnosis Center Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Quanjingdekang Medical Imaging Diagnosis Center Co ltd filed Critical Beijing Quanjingdekang Medical Imaging Diagnosis Center Co ltd
Priority to CN202010262936.7A priority Critical patent/CN111476292B/en
Publication of CN111476292A publication Critical patent/CN111476292A/en
Application granted granted Critical
Publication of CN111476292B publication Critical patent/CN111476292B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • G06V10/751Comparing pixel values or logical combinations thereof, or feature values having positional relevance, e.g. template matching
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2413Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
    • G06F18/24133Distances to prototypes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/40Scaling the whole image or part thereof
    • G06T3/4038Scaling the whole image or part thereof for image mosaicing, i.e. plane images composed of plane sub-images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Abstract

The small sample element learning training method for medical image classification processing artificial intelligence is characterized by constructing three element learners including a multi-scale CNN feature extractor, a measurement learner and a classification discriminator and designing measurement standards on the element learners at the same time; on each task of the training set, a target set is learned through support set distance measurement, a measurement standard is finally obtained through learning, and then for a new task of the test set, the target set can be rapidly and correctly classified only by means of a small number of samples of the support set; the practicability and reliability of artificial intelligent automatic detection in the field of difficult and complicated diseases are improved by adopting meta-learning, the defect that the single disease data amount is small and the disease is dispersed is overcome, the classification and sample-less learning training of a medical image processing intelligent system is completed, the accuracy of filtration is obviously improved, the accuracy of medical image classification is effectively enhanced, the production efficiency is greatly improved, and the upgrading of an intelligent diagnosis deep learning technology in the medical industry is facilitated.

Description

Small sample element learning training method for medical image classification processing artificial intelligence
Technical Field
The invention relates to a structure improvement technology of equipment, belonging to the technical field of machine learning, pattern recognition and medical image processing of G06T7/11 in IPC classification, or G06K9/00 for reading or recognizing printed or written characters or for recognizing graphs, in particular to a small sample meta-learning training method of artificial intelligence for medical image classification processing.
Background
At present, medical images including X-ray imaging, CT imaging, magnetic resonance imaging, ultrasonic imaging, nuclear medicine imaging and the like can enable doctors to know changes of internal forms, functions, metabolism and the like of the bodies of patients besides contact and anatomy, and have important effect on diagnosing causes and conditions of the patients. Medical imaging plays an extremely important role in medical clinical diagnosis, and modern medicine is not away from medical imaging technology. The medical image auxiliary diagnosis system developed by science development, Tencent development, Ali health and the like can assist doctor diagnosis in specific medical fields such as lung CT image diagnosis, early esophageal cancer screening and the like by means of high-quality high-resolution images, a computer can automatically cut medical image images and classify the images by learning a large amount of images and diagnosis data, suspicious tissues are amplified and observed for multiple times in different scale scales, important information is extracted, and finally suggestions of whether lesions exist are output to assist image doctors to make decisions, so that the diagnosis efficiency is greatly improved, meanwhile, the phenomena of missed diagnosis and misdiagnosis can be effectively reduced, and the future medical development is greatly promoted.
With the development of internet technology and the improvement of informatization level, the image data volume also shows explosive growth. In a huge image database, how a computer efficiently picks up image data poses certain challenges to image classification technology. In recent years, deep learning is rapidly developed, and by means of higher accuracy and recognition efficiency of the deep learning, image classification technology by means of the deep learning gradually replaces manual labeling features to classify images, but the deep learning often has the defects that parameters are difficult to adjust in the training process, the requirement of training samples is large, and the training time is too long.
In order to solve the problems, it is meaningful to research how to efficiently utilize deep learning to perform image recognition under the condition of a small number of samples, and the adaptability of a deep learning model under different sample conditions is enhanced.
One key feature of human intelligence is versatility (versatility): the ability to accomplish different tasks. The current AI systems are good at mastering individual skills such as go, Jeopardy games and even stunt flight of helicopters. However, letting AI systems do something that seems simple, it is rather difficult for them to do. The program that takes Jeopardy champion cannot talk, and a professional helicopter stunt flight controller cannot navigate in a completely new simple environment, such as locating a fire scene, flying past, and then extinguishing it. However, humans can flexibly cope with and adapt spontaneously in the face of a variety of new situations. Humans are expected to achieve such versatility for artificial intelligence.
Since the development of deep learning, various intelligent devices are rapidly developed, but tens of thousands or even millions of data are required for intelligent programs to train, and for example, in the case of the recently named AlphaGo, the chess playing is a computer, but the subsequent stage of the chess needs huge data to support training to achieve the effect. It is also a short board for deep learning, and the training data needs too much, but sometimes some training tasks cannot find so much training data. In this case, if the deep learner is trained with less training data, an over-fitting phenomenon occurs, which is not beneficial to application, and therefore, for the task with less samples, a learner is required to successfully apply a small amount of data to perform good training to achieve the functions required by the learner.
It is reported that the recent technology developed in some laboratories, model-independent meta-learning (model-independent learning), was tried, and the results were unknown. For a person to get a new item first, you can quickly learn and become familiar with it by exploring it, and this learning ability is not available with current machines. Therefore, people think that if the machine can also have the learning capability, the machine can quickly learn when the sample size is small, which is called meta learning (meta learning). For example, a small sample learning article, "Human-level content learning through basic programming index", was published in the Science journal cover 2015, which proposes a Bayesian Programming Learning (BPL) framework to implement One-shot learning, and data sets adopt handwritten characters from different world alphabets. It was found that the task of small sample classification can be well achieved through meta-learning.
In recent years, the learning and classification of small samples are rapidly developed, and in the face of a plurality of classification tasks, task requirements can be met by training one model. The meta-learning method is more, and for the greatest applicability, the mechanism of the meta-learning is the universality of tasks, namely different tasks are faced without constructing different models, and the same set of learning algorithm can be used for solving various different tasks. The learnable parameter theta of one model is defined, and the corresponding task can be solved by changing the value of the parameter theta in different tasks. The value of the parameter theta can be learned through the meta-learner, when different tasks are faced, the theta value is continuously updated through a gradient descending method according to a loss function, the model is continuously close to the model capable of solving the tasks, and when the theta value is finally converged, the meta-learner learns a better parameter theta to enable the model to adaptively solve the corresponding tasks. The algorithm does not build or introduce other parameters for the learner, and the strategy for training the learner uses known optimization processes such as gradient descent.
Aiming at a plurality of problems of small sample identification, such as high possibility of overfitting, poor model generalization and the like, a small sample classification identification model based on deep learning is provided through experimental research, and the optimization direction is mainly based on the following two aspects, namely the enhancement of the number of image samples and the optimization of the identification model. When sample number enhancement is carried out, a technology of combining a generative model and image preprocessing is proposed. Firstly, a full-connection generating model is used for sample enhancement, and a convolutional neural network is used for replacing a full-connection neural network to perform image training aiming at the problem of excessive parameters of the full-connection neural network. Because the generated sample image has randomness, the sample is generated by using the condition generating model, and the generated sample set contains labels and can be well applied to subsequent supervised classification learning. Aiming at the problem that a sample generated in a condition generation model is fuzzy, an image edge detection technology based on wavelet transformation and self-adaptive mathematical morphology is provided. The method can well overcome the problem of edge blurring, and in addition, because the generated samples often have noise, the optimized denoising model can well remove the noise and simultaneously reserve the edge details of the image based on the image denoising technology combining empirical mode decomposition and sparse representation. Therefore, the image sample number enhancement by the fusion model can well enlarge the samples for the classification model of the next part. In the small sample classification model, the model classification is carried out by using the transfer learning, and the transfer learning can always achieve higher recognition efficiency in a small training sample set. And combining the generation model with the transfer learning, and carrying out small sample image identification by using the fusion model. The transfer learning is trained by adopting an input-V3 model, compared with the classification training of a convolutional neural network, the generalization capability of the model can be improved by adding the transfer learning, the training efficiency of the model is also greatly improved, and when a sample is missing, the accuracy of identification is also greatly improved because the classification model based on the transfer learning only needs to be retrained. The method is combined with a sample enhancement technology, when a sample is missing, model fusion and classification are carried out by adding a mode of combining a simulated sample and transfer learning, a standard data set and a shot leaf image are used for experimental comparison, and in the aspect of accuracy, compared with the mode of simply using transfer learning and a convolutional neural network, the method is improved to some extent, and has good reference significance for deep learning classification in an experimental environment with insufficient sample amount.
In the medical industry, the greatest contribution of computer functions is to help doctors to filter a large number of medical images, so that the use of diagnosis is at most an auxiliary diagnostic tool, and machine learning reaches the time of deep learning, and tasks which are not possible are overcome by deep learning algorithms.
The related patent documents are less published.
Nanjing university and Jiangsu Wanweieisi network intelligent industry innovation center limited company 201910374698.6 disclose a point interactive medical image segmentation method based on a deep neural network, and a point interactive deep learning segmentation algorithm is provided specially for the problem of renal tumor segmentation in medical images. The algorithm consists of a point interaction preprocessing module, a bidirectional ConvRNN unit and a core depth segmentation network. The algorithm starts from a tumor center position provided by an expert, 16 image blocks with the size of 32 multiplied by 32 are densely collected from inside to outside according to 4 pixel step lengths in 16 directions with uniform intervals to form an image block sequence, a deep segmentation network with sequence learning is used for learning the change trend of the inside and the outside of a target, the edge of the target is determined, and the renal tumor is segmented. The method can overcome the influences of low contrast of medical images, variable target positions and fuzzy target edges, and is suitable for organ segmentation and tumor segmentation tasks. The technology has the following characteristics: 1) the interaction mode is simple and convenient; 2) the Sequential Patch Learning concept is put forward, the long-range semantic relation is captured by using sequence image blocks, and a large receptive field can be obtained even for a shallow network; 3) a brand-new ConvRNN unit is provided, the internal and external change trend of a target is learned, the interpretability is strong, the actual working mode of a doctor is met, and the final model is high in precision and strong in applicability.
Nanjing aerospace university in China patent application 201610654316.1 discloses a color image significance detection method based on background and foreground information, which comprises the following steps: performing over-segmentation processing on an input color image to obtain a series of super pixel blocks; selecting background seeds, and obtaining roughness significance by comparing the characteristics of each superpixel block and the background seeds; defining the background weight of each super pixel block based on the characteristic distribution of the background seeds, and improving the rough significance through the background weight to obtain the significance based on the background information; segmenting the saliency map based on the background information formed in the previous step, selecting a compact foreground region from all segmentation results, extracting foreground region characteristics, and obtaining the saliency based on the foreground information through characteristic comparison; and integrating the significance degrees based on the background and foreground information obtained in the first two steps, and performing smoothing operation to obtain the optimized significance degrees of all the superpixel blocks. The foreground object in the image can be highlighted more consistently, and the suppression effect on the background noise in the image is good.
Microsoft licensing, llc, in chinese patent application 201510362657.7 proposes a method involving a machine vision processing system. Techniques and constructs may determine an albedo map and a shadowed map from a digital image. The albedo and shadowed map may be determined based at least in part on a color difference threshold. A color-shaded map can be determined based at least in part on the albedo map, and the illumination factor can be determined based on the color-shaded map. The digital image may be adjusted based at least in part on the illumination coefficient. In some examples, respective shaded maps for respective color channels of the digital image may be generated. The color-shaded map may be generated based at least in part on the shaded map. In some examples, multiple regions of a digital image may be determined, and proximity relationships between the various regions may be determined. The albedo shaded map may be determined based at least in part on the proximity relationship.
Tianjin university discloses an image classification method for two-way channel attention element learning in Chinese patent application 201910770542.X, wherein the relationship between a support set and a query set in the element learning is learned through an attention model mechanism, so that the support set and the query set pay attention to each other, and the relationship between the support set and the query set is mined on the assumption that when the visual features of an image sample in the support set are mapped to category features, the attention of the sample in the support set to the significant region of the sample in the support set is considered, and the attention of the sample in the support set to the significant region of the sample in the query set is also considered, so that the attention of a network to the significant and detailed region of the image features is improved, the convergence speed is accelerated, and the performance of the image classification method based on element learning and few samples is further improved.
The Chinese patent application 201910143448.1 of Zhongshan university proposes a method for classifying small and zero sample images based on metric learning and meta learning, relating to the field of computer vision recognition and transfer learning, and comprising the following steps: constructing a training data set and a target task data set; selecting a support set and a test set from a training data set; respectively inputting the samples of the test set and the support set into a feature extraction network to obtain feature vectors; sequentially inputting the feature vectors of the test set and the support set into a feature attention module and a distance measurement module, calculating the class similarity of the test set sample and the support set sample, and updating the parameters of each module by using a loss function; repeating the steps until the parameters of the network of each module are converged, and finishing the training of each module; and sequentially passing the picture to be tested and the training picture in the target task data set through the feature extraction network, the feature attention module and the distance measurement module, and outputting a class label with the highest class similarity with the test set, namely the classification result of the picture to be tested.
Haqing encyclopedia of technology, Inc. in Chinese patent application 201910684593.0, proposes an ensemble learning based image classification system comprising a plurality of Cellular Neural Network (CNN) based Integrated Circuits (ICs) operatively coupled together as a set of basic learners for the image classification task. Each CNN-based IC is configured with at least one distinct deep learning model in the form of filter coefficients. The ensemble learning based image classification system further includes: a controller configured as the integrated meta learner; and a memory-based data buffer for holding various data used by the controller and the CNN-based IC in the integration. The various data may include input image data to be classified. The various data may also include extracted feature vectors or image classification outputs from the set of base learners. The extracted feature vectors or image classification outputs are then used by the meta learner to further perform the image classification task.
According to the application of extensive literature in medical image processing by deep learning, most of the work is based on 2D images, in fact, in medical images, both CT and MRI are the result of 2D transformation of 3D data, and 3D reconstruction and the like are also a very important large class of algorithms in medical image processing, however, the existing 3D-based algorithm is time-consuming and not much improved compared with the 2D-based algorithm, the more traditional algorithm regards CNN as an operator for feature extraction to obtain the description features of the image, and the latest methods directly use the CNN result as the final output result, which involves the limitation of the black box, has poor interpretability, and is more suitable for feature interpretation, the interpretability is better because of the more rule classes that can be added in some of the subsequent post-processing.
The artificial intelligence technology is used for assisting in the field of diagnosis in medical images and mainly depends on massive data for training. But most medical image data originates from hospitals. And the third-level hospital has most of image data, but in practical application, the data volume of a single disease type is insufficient, the data is dispersed, the disease type is dispersed, so that the artificial intelligent medical auxiliary diagnosis system does not have enough training data for training, and the accuracy rate cannot be guaranteed. Although the purpose of small sample learning and meta learning is to learn quickly, and the amount of training samples adopted in all researches in recent years is small, a large sample is still required for supporting the autonomous learning of a machine.
Disclosure of Invention
The invention aims to provide a small sample meta-learning training method for artificial intelligence of medical image classification processing, which solves the problem of medical small samples aiming at the phenomenon of less data in the medical field, realizes small sample classification by adopting a meta-learning method, improves the practicability and reliability of artificial intelligence automatic detection in the field of difficult and complicated diseases, and effectively enhances the accuracy of medical image classification.
The aim of the invention is achieved by the following technical measures: constructing three meta-learners including a multi-scale CNN feature extractor, a metric learner and a classification discriminator, and designing metric standards on the meta-learners; on each task of the training set, a target set is learned through support set distance measurement, a measurement standard is finally obtained through learning, and then for a new task of the testing set, the target set can be rapidly and correctly classified only by means of a small number of samples of the support set. The method specifically comprises the following steps:
reading a data set, namely a training data set, which is a medical image data set labeled for 1000 cases, wherein the training data set comprises a training set, a verification set and a test set;
step 2, carrying out image preprocessing operation, carrying out primary processing on the data in the training data set, specifying the range of the data, specifying the range and the size of the data, and removing irrelevant data;
step 3, data augmentation, namely normalizing the medical image data, and performing image rotation, noise addition and amplification treatment;
step 4, building a multi-scale feature extractor for algorithm model construction, wherein the feature extractor is a CNN feature extractor for extracting multi-scale super-pixel features of the image;
step 5, constructing a metric learning device and calculating the distance between image features;
step 6, setting up a classification judger, and designing a network layer structure and training network parameters;
step 7, entering a model training process, wherein a multi-scale relation network related to the model training process is an end-to-end differentiable structure, and estimating and adjusting the values of parameters in the model by adopting a back propagation algorithm and an adaptive moment;
step 8, finishing output; and obtaining a relation score after learning through a metric learning device according to the relation characteristics, wherein the relation score is a result of the image classification task of the few samples at one time.
Particularly, the superpixel is a small area consisting of a series of pixel points which are adjacent in position and similar in color, brightness or texture characteristics; most of the small areas keep effective information for further image segmentation, and the boundary information of objects in the images is generally not damaged; superpixels are graphs which divide a pixel-level graph into area-level graphs, and are abstractions of basic information elements; and for each block of superpixel, extracting a plurality of image patches with different sizes by taking the center of the superpixel as the center to serve as the input of a multi-scale CNN model, and performing feature extraction operation on the image patches corresponding to each block of superpixel by using the multi-scale CNN model by the multi-scale feature extractor based on a relation network.
Particularly, the specific network layer structure consists of two convolution modules, namely a convolution layer, a batch normalization layer, a modified linear unit layer and a maximum pooling layer; wherein the convolutional layers are composed of 64 filters with the size of 3x3, zero padding is not carried out, and the number of channels of the input signal of the first convolutional layer is determined by the number of channels of the input picture; the size of the pooling layer is 2x2, no zero padding is performed; the third and fourth convolution modules are composed of convolution layers, batch normalization layers and modified linear unit layers, wherein each convolution layer is composed of 64 filters with the size of 3x3 and is subjected to zero padding; inputting the picture into a feature extractor to obtain the size of (W, M, 64); a second, third and fourth layer feature map, wherein M represents the width and height of the feature map and 64 represents the depth of the feature map; and then, splicing the second layer characteristic diagram, the third layer characteristic diagram and the fourth layer characteristic diagram in the depth direction to obtain the multi-scale characteristic with the size (M, M, 128).
In particular, the specific structure of the metric learner is composed of two convolution modules and two full-connection layers; first of allThe first convolution module and the second convolution module consist of convolution layers, batch normalization layers, correction and modification linear unit layers and maximum pooling layers; wherein the convolutional layers are composed of 64 filters with the size of 3x3 or 2x2, and the number of channels of the input signal of the first convolutional layer is determined by the number of channels of the relation characteristic, here, the value is 128, the size is 2x2, and zero padding is performed; obtaining a feature graph with the size of (N, N,64) by the relationship features through a multi-scale learning device, and straightening the feature graph and transmitting the feature graph into a full connection layer; the first layer of full connection layer adopts a ReLU activation function, and the second layer of full connection layer adopts a Sigmoid activation function; inputting the relation feature into the metric learner to obtain a value of [0,1 ]]A value of (a), which represents the similarity between two picture features, this similarity being referred to as the relationship score; obtaining N under the condition of one-time N-wayk-shot few-sample image classification task setting2B similarity values; for each sample in the target set, taking the maximum value in the relation scores of the mean values of the samples of each type of the support set as a prediction label: finally, obtaining classification labels of the N x b target set samples; formulating a multi-scale relationship network as follows:
Figure BDA0002439354360000081
Figure BDA0002439354360000082
Figure BDA0002439354360000083
Figure BDA0002439354360000084
Figure BDA0002439354360000085
wherein the content of the first and second substances,i=1,…k;j=1,…b;n=1,…N;m-1,…N;
Figure BDA00024393543600000816
a representation feature extractor;
Figure BDA00024393543600000817
a representative metric learner;
Figure BDA0002439354360000086
and
Figure BDA0002439354360000087
representing the original pixel characteristics of the sample pictures of the support set and the target set;
Figure BDA0002439354360000088
and
Figure BDA0002439354360000089
representing multi-scale characteristics of the support set and the target set samples;
Figure BDA00024393543600000810
a prototype representing each type of sample of the support set;
Figure BDA00024393543600000820
representing the three-dimensional vector reconnection subtraction and then taking an absolute value;
Figure BDA00024393543600000811
representing a relationship score;
Figure BDA00024393543600000812
a prediction tag representing a sample of the target set.
In particular, the network parameter design comprises the design of loss function, the selection of an optimizer and the design of an optimization algorithm, and the method comprises the following processes: for image classification tasks, cross entropy is generally selected to calculate the distance between a predicted probability distribution and an expected probability distribution; in the multi-scale relation network, the measurement learner describes the similarity between two images, and the last layer of network of the measurement learner has only one output node; the designed loss function has the following specific formula:
Figure BDA00024393543600000813
wherein the content of the first and second substances,
Figure BDA00024393543600000814
represents the parameters of the feature extractor and is,
Figure BDA00024393543600000815
a proxy metric learner parameter; argmin means when the following formula reaches the minimum
Figure BDA00024393543600000818
And
Figure BDA00024393543600000819
taking the value of (A); i.e. labels for each category of the support set, ynA label is predicted for the target set picture sample,
Figure BDA0002439354360000091
the similarity of the same matching pair is 1, and the similarity of different matching pairs is 0;
Figure BDA0002439354360000092
calculated by the formula (1-4).
In particular, in the training process, after an epoch is finished, namely after the epoch training of different tasks is finished, the accuracy rate on the verification set is calculated, and the highest accuracy rate on the verification set up to now is recorded; when the continuous epochs are not optimal for multiple times, the accuracy rate can be considered not to be improved any more; when the accuracy is not improved any more or the accuracy is gradually reduced, the iteration can be stopped, a model corresponding to the highest accuracy is output, and then the model is used for testing; the calculation formula of the accuracy is as follows:
Figure BDA0002439354360000093
the epamode represents the number of tasks, and one task is a few-sample image classification task; n isTTable total number of samples on non-target set;
Figure BDA0002439354360000094
a predictive tag value representing a sample of the target set,
Figure BDA0002439354360000095
a label value representing a sample of the target set; b represents the value of each type of picture of the target set, namely the batch number; and N represents the number of sample categories in one small-sample image classification task.
Particularly, the overall output process comprises the steps of inputting a support set and a target set to a feature extractor, obtaining a relation feature by obtaining extracted multi-scale features of the support set and multi-scale features of the target set, and finally obtaining a relation score by a measurement learning device.
The invention has the advantages and effects that: by utilizing meta-learning, the defect that the data amount of a single disease is small and the disease is dispersed is overcome, the learning training of classifying few samples of the medical image processing intelligent system is completed, the filtering accuracy is obviously improved, the production efficiency is greatly improved, and the upgrading of the intelligent diagnosis deep learning technology in the medical industry is facilitated. Feasibility support is provided for the development and application of medical imaging equipment.
Drawings
FIG. 1 is a diagram of a network topology used in the present invention;
FIG. 2 is a functional diagram of a multi-scale CNN feature extractor employed in the present invention;
FIG. 3 is a diagram of a multi-scale CNN feature extractor employed in the present invention;
FIG. 4 is a diagram of a metric learner used in accordance with the present invention;
FIG. 5 is a flow chart of the overall training process employed by the present invention;
Detailed Description
The principle of the invention is that the task of the low sample learning based on the meta-learning is to establish a model with excellent generalization performance, and further, the low sample learning problem of the deep learning method is processed by utilizing the meta-learning. For new category picture tasks which are not seen in the training process, the previous experience knowledge can be well utilized, so that the deep learning method can be adjusted according to the new tasks, the learning ability of the learning society of the network is enhanced, the deep learning and generalization are fast, and the correct classification can be carried out only by means of a small amount of samples of each new category.
In the present invention, the metric (metric), that is, the correlation between two samples expressed by some way, such as euclidean distance, may be considered that in a certain embedding projection space, a neural network with parameters is generally used to complete the embedding process, and samples with closer distances are more similar, that is, they may be classified into the same category. Of course, there are many metric methods, but in recent years, cosine distance and euclidean distance are used more, and the metric-based method is used to realize small sample learning. One key point in metric learning is how to project samples into a space, where samples that are close together are of the same class. The reason why it is possible to express similarity using distance on the omniroot handwritten character data set is because it is possible to use distance to explain similarity in the case where handwritten number recognition is implemented similarly to KNN, since the shape of a handwritten character can be seen in a matrix obtained by projecting the feature matrix of the picture into a space, which is the shape of the handwritten character. In a task with a small sample size based on optimized meta-learning, some common learners have an overfitting phenomenon because of fewer samples, and meanwhile, the learning effect is good in a training process, and the learners can be generally trained to iterate millions or even tens of millions and then can be converged.
The invention adds LSTM network in the meta-learner, uses a small amount of marked samples in the task, optimizes and updates the initialization parameters of the learner through the meta-learner, so that the learner can use only a small amount of samples when facing the corresponding task, and simultaneously, the convergence is faster, thereby achieving the purpose of rapidly learning and solving the task.
The invention realizes picture analysis and target positioning based on deep learning, converts accumulated medical data into an available algorithm model through the combination of computer vision, big data and deep learning technology, analyzes images by utilizing an image processing technology and a machine learning algorithm, detects the position of a focus in the medical images, overcomes the subjective difference between different operators by utilizing an AI technology, lightens the workload of manual processing, helps a doctor to improve the diagnosis efficiency on the aspects of accuracy and speed of a computer, and solves the problem of low accuracy caused by insufficient training data of artificial intelligence in the development of a medical auxiliary diagnosis system.
As shown in fig. 1, the present invention specifically comprises the following steps:
reading a data set, namely a training data set, which is a medical image data set labeled for 1000 cases, wherein the training data set comprises a training set, a verification set and a test set;
step 2, carrying out image preprocessing operation, carrying out primary processing on the data in the training data set, specifying the range of the data, specifying the range and the size of the data, and removing irrelevant data;
step 3, data augmentation, namely normalizing the medical image data, and performing image rotation, noise addition and amplification treatment;
step 4, building a multi-scale feature extractor for algorithm model construction, wherein the feature extractor is a CNN feature extractor for extracting multi-scale super-pixel features of the image;
step 5, constructing a metric learning device and calculating the distance between image features;
step 6, setting up a classification judger, and designing a network layer structure and training network parameters;
and 7, entering a model training process, wherein a multi-scale relation network related to the model training process is an end-to-end differentiable structure, and estimating and adjusting the values of parameters in the model by adopting a back propagation algorithm and an adaptive moment.
Step 8, finishing output; and obtaining a relation score after learning through a metric learning device according to the relation characteristics, wherein the relation score is a result of the image classification task of the few samples at one time.
In the invention, three meta-learners, namely a multi-scale CNN feature extractor, a metric learner and a classification discriminator are set up, and simultaneously, the metric standard on the meta-learners is designed; on each task of the training set, a target set is learned through support set distance measurement, a measurement standard is finally obtained through learning, and then for a new task of the testing set, the target set can be rapidly and correctly classified only by means of a small number of samples of the support set. The similarity measurement is learned from a wide task space, so that the experience extracted from the past learning task is used for guiding the learning of a new task, and the purpose of learning how to learn is achieved.
The invention has wide application prospect in the fields of computer-aided diagnosis systems of future medical diseases and the like.
The invention is further illustrated by the following figures and examples.
Example 1: the meta learning is applied to the learning problem of few samples in the field of medical image classification processing, a training network is built, and network parameters are set. The training network supports a small amount of medical data samples to realize rapid and high-precision classification;
in the foregoing, a superpixel is a small region composed of a series of pixel points adjacent in position and similar in color, brightness or texture characteristics; most of the small areas keep effective information for further image segmentation, and the boundary information of objects in the images is generally not damaged; superpixels are graphs obtained by dividing a picture element level (pixel-level) into a region level (discrete-level) and are abstractions of basic information elements. As shown in fig. 2, for each super pixel, a plurality of image patches with different sizes are extracted with the center of the super pixel as the center to be used as the input of the multi-scale CNN model, and the multi-scale feature extractor performs feature extraction operation on the image patches corresponding to each super pixel by using the multi-scale CNN model based on the relational network; as shown in fig. 3, the specific network layer structure is composed of two convolution modules, including convolution layer, batch normalization layer, modified linear unit layer and max pooling layer; wherein the convolutional layers are composed of 64 filters with the size of 3x3, zero padding is not carried out, and the number of channels of the input signal of the first convolutional layer is determined by the number of channels of the input picture; the size of the pooling layer is 2x2, no zero padding is performed; the third and fourth convolution modules consist of convolutional layers, which consist of 64 filters of size 3x3, zero-padded, batch normalization layers, and modified linear cell layers. The picture is input to a feature extractor, resulting in a size of (W, M, 64). Second, third, and fourth layer feature maps, where M represents the width and height of the feature map and 64 represents the depth of the feature map. And then, splicing the second layer characteristic diagram, the third layer characteristic diagram and the fourth layer characteristic diagram in the depth direction to obtain the multi-scale characteristic with the size (M, M, 128).
In the foregoing, the specific structure of the metric learner is shown in fig. 4, and the metric learner is composed of two convolution modules and two full-connection layers in total. The first and second convolution modules are comprised of convolution layers, batch normalization layers, modification linear unit layers, and max pooling layers. Wherein the convolutional layers are composed of 64 filters with the size of 3x3 or 2x2, and the number of channels of the input signal of the first convolutional layer is determined by the number of channels of the relation characteristic, here the value is 128, and the size is 2x2, and zero padding is performed. Obtaining a feature graph with the size of (N, N,64) by the relationship features through a multi-scale learning device, and straightening the feature graph and transmitting the feature graph into a full connection layer; the first layer of full connection layer adopts a ReLU activation function, and the second layer of full connection layer adopts a Sigmoid activation function. Inputting the relation feature into the metric learner to obtain a value of [0,1 ]]The value of (c), which represents the similarity between two picture features, this similarity is called the relationship score. Obtaining N under the condition of one-time N-wayk-shot few-sample image classification task setting2B similarity values; for each sample in the target set, taking the maximum value in the relation scores of the mean values of the samples of each type of the support set as a prediction label: finally, we get the classification labels of the N x b target set samples. Formulating a multi-scale relationship network as follows:
Figure BDA0002439354360000121
Figure BDA0002439354360000122
Figure BDA0002439354360000123
Figure BDA0002439354360000124
Figure BDA0002439354360000125
wherein, i is 1, … k; j is 1, … b; n is 1, … N; m-1, … N;
Figure BDA0002439354360000129
a representation feature extractor;
Figure BDA00024393543600001210
a representative metric learner;
Figure BDA0002439354360000126
and
Figure BDA0002439354360000127
representing the original pixel characteristics of the sample pictures of the support set and the target set;
Figure BDA0002439354360000128
and
Figure BDA0002439354360000131
representing multi-scale characteristics of the support set and the target set samples;
Figure BDA0002439354360000132
a prototype representing each type of sample of the support set;
Figure BDA00024393543600001315
representation of absolute value of three-dimensional vector after reconnection subtractionA value;
Figure BDA0002439354360000133
representing relationship scores
Figure BDA0002439354360000134
A prediction tag representing a sample of the target set.
In the foregoing, the network parameter design includes a loss function design, an optimizer selection, and an optimization algorithm design, and the process includes:
for image classification tasks, the cross entropy is typically chosen to compute the distance between the predicted probability distribution and the desired probability distribution. However, in the multi-scale relationship network, the similarity between two images is characterized by the metric learner, and the last layer of the network of the metric learner has only one output node. The designed loss function has the following specific formula:
Figure BDA0002439354360000135
wherein the content of the first and second substances,
Figure BDA00024393543600001311
represents the parameters of the feature extractor and is,
Figure BDA00024393543600001312
a proxy metric learner parameter; argmin means when the following formula reaches the minimum
Figure BDA00024393543600001313
And
Figure BDA00024393543600001314
taking the value of (A); i.e. labels for each category of the support set, ynA label is predicted for the target set picture sample,
Figure BDA0002439354360000136
the similarity of the same matching pair is 1, and the similarity of different matching pairs is 0;
Figure BDA0002439354360000137
obtained by calculation of formula (1-4);
in the training process, after an epoch is finished, that is, after the epoch training of different tasks is completed, the accuracy on the verification set is calculated, and the highest accuracy on the verification set up so far is recorded. When the continuous epochs are not optimal, the accuracy rate is not increased any more. When the accuracy is not improved any more or the accuracy is gradually reduced, the iteration can be stopped, the model corresponding to the highest accuracy is output, and then the model is used for testing. The calculation formula of the accuracy is as follows:
Figure BDA0002439354360000138
the epamode represents the number of tasks, and one task is a few-sample image classification task; n isTTable total number of samples on non-target set;
Figure BDA0002439354360000139
a predictive tag value representing a sample of the target set,
Figure BDA00024393543600001310
a label value representing a sample of the target set; b represents the value of each type of picture of the target set, namely the batch number; and N represents the number of sample categories in one small-sample image classification task.
Among the foregoing, as shown in FIG. 5; the whole output process comprises the steps of enabling the support set and the target set to go to a feature extractor, obtaining relation features by obtaining extracted support set multi-scale features and target set multi-scale features, and obtaining relation scores through a measurement learning device.
In the embodiment of the invention, the experimental conditions are IntelCorei 73770 type CPU, a memory 16GBytes, an NVIDIA GeForce GTX 0686 GB type accelerator and a Pytrch 0.4.0 deep learning frame.
In the embodiment of the invention, the similarity measurement is learned from a wide task space, so that the experience extracted from the past learning task is used for guiding the learning of a new task, and the purpose of learning how to learn is achieved.
In the embodiment of the invention, the larger the image patch scale is, the more context relations can be acquired, and the more context information is included in the extracted features, which is more beneficial to improving the function of the meta-learner.

Claims (3)

1. A small sample element learning training method for medical image classification processing artificial intelligence is characterized in that three element learners including a multi-scale CNN feature extractor, a measurement learner and a classification discriminator are set up, and simultaneously, the element learners are subjected to measurement standard design; on each task of the training set, a target set is learned through support set distance measurement, a measurement standard is finally obtained through learning, and then for a new task of the test set, the target set can be rapidly and correctly classified only by means of a small number of samples of the support set; the method specifically comprises the following steps:
reading a data set, namely a training data set, which is a medical image data set labeled for 1000 cases, wherein the training data set comprises a training set, a verification set and a test set;
step 2, carrying out image preprocessing operation, carrying out primary processing on the data in the training data set, specifying the range of the data, specifying the range and the size of the data, and removing irrelevant data;
step 3, data augmentation, namely normalizing the medical image data, and performing image rotation, noise addition and amplification treatment;
step 4, building a multi-scale feature extractor for algorithm model construction, wherein the feature extractor is a CNN feature extractor for extracting multi-scale super-pixel features of the image;
step 5, constructing a metric learning device and calculating the distance between image features;
step 6, setting up a classification judger, and designing a network layer structure and training network parameters;
step 7, entering a model training process, wherein a multi-scale relation network related to the model training process is an end-to-end differentiable structure, and estimating and adjusting the values of parameters in the model by adopting a back propagation algorithm and an adaptive moment;
step 8, finishing output; obtaining a relation score after learning through a metric learning device according to the relation characteristics, wherein the relation score is a result of a one-time few-sample image classification task;
the super-pixel is a small area formed by a series of pixel points which are adjacent in position and similar in color, brightness or texture characteristics; most of the small areas keep effective information for further image segmentation, and the boundary information of objects in the images is generally not damaged; superpixels are graphs which divide a pixel-level graph into area-level graphs, and are abstractions of basic information elements; for each super pixel, extracting a plurality of image patches with different sizes by taking the center of the super pixel as the center to serve as the input of a multi-scale CNN model, and performing feature extraction operation on the image patches corresponding to each super pixel by the multi-scale CNN model by the multi-scale feature extractor based on a relation network;
the first two convolution modules of the specific network layer structure consist of a convolution layer, a batch standardization layer, a modified linear unit layer and a maximum pooling layer; wherein the convolutional layers are composed of 64 filters with the size of 3x3, zero padding is not carried out, and the number of channels of the input signal of the first convolutional layer is determined by the number of channels of the input picture; the size of the pooling layer is 2x2, no zero padding is performed; the third and fourth convolution modules are composed of convolution layers, batch normalization layers and modified linear unit layers, wherein each convolution layer is composed of 64 filters with the size of 3x3 and is subjected to zero padding; inputting the picture into a feature extractor to obtain a second, third and fourth layer feature map with the size of (M, M, 64), wherein M represents the width and height of the feature map, and 64 represents the depth of the feature map; then, splicing the second layer characteristic diagram, the third layer characteristic diagram and the fourth layer characteristic diagram in the depth direction to obtain a multi-scale characteristic with the size of (M, M, 128); the specific structure of the metric learner is composed of two convolution modules and two full-connection layers; the first convolution module and the second convolution module are composed of a convolution layer, a batch normalization layer, a modified linear unit layer and a maximum pooling layer; wherein the convolutional layer comprises 64 layers with the size of 3x3 or 2x2The number of channels of the input signal of the first convolutional layer is determined by the number of channels of the relation characteristic, the number of channels is 128, the size of the filter is 2x2, and zero padding is carried out; obtaining a feature graph with the size of (N, N,64) by the relationship features through a multi-scale learning device, and straightening the feature graph and transmitting the feature graph into a full connection layer; the first layer of full connection layer adopts a ReLU activation function, and the second layer of full connection layer adopts a Sigmoid activation function; inputting the relation feature into the metric learner to obtain a value of [0,1 ]]A value of (a), which represents the similarity between two picture features, this similarity being referred to as the relationship score; obtaining N under the condition of one-time N-way k-shot few-sample image classification task setting2B similarity values; for each sample in the target set, taking the maximum value in the relation scores of the mean values of the samples of each type of the support set as a prediction label: finally, obtaining classification labels of the N x b target set samples; formulating a multi-scale relationship network as follows:
Figure FDA0002864105380000031
Figure FDA0002864105380000032
Figure FDA0002864105380000033
Figure FDA0002864105380000034
Figure FDA0002864105380000035
wherein, i is 1, … k; j is 1, … b; n is 1, … N; m is 1, … N;
Figure FDA0002864105380000036
a representation feature extractor; gφA representative metric learner;
Figure FDA0002864105380000037
and
Figure FDA0002864105380000038
representing the original pixel characteristics of the sample pictures of the support set and the target set;
Figure FDA0002864105380000039
and
Figure FDA00028641053800000310
representing the multi-scale superpixel characteristics of the support set and the target set samples;
Figure FDA00028641053800000311
means representing each type of sample of the support set; [ … …]Representing and subtracting the three-dimensional vector reconnection and then taking an absolute value;
Figure FDA00028641053800000312
representing a relationship score;
Figure FDA00028641053800000313
a prediction tag representing a sample of the target set;
the network parameter design comprises the design of loss function, the selection of an optimizer and the design of an optimization algorithm, and the method comprises the following processes: for image classification tasks, cross entropy is generally selected to calculate the distance between a predicted probability distribution and an expected probability distribution; in the multi-scale relation network, the measurement learner describes the similarity between two images, and the last layer of network of the measurement learner has only one output node; the designed loss function specific formula is shown as follows;
Figure FDA0002864105380000041
wherein the content of the first and second substances,
Figure FDA0002864105380000042
represents a feature extractor parameter, phi represents a metric learner parameter; argmin means when the following formula reaches the minimum
Figure FDA0002864105380000043
And phi is taken; n represents the sample category number in a few-sample image classification task at a time; b represents the value of each type of picture of the target set, namely the batch number, ynA label is predicted for the target set picture sample,
Figure FDA0002864105380000044
i.e., labels for each category of the support set;
Figure FDA0002864105380000045
the similarity of the same matching pair is 1, and the similarity of different matching pairs is 0;
Figure FDA0002864105380000046
calculated by the formula (1-4).
2. The method for learning and training the small sample elements of the medical image classification processing artificial intelligence as claimed in claim 1, wherein in the training process, after an epoch is finished, namely after the epoch different tasks are trained, the accuracy on the verification set is calculated, and the highest accuracy on the verification set up to now is recorded; when the continuous epochs are not optimal for multiple times, the accuracy rate can be considered not to be improved any more; when the accuracy is not improved any more or the accuracy is gradually reduced, the iteration can be stopped, a model corresponding to the highest accuracy is output, and then the model is used for testing; the calculation formula of the accuracy is as follows:
Figure FDA0002864105380000047
the epamode represents the number of tasks, and one task is a few-sample image classification task; n isrRepresenting the total number of samples on the target set; b represents the value of each type of picture of the target set, namely the batch number; n represents the sample category number in a few-sample image classification task at a time;
Figure FDA0002864105380000052
i.e. labels for each category of the support set, ynPredicting labels for the target set picture samples;
Figure FDA0002864105380000051
it means that the similarity of the same matching pair is 1 and the similarity of different matching pairs is 0.
3. The method as claimed in claim 1, wherein the process of outputting the whole includes inputting a support set and a target set into a feature extractor, obtaining a relationship feature by obtaining extracted multi-scale features of the support set and multi-scale features of the target set, and finally obtaining a relationship score by a metric learner.
CN202010262936.7A 2020-04-03 2020-04-03 Small sample element learning training method for medical image classification processing artificial intelligence Active CN111476292B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010262936.7A CN111476292B (en) 2020-04-03 2020-04-03 Small sample element learning training method for medical image classification processing artificial intelligence

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010262936.7A CN111476292B (en) 2020-04-03 2020-04-03 Small sample element learning training method for medical image classification processing artificial intelligence

Publications (2)

Publication Number Publication Date
CN111476292A CN111476292A (en) 2020-07-31
CN111476292B true CN111476292B (en) 2021-02-19

Family

ID=71749916

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010262936.7A Active CN111476292B (en) 2020-04-03 2020-04-03 Small sample element learning training method for medical image classification processing artificial intelligence

Country Status (1)

Country Link
CN (1) CN111476292B (en)

Families Citing this family (41)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112116957A (en) * 2020-08-20 2020-12-22 澳门科技大学 Disease subtype prediction method, system, device and medium based on small sample
CN112069961B (en) * 2020-08-28 2022-06-14 电子科技大学 Few-sample document layout analysis method based on metric learning
CN112016506B (en) * 2020-09-07 2022-10-11 重庆邮电大学 Classroom attitude detection model parameter training method capable of quickly adapting to new scene
CN111985581B (en) * 2020-09-09 2022-07-05 福州大学 Sample-level attention network-based few-sample learning method
CN112115993B (en) * 2020-09-11 2023-04-07 昆明理工大学 Zero sample and small sample evidence photo anomaly detection method based on meta-learning
CN112116002A (en) * 2020-09-18 2020-12-22 北京旋极信息技术股份有限公司 Determination method, verification method and device of detection model
CN112101473B (en) * 2020-09-22 2022-07-15 南京邮电大学 Smoke detection algorithm based on small sample learning
CN112329827B (en) * 2020-10-26 2022-08-23 同济大学 Increment small sample target detection method based on meta-learning
CN112364990B (en) * 2020-10-29 2021-06-04 北京语言大学 Method and system for realizing grammar error correction and less sample field adaptation through meta-learning
CN112348792A (en) * 2020-11-04 2021-02-09 广东工业大学 X-ray chest radiography image classification method based on small sample learning and self-supervision learning
CN112686850B (en) * 2020-12-24 2021-11-02 上海体素信息科技有限公司 Method and system for few-sample segmentation of CT image based on spatial position and prototype network
CN112632315B (en) * 2020-12-30 2022-03-29 珠海大横琴科技发展有限公司 Method and device for retrieving remote sensing image
CN112784879A (en) * 2020-12-31 2021-05-11 前线智能科技(南京)有限公司 Medical image segmentation or classification method based on small sample domain self-adaption
CN112863549A (en) * 2021-01-20 2021-05-28 广东工业大学 Voice emotion recognition method and device based on meta-multitask learning
CN112989085B (en) * 2021-01-29 2023-07-25 腾讯科技(深圳)有限公司 Image processing method, device, computer equipment and storage medium
CN112800257A (en) * 2021-02-10 2021-05-14 上海零眸智能科技有限公司 Method for quickly adding sample training data based on image searching
CN112967187A (en) * 2021-02-25 2021-06-15 深圳海翼智新科技有限公司 Method and apparatus for target detection
CN112927152B (en) * 2021-02-26 2024-02-02 平安科技(深圳)有限公司 CT image denoising processing method, device, computer equipment and medium
CN113052802B (en) * 2021-03-11 2024-04-09 南京大学 Small sample image classification method, device and equipment based on medical image
CN112861995B (en) * 2021-03-15 2023-03-31 中山大学 Unsupervised few-sample image classification method and system based on model independent meta learning and storage medium
CN112949740B (en) * 2021-03-17 2022-11-25 重庆邮电大学 Small sample image classification method based on multilevel measurement
CN113096080B (en) * 2021-03-30 2024-01-16 四川大学华西第二医院 Image analysis method and system
CN112750074B (en) * 2021-04-06 2021-07-02 南京智莲森信息技术有限公司 Small sample image feature enhancement method and system and image classification method and system
CN113221110B (en) * 2021-04-08 2022-06-28 浙江工业大学 Remote access Trojan intelligent analysis method based on meta-learning
CN112949770B (en) * 2021-04-08 2023-12-26 深圳市医诺智能科技发展有限公司 Medical image identification and classification method and terminal
CN113269734B (en) * 2021-05-14 2023-04-07 成都市第三人民医院 Tumor image detection method and device based on meta-learning feature fusion strategy
CN113255701B (en) * 2021-06-24 2021-10-22 军事科学院系统工程研究院网络信息研究所 Small sample learning method and system based on absolute-relative learning framework
CN113505799B (en) * 2021-06-30 2022-12-23 深圳市慧鲤科技有限公司 Significance detection method and training method, device, equipment and medium of model thereof
CN113627434A (en) * 2021-07-07 2021-11-09 中国科学院自动化研究所 Method and device for building processing model applied to natural image
CN113535953B (en) * 2021-07-15 2022-05-27 湖南大学 Meta learning-based few-sample classification method
CN113642465B (en) * 2021-08-13 2022-07-08 石家庄铁道大学 Bearing health assessment method based on relational network
CN113505861B (en) * 2021-09-07 2021-12-24 广东众聚人工智能科技有限公司 Image classification method and system based on meta-learning and memory network
CN114386454B (en) * 2021-12-09 2023-02-03 首都医科大学附属北京友谊医院 Medical time sequence signal data processing method based on signal mixing strategy
CN114266977B (en) * 2021-12-27 2023-04-07 青岛澎湃海洋探索技术有限公司 Multi-AUV underwater target identification method based on super-resolution selectable network
CN114491039B (en) * 2022-01-27 2023-10-03 四川大学 Primitive learning few-sample text classification method based on gradient improvement
CN114693662A (en) * 2022-04-12 2022-07-01 吉林大学 Histopathological section subtype classification method and device, medium and terminal
CN114723998B (en) * 2022-05-05 2023-06-20 兰州理工大学 Small sample image classification method and device based on large-boundary Bayesian prototype learning
CN115082770B (en) * 2022-07-04 2024-02-23 青岛科技大学 Image center line structure extraction method based on machine learning
CN115393338A (en) * 2022-09-02 2022-11-25 复旦大学附属中山医院 Biological tissue identification model construction method and device and electronic equipment
CN115527269B (en) * 2022-10-10 2023-05-16 动自由(北京)科技有限公司 Intelligent human body posture image recognition method and system
CN115619810B (en) * 2022-12-19 2023-10-03 中国医学科学院北京协和医院 Prostate partition segmentation method, system and equipment

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107220506A (en) * 2017-06-05 2017-09-29 东华大学 Breast cancer risk assessment analysis system based on depth convolutional neural networks
CN107507189A (en) * 2017-07-04 2017-12-22 西北大学 Mouse CT image kidney dividing methods based on random forest and statistical model
JP7213241B2 (en) * 2017-11-14 2023-01-26 マジック リープ, インコーポレイテッド Meta-learning for Multitask Learning on Neural Networks
CN108876797B (en) * 2018-06-08 2021-09-24 长安大学 Image segmentation system and method based on Spiking-SOM neural network clustering
CN109492711B (en) * 2018-12-07 2020-08-25 杭州电子科技大学 Malignant melanoma and non-malignant melanoma classification system based on deep learning
CN109753997B (en) * 2018-12-19 2022-11-22 湖南科技大学 Automatic accurate robust segmentation method for liver tumor in CT image
CN109685135B (en) * 2018-12-21 2022-03-25 电子科技大学 Few-sample image classification method based on improved metric learning
CN110020682B (en) * 2019-03-29 2021-02-26 北京工商大学 Attention mechanism relation comparison network model method based on small sample learning

Also Published As

Publication number Publication date
CN111476292A (en) 2020-07-31

Similar Documents

Publication Publication Date Title
CN111476292B (en) Small sample element learning training method for medical image classification processing artificial intelligence
WO2020215984A1 (en) Medical image detection method based on deep learning, and related device
CN108921851B (en) Medical CT image segmentation method based on 3D countermeasure network
CN106203432B (en) Positioning system of region of interest based on convolutional neural network significance map
WO2019200747A1 (en) Method and device for segmenting proximal femur, computer apparatus, and storage medium
CN107862694A (en) A kind of hand-foot-and-mouth disease detecting system based on deep learning
CN107169974A (en) It is a kind of based on the image partition method for supervising full convolutional neural networks more
CN108257135A (en) The assistant diagnosis system of medical image features is understood based on deep learning method
CN109544518B (en) Method and system applied to bone maturity assessment
CN112150428A (en) Medical image segmentation method based on deep learning
CN110689025A (en) Image recognition method, device and system, and endoscope image recognition method and device
CN112102266A (en) Attention mechanism-based cerebral infarction medical image classification model training method
Pan et al. Cell detection in pathology and microscopy images with multi-scale fully convolutional neural networks
CN110059656B (en) Method and system for classifying white blood cells based on convolution countermeasure generation neural network
CN108549912A (en) A kind of medical image pulmonary nodule detection method based on machine learning
CN110853070A (en) Underwater sea cucumber image segmentation method based on significance and Grabcut
CN114897914A (en) Semi-supervised CT image segmentation method based on confrontation training
CN113421228A (en) Thyroid nodule identification model training method and system based on parameter migration
Tsai et al. Machine learning based common radiologist-level pneumonia detection on chest X-rays
Mamdouh et al. A New Model for Image Segmentation Based on Deep Learning.
CN117237351B (en) Ultrasonic image analysis method and related device
CN116630964A (en) Food image segmentation method based on discrete wavelet attention network
Qian et al. Multi-scale context UNet-like network with redesigned skip connections for medical image segmentation
CN110008902A (en) A kind of finger vein identification method and system merging essential characteristic and deformation characteristics
CN113936006A (en) Segmentation method and device for processing high-noise low-quality medical image

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant