CN113888470B - Diagnosis method and device based on convolutional neural network and multi-modal medical image - Google Patents

Diagnosis method and device based on convolutional neural network and multi-modal medical image Download PDF

Info

Publication number
CN113888470B
CN113888470B CN202111038996.1A CN202111038996A CN113888470B CN 113888470 B CN113888470 B CN 113888470B CN 202111038996 A CN202111038996 A CN 202111038996A CN 113888470 B CN113888470 B CN 113888470B
Authority
CN
China
Prior art keywords
convolutional neural
neural network
network
medical image
multiple modes
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111038996.1A
Other languages
Chinese (zh)
Other versions
CN113888470A (en
Inventor
徐枫
周展平
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tsinghua University
Original Assignee
Tsinghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tsinghua University filed Critical Tsinghua University
Priority to CN202111038996.1A priority Critical patent/CN113888470B/en
Publication of CN113888470A publication Critical patent/CN113888470A/en
Application granted granted Critical
Publication of CN113888470B publication Critical patent/CN113888470B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H30/00ICT specially adapted for the handling or processing of medical images
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/20ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Medical Informatics (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Public Health (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Computational Linguistics (AREA)
  • Radiology & Medical Imaging (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Primary Health Care (AREA)
  • Biophysics (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Epidemiology (AREA)
  • Measuring And Recording Apparatus For Diagnosis (AREA)
  • Pathology (AREA)
  • Probability & Statistics with Applications (AREA)
  • Quality & Reliability (AREA)
  • Databases & Information Systems (AREA)

Abstract

The application provides a diagnosis method based on a convolutional neural network and multi-mode medical images, which relates to the technical fields of computer vision, deep learning and medical image processing, and comprises the following steps: collecting training data, wherein the training data comprises medical image samples of multiple modes and diagnosis marking information corresponding to the medical image samples of the multiple modes; optimizing network parameters of the convolutional neural network according to the training data to obtain an optimized classification network based on the convolutional neural network; medical images of multiple modes are acquired, the medical images of the multiple modes are input into a classification network based on a convolutional neural network, so that the prediction probability of diseases corresponding to the medical images of each mode is output, and the prediction probability is used for representing the probability of suffering from the diseases and the probability of not suffering from the diseases respectively. The application integrates medical images of multiple modes to diagnose, thereby improving the accuracy of disease diagnosis.

Description

Diagnosis method and device based on convolutional neural network and multi-modal medical image
Technical Field
The application relates to the technical fields of computer vision, deep learning and medical image processing, in particular to a diagnosis method and device based on a convolutional neural network and multi-mode medical images.
Background
In the fields of computer vision, natural language processing and the like, comprehensively utilizing data of multiple modes is a common method for improving algorithm performance. In practice, doctors often make use of multiple modality data for diagnosis. Convolutional Neural Networks (CNNs) are also widely used in the medical image processing field as an excellent image processing method. Common medical image types include CT, MRI, ultrasound images, pathology images, and the like. Existing computer-automated diagnostic methods based on medical images only utilize images of a single modality. Medical images of different modes can reflect information of physiological structures from multiple angles, so that the accuracy of diagnosis is improved.
Disclosure of Invention
The present application aims to solve at least one of the technical problems in the related art to some extent.
Therefore, a first object of the present application is to provide a diagnosis method based on convolutional neural network and multi-modal medical image, which solves the technical problem that the existing computer automated diagnosis method only uses single-modal image, realizes the purpose of diagnosing the multi-modal medical image by using deep learning method, and improves the accuracy of disease diagnosis.
A second object of the present application is to provide a diagnostic device based on a convolutional neural network and a multi-modal medical image.
A third object of the present application is to propose a non-transitory computer readable storage medium.
To achieve the above object, an embodiment of a first aspect of the present application provides a diagnostic method based on a convolutional neural network and a multi-modal medical image, including: collecting training data, wherein the training data comprises medical image samples of multiple modes and diagnosis marking information corresponding to the medical image samples of the multiple modes; optimizing network parameters of the convolutional neural network according to the training data to obtain an optimized classification network based on the convolutional neural network; medical images of multiple modes are acquired, the medical images of the multiple modes are input into a classification network based on a convolutional neural network, so that the prediction probability of diseases corresponding to the medical images of each mode is output, and the prediction probability is used for representing the probability of suffering from the diseases and the probability of not suffering from the diseases respectively.
Optionally, in an embodiment of the present application, the convolutional neural network-based classification network includes a plurality of convolutional neural network-based feature extraction networks and a multi-layer perceptron, each feature extraction network corresponds to a mode, and the network parameters of the convolutional neural network are optimized according to training data, so as to obtain an optimized convolutional neural network-based classification network, including:
Inputting the medical image samples of each mode in the medical image samples of a plurality of modes into a corresponding feature extraction network to perform feature extraction, wherein the extracted features are vectors;
The features of all modes are spliced into long vectors, the long vectors are input into a multi-layer perceptron, corresponding values are output, and the corresponding values are calculated by using softmax, so that the prediction probability of each category is obtained;
And acquiring a cross entropy loss function of the prediction probability according to the prediction probability of each category, and optimizing network parameters of the convolutional neural network according to the cross entropy loss function.
To achieve the above object, a second aspect of the present invention provides a diagnostic apparatus based on a convolutional neural network and a multi-modal medical image, comprising:
The acquisition module is used for acquiring training data, wherein the training data comprises medical image samples of multiple modes and diagnosis marking information corresponding to the medical image samples of the multiple modes;
The optimizing module is used for optimizing the network parameters of the convolutional neural network according to the training data to obtain an optimized classification network based on the convolutional neural network;
the classification module is used for acquiring medical images of multiple modes, inputting the medical images of multiple modes into a classification network based on a convolutional neural network so as to output the prediction probability of the diseases corresponding to the medical images of each mode, wherein the prediction probability is used for representing the probability of the diseases and the probability of the diseases not.
Optionally, in an embodiment of the present application, the classification network based on the convolutional neural network includes a plurality of feature extraction networks based on the convolutional neural network and a multi-layer perceptron, each feature extraction network corresponds to a mode, and the optimization module is specifically configured to:
Inputting the medical image samples of each mode in the medical image samples of a plurality of modes into a corresponding feature extraction network to perform feature extraction, wherein the extracted features are vectors;
The features of all modes are spliced into long vectors, the long vectors are input into a multi-layer perceptron, corresponding values are output, and the corresponding values are calculated by using softmax, so that the prediction probability of each category is obtained;
And acquiring a cross entropy loss function of the prediction probability according to the prediction probability of each category, and optimizing network parameters of the convolutional neural network according to the cross entropy loss function.
In order to achieve the above object, an embodiment of a third aspect of the present invention proposes a non-transitory computer-readable storage medium capable of performing a diagnosis method based on a convolutional neural network and a multi-modal medical image when instructions in the storage medium are executed by a processor.
The diagnosis method based on the convolutional neural network and the multi-modal medical image, the diagnosis device based on the convolutional neural network and the multi-modal medical image and the non-transitory computer readable storage medium solve the technical problem that the computer automated diagnosis method only uses single-modal images in the existing method, realize the purpose of diagnosing the multi-modal medical image by using a deep learning method, and improve the accuracy of disease diagnosis.
Additional aspects and advantages of the application will be set forth in part in the description which follows and, in part, will be obvious from the description, or may be learned by practice of the application.
Drawings
The foregoing and/or additional aspects and advantages of the application will become apparent and readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings, in which:
FIG. 1 is a flowchart of a diagnostic method based on convolutional neural network and multi-modal medical image according to an embodiment of the present application;
fig. 2 is a schematic structural diagram of a diagnostic device based on a convolutional neural network and a multi-modal medical image according to a second embodiment of the present application.
Detailed Description
Embodiments of the present application are described in detail below, examples of which are illustrated in the accompanying drawings, wherein like or similar reference numerals refer to like or similar elements or elements having like or similar functions throughout. The embodiments described below by referring to the drawings are illustrative and intended to explain the present application and should not be construed as limiting the application.
The following describes a diagnosis method and apparatus based on a convolutional neural network and a multi-modal medical image according to an embodiment of the present application with reference to the accompanying drawings.
Fig. 1 is a flowchart of a diagnosis method based on a convolutional neural network and a multi-modal medical image according to an embodiment of the application.
As shown in fig. 1, the diagnosis method based on the convolutional neural network and the multi-modal medical image comprises the following steps:
step 101, acquiring training data, wherein the training data comprises medical image samples of multiple modes and diagnosis marking information corresponding to the medical image samples of the multiple modes;
Step 102, optimizing network parameters of the convolutional neural network according to training data to obtain an optimized classification network based on the convolutional neural network;
Step 103, acquiring medical images of multiple modes, inputting the medical images of multiple modes into a classification network based on a convolutional neural network, and outputting the prediction probability of the disease corresponding to the medical images of each mode, wherein the prediction probability is used for representing the probability of suffering from the disease and the probability of not suffering from the disease respectively.
According to the diagnosis method based on the convolutional neural network and the multi-mode medical image, training data are collected, and the training data comprise medical image samples of multiple modes and diagnosis marking information corresponding to the medical image samples of the multiple modes; optimizing network parameters of the convolutional neural network according to the training data to obtain an optimized classification network based on the convolutional neural network; medical images of multiple modes are acquired, the medical images of the multiple modes are input into a classification network based on a convolutional neural network, so that the prediction probability of diseases corresponding to the medical images of each mode is output, and the prediction probability is used for representing the probability of suffering from the diseases and the probability of not suffering from the diseases respectively. Therefore, the technical problem that the computer automatic diagnosis method only utilizes a single-mode image in the existing method can be solved, the purpose of diagnosing the medical images in multiple modes comprehensively by utilizing a deep learning method is achieved, and the accuracy of disease diagnosis is improved.
After the multi-mode medical images and diagnostic labels for training are obtained, the multi-mode medical images are respectively input into different convolutional neural networks to extract features, all the features are spliced into a vector to serve as multi-mode comprehensive features, and finally the multi-mode comprehensive features are classified through a multi-layer perceptron. The parameters of the network are optimized through training, so that the automatic diagnosis of the diseases by using a computer is realized, and the method mainly comprises the following steps: collecting medical images of multiple modes of a patient and diagnosing and labeling; training by using a classification network based on a convolutional neural network by utilizing data, and optimizing network parameters; when in use, medical images of multiple modes of a patient are acquired, and the medical images are input into a classification network to obtain the prediction probability.
Further, in an embodiment of the present application, a convolutional neural network-based classification network includes a plurality of convolutional neural network-based feature extraction networks and a multi-layer perceptron, each feature extraction network corresponds to a mode, and network parameters of the convolutional neural network are optimized according to training data to obtain an optimized convolutional neural network-based classification network, including:
Inputting the medical image samples of each mode in the medical image samples of a plurality of modes into a corresponding feature extraction network to perform feature extraction, wherein the extracted features are vectors;
The features of all modes are spliced into long vectors, the long vectors are input into a multi-layer perceptron, corresponding values are output, and the corresponding values are calculated by using softmax, so that the prediction probability of each category is obtained;
And acquiring a cross entropy loss function of the prediction probability according to the prediction probability of each category, and optimizing network parameters of the convolutional neural network according to the cross entropy loss function.
For example, for a two-class problem that diagnoses whether a disease is present, the final output is a vector of length 2, where the values are the probability of having the disease and the probability of not having the disease, respectively.
And performing end-to-end training on the whole multi-mode ultrasonic image classification network by using the medical images of multiple modes so as to optimize network parameters by using a cross entropy loss function of predictive probability.
Fig. 2 is a schematic structural diagram of a diagnostic device based on a convolutional neural network and a multi-modal medical image according to a second embodiment of the present application.
As shown in fig. 2, the diagnosis device based on convolutional neural network and multi-modal medical image comprises:
the acquisition module 10 is configured to acquire training data, where the training data includes medical image samples of multiple modes and diagnostic labeling information corresponding to the medical image samples of multiple modes;
the optimizing module 20 is configured to optimize network parameters of the convolutional neural network according to the training data, and obtain an optimized classification network based on the convolutional neural network;
The classification module 30 is configured to acquire medical images of multiple modes, input the medical images of multiple modes into a classification network based on a convolutional neural network, and output prediction probabilities of diseases corresponding to the medical images of each mode, where the prediction probabilities are used to represent a probability of having a disease and a probability of not having the disease, respectively.
Further, in the embodiment of the present application, the classification network based on the convolutional neural network includes a plurality of feature extraction networks based on the convolutional neural network and a multi-layer perceptron, each feature extraction network corresponds to a mode, and the optimization module is specifically configured to:
Inputting the medical image samples of each mode in the medical image samples of a plurality of modes into a corresponding feature extraction network to perform feature extraction, wherein the extracted features are vectors;
The features of all modes are spliced into long vectors, the long vectors are input into a multi-layer perceptron, corresponding values are output, and the corresponding values are calculated by using softmax, so that the prediction probability of each category is obtained;
And acquiring a cross entropy loss function of the prediction probability according to the prediction probability of each category, and optimizing network parameters of the convolutional neural network according to the cross entropy loss function.
The diagnosis device based on the convolutional neural network and the multi-mode medical image comprises an acquisition module, a diagnosis module and a diagnosis module, wherein the acquisition module is used for acquiring training data, and the training data comprises medical image samples of multiple modes and diagnosis marking information corresponding to the medical image samples of the multiple modes; the optimizing module is used for optimizing the network parameters of the convolutional neural network according to the training data to obtain an optimized classification network based on the convolutional neural network; the classification module is used for acquiring medical images of multiple modes, inputting the medical images of multiple modes into a classification network based on a convolutional neural network so as to output the prediction probability of the diseases corresponding to the medical images of each mode, wherein the prediction probability is used for representing the probability of the diseases and the probability of the diseases not. Therefore, the technical problem that the computer automatic diagnosis method only utilizes a single-mode image in the existing method can be solved, the purpose of diagnosing the medical images in multiple modes comprehensively by utilizing a deep learning method is achieved, and the accuracy of disease diagnosis is improved.
In order to implement the above embodiment, the present invention also proposes a non-transitory computer readable storage medium having stored thereon a computer program which, when executed by a processor, implements the diagnostic method based on a convolutional neural network and a multi-modal medical image of the above embodiment.
In the description of the present specification, a description referring to terms "one embodiment," "some embodiments," "examples," "specific examples," or "some examples," etc., means that a particular feature, structure, material, or characteristic described in connection with the embodiment or example is included in at least one embodiment or example of the present application. In this specification, schematic representations of the above terms are not necessarily directed to the same embodiment or example. Furthermore, the particular features, structures, materials, or characteristics described may be combined in any suitable manner in any one or more embodiments or examples. Furthermore, the different embodiments or examples described in this specification and the features of the different embodiments or examples may be combined and combined by those skilled in the art without contradiction.
Furthermore, the terms "first," "second," and the like, are used for descriptive purposes only and are not to be construed as indicating or implying a relative importance or implicitly indicating the number of technical features indicated. Thus, a feature defining "a first" or "a second" may explicitly or implicitly include at least one such feature. In the description of the present application, the meaning of "plurality" means at least two, for example, two, three, etc., unless specifically defined otherwise.
Any process or method descriptions in flow charts or otherwise described herein may be understood as representing modules, segments, or portions of code which include one or more executable instructions for implementing specific logical functions or steps of the process, and additional implementations are included within the scope of the preferred embodiment of the present application in which functions may be executed out of order from that shown or discussed, including substantially concurrently or in reverse order from that shown or discussed, depending on the functionality involved, as would be understood by those reasonably skilled in the art of the embodiments of the present application.
Logic and/or steps represented in the flowcharts or otherwise described herein, e.g., a ordered listing of executable instructions for implementing logical functions, can be embodied in any computer-readable medium for use by or in connection with an instruction execution system, apparatus, or device, such as a computer-based system, processor-containing system, or other system that can fetch the instructions from the instruction execution system, apparatus, or device and execute the instructions. For the purposes of this description, a "computer-readable medium" can be any means that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device. More specific examples (a non-exhaustive list) of the computer-readable medium would include the following: an electrical connection (electronic device) having one or more wires, a portable computer diskette (magnetic device), a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber device, and a portable compact disc read-only memory (CDROM). In addition, the computer readable medium may even be paper or other suitable medium on which the program is printed, as the program may be electronically captured, via, for instance, optical scanning of the paper or other medium, then compiled, interpreted or otherwise processed in a suitable manner, if necessary, and then stored in a computer memory.
It is to be understood that portions of the present application may be implemented in hardware, software, firmware, or a combination thereof. In the above-described embodiments, the various steps or methods may be implemented in software or firmware stored in a memory and executed by a suitable instruction execution system. As with the other embodiments, if implemented in hardware, may be implemented using any one or combination of the following techniques, as is well known in the art: discrete logic circuits having logic gates for implementing logic functions on data signals, application specific integrated circuits having suitable combinational logic gates, programmable Gate Arrays (PGAs), field Programmable Gate Arrays (FPGAs), and the like.
Those of ordinary skill in the art will appreciate that all or a portion of the steps carried out in the method of the above-described embodiments may be implemented by a program to instruct related hardware, where the program may be stored in a computer readable storage medium, and where the program, when executed, includes one or a combination of the steps of the method embodiments.
In addition, each functional unit in the embodiments of the present application may be integrated in one processing module, or each unit may exist alone physically, or two or more units may be integrated in one module. The integrated modules may be implemented in hardware or in software functional modules. The integrated modules may also be stored in a computer readable storage medium if implemented in the form of software functional modules and sold or used as a stand-alone product.
The above-mentioned storage medium may be a read-only memory, a magnetic disk or an optical disk, or the like. While embodiments of the present application have been shown and described above, it will be understood that the above embodiments are illustrative and not to be construed as limiting the application, and that variations, modifications, alternatives and variations may be made to the above embodiments by one of ordinary skill in the art within the scope of the application.

Claims (3)

1. A diagnosis method based on a convolutional neural network and a multi-modal medical image is characterized by comprising the following steps:
Collecting training data, wherein the training data comprises medical image samples of multiple modes and diagnosis marking information corresponding to the medical image samples of the multiple modes;
optimizing the network parameters of the convolutional neural network according to the training data to obtain an optimized classification network based on the convolutional neural network;
Acquiring medical images of multiple modes, inputting the medical images of the multiple modes into the classification network based on the convolutional neural network so as to output the prediction probability of diseases corresponding to the medical images of each mode, wherein the prediction probability is used for representing the probability of the diseases and the probability of the diseases not;
The convolutional neural network-based classification network comprises a plurality of convolutional neural network-based feature extraction networks and a multi-layer perceptron, wherein each feature extraction network corresponds to a mode, network parameters of the convolutional neural network are optimized according to the training data, and the optimized convolutional neural network-based classification network is obtained, and the convolutional neural network-based classification network comprises:
Inputting the medical image samples of each mode in the medical image samples of the plurality of modes into a corresponding feature extraction network to perform feature extraction, wherein the extracted features are vectors;
the features of all modes are spliced into long vectors, the long vectors are input into the multi-layer perceptron, corresponding values are output, the corresponding values are calculated by using softmax, and the prediction probability of each category is obtained;
And acquiring a cross entropy loss function of the prediction probability according to the prediction probability of each category, and optimizing network parameters of the convolutional neural network according to the cross entropy loss function.
2. A diagnostic device based on a convolutional neural network and a multi-modal medical image, comprising:
the acquisition module is used for acquiring training data, wherein the training data comprises medical image samples of multiple modes and diagnosis marking information corresponding to the medical image samples of the multiple modes;
the optimizing module is used for optimizing the network parameters of the convolutional neural network according to the training data to obtain an optimized classification network based on the convolutional neural network;
The classification module is used for acquiring medical images of multiple modes, inputting the medical images of multiple modes into the classification network based on the convolutional neural network so as to output the prediction probability of diseases corresponding to the medical images of each mode, wherein the prediction probability is used for representing the probability of the diseases and the probability of the diseases not;
The classification network based on the convolutional neural network comprises a plurality of characteristic extraction networks based on the convolutional neural network and a multi-layer perceptron, wherein each characteristic extraction network corresponds to one mode, and the optimization module is specifically used for:
Inputting the medical image samples of each mode in the medical image samples of the plurality of modes into a corresponding feature extraction network to perform feature extraction, wherein the extracted features are vectors;
the features of all modes are spliced into long vectors, the long vectors are input into the multi-layer perceptron, corresponding values are output, the corresponding values are calculated by using softmax, and the prediction probability of each category is obtained;
And acquiring a cross entropy loss function of the prediction probability according to the prediction probability of each category, and optimizing network parameters of the convolutional neural network according to the cross entropy loss function.
3. A non-transitory computer readable storage medium having stored thereon a computer program, which when executed by a processor, implements a diagnostic method based on a convolutional neural network and a multimodal medical image as claimed in claim 1.
CN202111038996.1A 2021-09-06 2021-09-06 Diagnosis method and device based on convolutional neural network and multi-modal medical image Active CN113888470B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111038996.1A CN113888470B (en) 2021-09-06 2021-09-06 Diagnosis method and device based on convolutional neural network and multi-modal medical image

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111038996.1A CN113888470B (en) 2021-09-06 2021-09-06 Diagnosis method and device based on convolutional neural network and multi-modal medical image

Publications (2)

Publication Number Publication Date
CN113888470A CN113888470A (en) 2022-01-04
CN113888470B true CN113888470B (en) 2024-07-05

Family

ID=79008279

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111038996.1A Active CN113888470B (en) 2021-09-06 2021-09-06 Diagnosis method and device based on convolutional neural network and multi-modal medical image

Country Status (1)

Country Link
CN (1) CN113888470B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114511513B (en) * 2022-01-14 2024-06-28 清华大学 Brain aneurysm three-dimensional detection segmentation method based on deep convolutional neural network
CN115116594B (en) * 2022-06-06 2024-05-31 中国科学院自动化研究所 Method and device for detecting effectiveness of medical device
CN115018795B (en) * 2022-06-09 2023-04-07 北京医准智能科技有限公司 Method, device and equipment for matching focus in medical image and storage medium
CN116452851A (en) * 2023-03-17 2023-07-18 中山大学肿瘤防治中心(中山大学附属肿瘤医院、中山大学肿瘤研究所) Training method and device for disease classification model, terminal and readable storage medium
CN116705252B (en) * 2023-06-16 2024-05-31 脉得智能科技(无锡)有限公司 Construction method, image classification method, device and medium for prostate cancer diagnosis model

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109558498A (en) * 2018-11-07 2019-04-02 南京邮电大学 Multi-modal hash method based on deep learning
CN111444960A (en) * 2020-03-26 2020-07-24 上海交通大学 Skin disease image classification system based on multi-mode data input

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11361470B2 (en) * 2019-05-09 2022-06-14 Sri International Semantically-aware image-based visual localization
CN113158875B (en) * 2021-04-16 2022-07-01 重庆邮电大学 Image-text emotion analysis method and system based on multi-mode interaction fusion network

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109558498A (en) * 2018-11-07 2019-04-02 南京邮电大学 Multi-modal hash method based on deep learning
CN111444960A (en) * 2020-03-26 2020-07-24 上海交通大学 Skin disease image classification system based on multi-mode data input

Also Published As

Publication number Publication date
CN113888470A (en) 2022-01-04

Similar Documents

Publication Publication Date Title
CN113888470B (en) Diagnosis method and device based on convolutional neural network and multi-modal medical image
KR102125127B1 (en) Method of brain disorder diagnosis via deep learning
US10902588B2 (en) Anatomical segmentation identifying modes and viewpoints with deep learning across modalities
KR102177568B1 (en) Method for semi supervised reinforcement learning using data with label and data without label together and apparatus using the same
US20230342918A1 (en) Image-driven brain atlas construction method, apparatus, device and storage medium
CN112037171B (en) Multi-mode feature fusion-based multi-task MRI brain tumor image segmentation method
US11471096B2 (en) Automatic computerized joint segmentation and inflammation quantification in MRI
CN115830017B (en) Tumor detection system, method, equipment and medium based on image-text multi-mode fusion
CN112741651B (en) Method and system for processing ultrasonic image of endoscope
CN115359066B (en) Focus detection method and device for endoscope, electronic device and storage medium
CN114511513B (en) Brain aneurysm three-dimensional detection segmentation method based on deep convolutional neural network
CN109147927A (en) A kind of man-machine interaction method, device, equipment and medium
CN112634255A (en) Method and device for establishing brain focus detection model and computer equipment
CN115937192B (en) Unsupervised retina blood vessel segmentation method and system and electronic equipment
CN116883725A (en) Three-dimensional medical image disease category classification method and device
CN115089112B (en) Post-stroke cognitive impairment risk assessment model building method and device and electronic equipment
Salih et al. A review of evaluation approaches for explainable AI with applications in cardiology
KR102616961B1 (en) Method of providing disease information by domain adaptation between heterogeneous capsule endoscopes
CN112686912B (en) Acute stroke lesion segmentation method based on gradual learning and mixed samples
CN114974560A (en) Intelligent diagnosis device, equipment and storage medium for benign and malignant tumors
CN112258564B (en) Method and device for generating fusion feature set
CN114723710A (en) Method and device for detecting ultrasonic video key frame based on neural network
CN110147830B (en) Method for training image data generation network, image data classification method and device
CN114399462B (en) Medical image anomaly detection method and device based on depth characterization network
US11998318B2 (en) System and method of using visually-descriptive words to diagnose ear pathology

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant