WO2020259666A1 - 图像分类方法、装置、设备、存储介质和医疗电子设备 - Google Patents
图像分类方法、装置、设备、存储介质和医疗电子设备 Download PDFInfo
- Publication number
- WO2020259666A1 WO2020259666A1 PCT/CN2020/098407 CN2020098407W WO2020259666A1 WO 2020259666 A1 WO2020259666 A1 WO 2020259666A1 CN 2020098407 W CN2020098407 W CN 2020098407W WO 2020259666 A1 WO2020259666 A1 WO 2020259666A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- target
- feature
- image feature
- target image
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 86
- 238000012549 training Methods 0.000 claims description 136
- 238000013528 artificial neural network Methods 0.000 claims description 94
- 238000012545 processing Methods 0.000 claims description 56
- 230000008569 process Effects 0.000 claims description 23
- 239000013598 vector Substances 0.000 claims description 20
- 230000004927 fusion Effects 0.000 claims description 15
- 230000015654 memory Effects 0.000 claims description 14
- 238000003672 processing method Methods 0.000 claims description 9
- 238000001514 detection method Methods 0.000 claims description 7
- 238000003709 image segmentation Methods 0.000 claims description 3
- 210000000481 breast Anatomy 0.000 description 84
- 201000010099 disease Diseases 0.000 description 32
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 description 32
- 230000036541 health Effects 0.000 description 22
- 230000006870 function Effects 0.000 description 11
- 210000000056 organ Anatomy 0.000 description 9
- 206010006187 Breast cancer Diseases 0.000 description 8
- 208000026310 Breast neoplasm Diseases 0.000 description 8
- 238000002595 magnetic resonance imaging Methods 0.000 description 7
- 238000010586 diagram Methods 0.000 description 6
- 238000012216 screening Methods 0.000 description 6
- 238000009607 mammography Methods 0.000 description 5
- 230000002146 bilateral effect Effects 0.000 description 4
- 238000002591 computed tomography Methods 0.000 description 4
- 230000000694 effects Effects 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 4
- 238000002603 single-photon emission computed tomography Methods 0.000 description 4
- 241001465754 Metazoa Species 0.000 description 3
- 238000002059 diagnostic imaging Methods 0.000 description 3
- 238000003384 imaging method Methods 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 238000002604 ultrasonography Methods 0.000 description 3
- 241000282412 Homo Species 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 210000001508 eye Anatomy 0.000 description 2
- 210000004072 lung Anatomy 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 238000002600 positron emission tomography Methods 0.000 description 2
- 238000005070 sampling Methods 0.000 description 2
- 210000000515 tooth Anatomy 0.000 description 2
- 230000008859 change Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
- G06V10/443—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by matching or filtering
- G06V10/449—Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters
- G06V10/451—Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters with interaction between the filter responses, e.g. cortical complex cells
- G06V10/454—Integrating the filters into a hierarchical structure, e.g. convolutional neural networks [CNN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2415—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4038—Image mosaicing, e.g. composing plane images from plane sub-images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
- G06T7/0014—Biomedical image inspection using an image reference approach
- G06T7/0016—Biomedical image inspection using an image reference approach involving temporal comparison
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/80—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
- G06V10/806—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
- G06T2207/20221—Image fusion; Image merging
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30036—Dental; Teeth
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30041—Eye; Retina; Ophthalmic
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30061—Lung
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30068—Mammography; Breast
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/03—Recognition of patterns in medical or anatomical images
Definitions
- This application relates to the field of image processing, in particular to an image classification method, device, equipment, storage medium, medical electronic equipment, and an image processing method.
- Image classification refers to automatically classifying input images into a set of predefined categories according to certain classification rules. For example, according to the semantic information contained in the image, the input image can be classified into objects and scenes. For example, it is possible to identify preset target objects contained in the input image and classify them according to the identified objects. For another example, images with similar content can also be classified into the same category according to semantic information in the input image.
- the purpose of this application is to provide an image classification method, device, equipment, storage medium, medical electronic equipment, and image processing method.
- an image classification method executed by an electronic device, the method comprising: receiving a target image and a reference image about the target image, wherein the target image is a medical image; using the same Method to determine the first image feature of the target image and the second image feature of the reference image; fuse the first image feature and the second image feature to determine the image feature to be classified; and use the image feature to be classified.
- the step of determining the probability that the target image belongs to a preset category includes: using the features of the image to be classified to obtain a vector of multiple dimensions, and the elements in the vector respectively indicate that the target image and the reference image belong to the preset category.
- the confidence score of the category determines the probability that the target image belongs to the preset category according to the confidence score that the target image belongs to the preset category.
- an image classification device including: a receiving unit configured to receive a target image and a reference image related to the target image, wherein the target image is a medical image; and an image feature determination unit , Configured to use the same method to determine the first image feature of the target image and the second image feature of the reference image; a fusion unit configured to fuse the first image feature and the second image feature to determine the to-be-classified Image features; and a classification result generating unit configured to use the features of the image to be classified to determine the probability that the target image belongs to a preset category, wherein the classification result generating unit is configured to use the features of the image to be classified to obtain multiple dimensions
- the elements in the vector respectively represent the confidence scores that the target image and the reference image belong to a preset category. According to the confidence scores that the target image belongs to the preset category, it is determined that the target image belongs to the preset category Probability.
- an image processing method executed by an electronic device comprising: receiving a target image; using a first neural network to determine a first target image feature of the target image; Second, the neural network determines the second target image feature of the target image; determines the first image processing result and the second image processing result of the target image according to the first target image feature and the second target image feature; fusion The first image processing result and the second image processing result determine the image processing result of the target image, wherein the first neural network and the second neural network are different networks trained by the same training method
- the first neural network is trained using a first training set
- the second neural network is trained using a second training set
- the positive training images included in the first training set and the second training set are The negative sample ratio is different.
- a medical electronic device including: an image acquisition unit configured to acquire a target image and a reference image related to the target image, wherein the target image is a medical image; image feature determination Unit configured to determine the first image feature of the target image and the second image feature of the reference image in the same manner; the fusion unit configured to fuse the first image feature and the second image feature to determine the Classification image features; and a classification result generation unit configured to determine the probability that the target image belongs to a preset category using the characteristics of the image to be classified, wherein the classification result generation unit is configured to use the characteristics of the image to be classified to obtain a plurality of A vector of dimensions, the elements in the vector respectively represent the confidence scores that the target image and the reference image belong to a preset category, and according to the confidence scores that the target image belongs to the preset category, it is determined that the target image belongs to the preset category The probability.
- an image classification device including a memory and a processor, wherein the memory stores instructions, and when the instructions are executed by the processor, the The processor executes the image classification method as described above.
- a computer-readable storage medium having instructions stored thereon, and when the instructions are executed by a processor, the processor executes the image classification method described above.
- Fig. 1 shows an exemplary scene diagram of the image processing system according to the present application
- Fig. 2 shows a schematic flowchart of an image classification method according to an embodiment of the present application
- Figure 3A shows a schematic process of the image classification method according to the present application
- Figure 3B shows the cc-bit image and mlo-bit image of the left breast of the human body, and the cc-bit image and mlo-bit image of the right breast;
- Fig. 4 shows a schematic flowchart of an image processing method according to an embodiment of the present application
- Fig. 5 shows a schematic block diagram of an image classification device according to an embodiment of the present application
- Fig. 6 shows a schematic block diagram of a medical electronic device according to an embodiment of the present application.
- Fig. 7 shows the architecture of a computing device according to an embodiment of the present application.
- processing is usually performed only for the target image of interest.
- image classification processing is usually performed only for the breast images on the side of interest.
- the medical images of different individuals differ greatly in physiological characteristics such as tissue density and fat thickness, the visual effects of the breast medical images obtained for different individuals are also very different. Therefore, it is difficult to make accurate judgments if only image processing is performed on the medical image on the side of interest.
- the comparison information of the medical images of the bilateral organs of the same person is considered when performing image classification, more accurate image classification results and disease screening results can be obtained.
- Fig. 1 shows an exemplary scene diagram of the image classification system according to the present application.
- the image classification system 100 may include a user terminal 110, a network 120, a server 130 and a database 140.
- the user terminal 110 may be, for example, the computer 110-1 and the mobile phone 110-2 shown in FIG. 1. It is understandable that, in fact, the user terminal can be any other type of electronic device that can perform data processing, which can include but is not limited to desktop computers, laptops, tablets, smart phones, smart home devices, wearable devices, Car electronic equipment, monitoring equipment, etc.
- the user terminal may also be any equipment provided with electronic equipment, such as vehicles, robots, and so on.
- the user terminal provided according to the present application can be used to receive images to be processed, and use the method provided in the present application to realize image classification, thereby realizing disease screening.
- the user terminal may collect the image to be processed through an image acquisition device (such as a camera, a video camera, etc.) provided on the user terminal.
- the user terminal may also receive the image to be processed from an independently set image acquisition device.
- the user terminal may also receive the image to be processed from the server via the network.
- the image to be processed described here can be a single image or a frame in the video.
- the user terminal may also receive the image to be processed from the medical acquisition device.
- the medical image mentioned here can be, for example, CT (Computed Tomography), MRI (Magnetic Resonance Imaging, magnetic resonance imaging), ultrasound, X-ray, radionuclide imaging (such as SPECT (Single-Photon Emission computed Tomography) Medical images collected by methods such as single photon emission computed tomography, PET (Positron Emission Tomography, positron emission tomography), etc., can also be images that display human physiological information such as electrocardiogram, electroencephalogram, and optical photography.
- the processing unit of the user terminal may be used to execute the image classification method provided in this application.
- the user terminal may use an application built in the user terminal to execute the image classification method.
- the user terminal may execute the image classification method provided in this application by calling an application program stored externally of the user terminal.
- the user terminal sends the received image to be processed to the server 130 via the network 120, and the server 130 executes the image classification method.
- the server 130 may use an application built in the server to execute the image classification method.
- the server 130 may execute the image classification method by calling an application program stored outside the server.
- the network 120 may be a single network, or a combination of at least two different networks.
- the network 120 may include, but is not limited to, one or a combination of several of a local area network, a wide area network, a public network, and a private network.
- the server 130 may be a single server or a server group, and each server in the group is connected through a wired or wireless network.
- a server group can be centralized, such as a data center, or distributed.
- the server 130 may be local or remote.
- the database 140 can generally refer to a device having a storage function.
- the database 140 is mainly used to store various data used, generated and output from the work of the user terminal 110 and the server 130.
- the database 140 may be local or remote.
- the database 140 may include various memories, such as Random Access Memory (RAM), Read Only Memory (ROM), etc.
- RAM Random Access Memory
- ROM Read Only Memory
- the storage devices mentioned above are just a few examples, and the storage devices that can be used in the system are not limited to these.
- the database 140 may be connected or communicated with the server 130 or a part thereof via the network 120, or directly connected or communicated with the server 130, or a combination of the above two methods.
- the database 140 may be a stand-alone device. In other embodiments, the database 140 may also be integrated in at least one of the user terminal 110 and the server 130. For example, the database 140 may be set on the user terminal 110 or on the server 130. For another example, the database 140 may also be distributed, a part of which is set on the user terminal 110 and the other part is set on the server 130.
- Fig. 2 shows a schematic flowchart of an image processing method according to an embodiment of the present application.
- a target image and at least one reference image about the target image may be received.
- the reference image may be an image of the same type as the target image.
- the reference image may contain other target objects that are the same or the same type as the target object in the target image.
- the target image may be a medical image.
- the reference image may be a similar medical image of an organ on the other side of the same person.
- the organs mentioned here can be any organs with two or more numbers in the human body such as breast, eyes, lungs, and teeth.
- the target image mentioned here can also be any other type of image except medical images, as long as the reference image can contain the target object in the target image.
- Target objects of the same type are sufficient.
- the target image may be a face image.
- the reference image may be a face image taken at other times (for example, in a different background, under different lighting, and at different ages).
- the target image can be any animal or plant except humans.
- the target image may include multiple images for the same target object.
- the target image may include images of the target object obtained from at least two angles.
- the target image may include a mammography detection image of the human breast taken at a CC (Craniocaudal) perspective and a mammography detection image of the human breast taken at a Mediolateral-Oblique (MLO) perspective.
- MLO Mediolateral-Oblique
- the target image may also include target objects acquired by at least two different devices.
- the target image may include an image of the human breast acquired by an X-ray device and an image of the human breast acquired by an MRI device. It is understandable that when the target image includes other types of target objects, those skilled in the art can also arbitrarily set the manner of shooting the target object to obtain other target images obtained by different devices. For example, it is also possible to obtain a target image containing the target object through at least two cameras with different parameters.
- the reference image may include multiple reference images, respectively corresponding to each of the multiple target images. Each reference image is obtained in the same way as the corresponding target image.
- the first image feature of the target image and the second image feature of the reference image may be determined.
- the first image feature of the target image and the second image feature of the reference image can be determined in the same manner.
- a first neural network including at least one convolution layer may be used to perform convolution processing on the target image to obtain the first image feature. Further, the first neural network may be used to perform convolution processing on the reference image to obtain a second image feature. In other words, a neural network sharing parameters can be used to process the target image and the reference image.
- the first image feature and the second image feature may each include a plurality of image features.
- the first neural network may be any neural network that can obtain image features from an image.
- the first neural network may be any network including at least one convolutional layer, such as any one of the Inception series network (such as Googlenet, etc.), the VGG series network, the Resnet series network, etc., or at least a part of any one of the foregoing networks.
- At least one of the color feature, texture feature, shape feature, spatial relationship feature and other features in the target image can also be extracted as the first image feature. Further, the same method can be used to extract the features in the reference image as the second image feature.
- step S206 the first image feature and the second image feature may be combined to determine the image feature to be classified.
- the first image feature and the second image feature may be spliced to determine the image feature to be classified.
- the first image feature and the second image feature may have multiple channels.
- the first image feature can be a tensor of size H1*W1*C1, where H1 can be the size of the first image feature in the first direction (for example, the length direction), and W1 can be the size of the first image feature in the second
- H1 and W1 may be the size in units of the number of pixels
- C1 may be the number of channels of the first image feature.
- the second image feature can be a tensor of size H2*W2*C2, where H2 can be the size of the second image feature in the first direction (for example, the length direction), and W2 can be the second image feature in the second direction ( For example, the size in the width direction, H2 and W2 may be the size in units of the number of pixels, and C2 may be the number of channels of the second image feature. Wherein C1, C2 are integers greater than 1.
- fusing the first image feature and the second image feature to determine the image feature to be classified may include splicing the first image feature and the second image feature to determine the image feature to be classified .
- the first image feature may have 1024 channels
- the second image feature may also have 1024 channels.
- a 2048 channel image feature to be classified can be obtained.
- the image feature to be classified with 2048 channels is fused with the information of the first image feature and the second image feature, that is, the image information of the target image and the reference image are fused.
- each element at the corresponding position in the corresponding channel of the first image feature and the second image feature may also be added to determine the image feature to be classified into which the image information of the target image and the reference image are fused.
- step S208 the feature of the image to be classified obtained in step S206 may be used to determine the probability that the target image belongs to a preset category. For example, the fused image features to be classified are used to predict whether the left and right breasts are positive for breast cancer.
- the first fully connected network may be used to process the features of the image to be classified to obtain the probability that the target image belongs to a preset category.
- the first fully connected network may be configured such that the first fully connected network outputs a vector with multiple dimensions, and each element in the vector represents the confidence score that the target image and the reference image belong to a preset category.
- the preset categories may include health categories and disease categories.
- the target image may be a medical image of the breast on the left side of the human body
- the reference image may be a medical image of the breast on the right side of the human body.
- the vector output by the first fully connected network can be 4-dimensional. The four-dimensional elements of the vector respectively represent the confidence scores that the medical image of the left breast belongs to the health category and the disease category, and the medical image of the right breast belongs to the health category and the disease category.
- the preset category may be any possible classification categories such as predefined animal categories, plant categories, and scene categories.
- the first fully connected network can be configured to output vectors of predetermined dimensions. For example, when the number of input images is m and the number of preset categories is n, the vector output by the first fully connected network may have m n dimensions.
- the probability that the target image and the reference image belong to a preset category may be determined according to the confidence scores for multiple dimensions output by the first fully connected network.
- the softmax function can be used to normalize the two confidence scores for the left breast to obtain the medical image of the left breast that belongs to the health category. The probability and the probability that the medical image of the left breast belongs to the disease category. Similarly, the softmax function can be used to obtain the probability that the medical image of the right breast belongs to the healthy category and the probability that the medical image of the right breast belongs to the disease category.
- the target image when the probability that the target image belongs to the preset category is greater than the preset probability threshold (such as 0.5), the target image can be considered to belong to the preset category.
- the preset probability threshold such as 0.5
- the target image can be classified according to the image information of the reference image.
- the image information of the reference image and the target image can be fused during the image classification process, and the target image can be determined to belong to the predetermined image based on the image characteristics of the image information of the fused target image and the reference image. Set the probability of the category to achieve a more accurate classification of the target image.
- the target image and the reference image are medical images, the accuracy of disease screening is improved.
- the above-mentioned image classification method provided by the embodiments of the present application when a breast cancer patient has breast cancer and the contrast information of bilateral breast images is significantly different, it can be determined that one side of the person is positive for breast cancer with a high probability.
- Fig. 3A shows a schematic process of the image classification method according to the present application.
- the input image may include the cc-bit image and the mlo-bit image of the left breast of the human body, and the cc-bit image and the mlo-bit image of the right breast.
- Figure 3B shows the cc-bit image and mlo-bit image of the left breast of the human body, and the cc-bit image and mlo-bit image of the right breast.
- the above four input images can be respectively input to a googlenet network with shared parameters, so as to obtain the respective vectorized image characteristics of each image.
- the output result of any layer in the googlenet network (such as a certain later layer, such as the pool5/7 ⁇ 7_s1 layer in the googlenet network) can be selected as the vectorized image feature of the image.
- the pool5/7 ⁇ 7_s1 layer in the googlenet network can be used to generate a 1024-dimensional vectorized feature for each input image.
- a 4096-dimensional fusion feature F [f cc l ,f mlo l ,f cc R ,f mlo R ].
- a 4-dimensional vector can be obtained, representing the confidence scores s + l , s - l , s + R of the left breast and right breast respectively belonging to the health category and the disease category. , S - R .
- the softmax function can be used to normalize s + l and s - l to obtain the probability p + l and p - l that the left breast belongs to the health category and the disease category.
- the softmax function can be used to normalize s + R and s - R to obtain p + R and p - R of the probability that the right breast belongs to the health category and the disease category, where exp() is the exponential operation function .
- the probability that the target image and the reference image belong to the health category and the disease category can be obtained using the process shown in FIG. 3A.
- the probability p + l that the medical image of the left breast belongs to the healthy category is greater than the preset probability threshold (such as 0.5), it can be considered that the medical image of the left breast belongs to the healthy category.
- the right breast can be determined The category to which the medical image belongs.
- the image classification method provided by the embodiments of the present application can be used to identify whether the mammography scan of a patient suspected of breast cancer contains breast cancer. For example, based on the network structure used for the comparison of the left and right breasts, the multi-view scans of the left and right breasts can be received at the same time, and the features of each scan of each breast can be extracted separately, and then feature fusion is performed, and the fused features It is also used to predict whether the left and right breasts are positive for breast cancer.
- the first neural network and the first fully connected network may be trained in the following manner.
- a first training set of the first neural network may be determined, wherein the first training set includes a plurality of training images.
- the multiple training images are images of the same type as the target image.
- the multiple training images included in the first training set may be different breasts that are known to conform to the preset type.
- the first training set may include a first training image.
- a first reference training image for the first training image can be determined.
- the first training image is a medical image of the left breast
- the first reference training image may be a medical image of the right breast.
- the first neural network may be used to perform convolution processing on the first training image and the first reference training image to obtain the first training image feature and the second training image feature.
- the training image feature to be classified may be determined according to the first training image feature and the second training image feature.
- the feature of the training image to be classified is fused with image information of the first training image and the first training reference image.
- a first fully connected network may be used to process the characteristics of the training image to be classified to determine the probability that the first training image belongs to a preset category.
- the parameters of the first neural network and the first fully connected network can be adjusted so that the probability that the first training image belongs to the preset category and the The loss between the real categories to which the first training image belongs is the smallest.
- the cross-entropy loss function may be used to calculate the loss between the probability that the first training image belongs to the preset category and the real category to which the first training image belongs.
- the first fully connected network can output the probability that the left breast image belongs to the health category and the disease category and the right breast image belongs to the health category and the disease category.
- the loss of the left breast image can be calculated according to the following formula:
- the loss of the right breast image can be calculated according to the following formula:
- the first neural network and the first fully connected network are trained using multiple known types of bilateral breast images, the contrast information of the bilateral breast images can be learned during the training process. Therefore, even In the case of large differences in tissue density, fat thickness and other aspects of breast scans of different individuals, the first neural network and first fully connected network after training can also be based on the breast scans to determine whether there is breast cancer. Stable and accurate judgment results.
- the proportion of positive samples and negative samples in training data is not balanced. For example, taking medical images as an example, there are fewer samples belonging to the disease category in the training images, but more samples belonging to the health category, so there is a problem of imbalanced samples.
- the same method can be used to train at least two network models with different parameters (such as the aforementioned first neural network and first fully connected network) by using sample sets that include different proportions of positive samples and negative samples.
- Network model to process the input image and determine the probability that the input image belongs to the preset category according to the output results of different networks.
- the input image includes a target image and a reference image of the target image.
- the first neural network is trained using a first training set
- the second neural network is trained using a second training set
- the positive values in the training images included in the first training set and the second training set are The negative sample ratio is different.
- Fig. 4 shows a flowchart of a graph processing method according to an embodiment of the present application.
- a target image can be received.
- the target image may be a medical image or any other type of image.
- the first neural network may be used to determine the first target image feature of the target image.
- a first neural network including at least one convolutional layer may be used to perform convolution processing on the target image to obtain the first target image feature.
- the first target image feature may include multiple image features.
- a second neural network may be used to determine the second target image feature of the target image.
- a second neural network including at least one convolutional layer may be used to perform convolution processing on the target image to obtain the second target image feature.
- the second image target features may respectively include a plurality of image features.
- the first neural network and the second neural network may be any neural networks that can obtain image features from an image.
- the first neural network may be any network including at least one convolutional layer, such as any one of the Inception series network (such as Googlenet, etc.), the VGG series network, the Resnet series network, etc., or at least a part of any one of the foregoing networks.
- the first neural network and the second neural network are different networks trained using the same training method, the first neural network is trained using the first training set, and the second neural network The network is trained using the second training set, and the ratios of positive and negative samples in the training images included in the first training set and the second training set are different.
- step S408 the first image processing result and the second image processing result of the target image may be determined according to the first target image feature and the second target image feature.
- the first image processing result and the second image processing result are of the same type.
- the first image processing result and the second image processing result may be at least one of an image classification result, an image segmentation result, and a target detection result. This depends on the specific method and training set used by those skilled in the art to train the neural network.
- step S410 the first image processing result and the second image processing result are merged to determine the image processing result of the target image.
- the image processing result includes at least one of an image classification result, an image segmentation result, and a target detection result.
- the proportion of positive samples and negative samples that make the network's positive sample loss and negative sample loss close can be determined through experiments. For example, by calculating the loss function of all positive samples and the loss function of all negative samples in the training set, the positive sample loss and the negative sample loss of the network trained by the training set can be determined. In the following, when the ratio of positive samples to negative samples is 1:2, the network's positive sample loss and negative sample loss are similar as an example to describe the principle of this application.
- a first training set with a ratio of positive samples to negative samples of 1:1 and a second training set with a ratio of positive samples to negative samples of 1:3 can be determined by sampling.
- the ratio of positive samples to negative samples is 1:2
- the loss of positive samples and negative samples of the network are similar. Therefore, when the ratio of positive and negative samples in the training set is changed, the training generated The ratio of positive sample loss and negative sample loss of the network can also be changed accordingly. Therefore, since the positive sample ratio is increased in the first training set whose ratio of positive samples to negative samples is 1:1, the loss of positive samples of the network trained in the first training set is smaller than the loss of negative samples. Similarly, since the second training set with a ratio of positive samples to negative samples of 1:3 increases the ratio of negative samples, the loss of positive samples of the network trained in the second training set is greater than the loss of negative samples.
- the network trained in the first training set has a better classification effect on positive samples
- the network trained in the second training set has a better classification effect on negative samples.
- the output of the network trained by the first training set indicates that the probability of the input image belonging to the preset category is closer to 0 or 1, that is It is easier to distinguish whether the input image belongs to a preset category.
- the network trained in the first training set since the network trained in the first training set has a poor classification effect on negative samples, for input images of negative samples, the network trained in the first training set will output a probability closer to 0.5, that is, it is less easy to distinguish the input The category to which the image belongs. Based on the above characteristics, by fusing the output results of the network trained in the first training set and the network trained in the second training set, more accurate prediction results can be obtained.
- the first training set with a ratio of positive samples to negative samples of 1:1, the second training set with a ratio of positive samples to negative samples of 1:3, and the ratio of positive samples to negative samples of 1:2 can also be determined by sampling.
- the first training set, the second training set, and the third training set can be used to train the first neural network, the second neural network, and the third neural network, respectively.
- the probability that the left breast medical image and the right breast medical image belong to the health category and the disease category can be determined by fusing (for example, weighted average) the above three network output results obtained by using different training sets. For example, the probability that the left breast medical image and the right breast medical image belong to the health category and the disease category can be determined by the following formula.
- Is the probability that the breast medical image on the left belongs to the health category Is the probability that the left breast medical image belongs to the disease category
- Is the probability that the breast medical image on the right belongs to the health category Is the probability that the breast medical image on the right belongs to the health category
- first target reference image feature and the second target reference image feature of the reference image can be obtained through steps similar to steps S402-S410.
- the first image feature includes a first target image feature and a second target image feature.
- the first target image feature is obtained by convolution processing the target image using a first neural network including at least one convolutional layer
- the second target image feature is obtained by using a second neural network to perform convolution processing on the target image. It is obtained by convolution processing.
- the second image feature includes a first target reference image feature and a second target reference image feature, wherein the first target reference image feature uses a first neural network including at least one convolutional layer to convolve the reference image
- the feature of the second target reference image is obtained by performing convolution processing on the reference image by using the second neural network.
- the first neural network and the second neural network are different networks trained using the same training method, the first neural network is trained using the aforementioned first training set, and the second neural network is trained using the Obtained from the aforementioned second training set.
- the proportions of positive and negative samples in the training images included in the first training set and the second training set are different. Therefore, the first neural network and the second neural network have different parameters and therefore have different outputs. result.
- the image features to be classified include a first image feature to be classified and a second image feature to be classified.
- the first image feature to be classified may be determined by splicing the first target image feature and the first target reference image feature
- the second image feature to be classified may be determined by splicing the second target image feature.
- the image feature and the second target reference image feature are determined.
- the first fully connected network may be used to process the features of the first image to be classified to obtain the first probability that the target image belongs to a preset category.
- a second fully connected network may be used to process the features of the second image to be classified to obtain the second probability that the target image belongs to a preset category.
- the probability that the target image belongs to the preset category can be determined by fusing the first probability and the second probability. For example, the probability that the target image belongs to the preset category may be determined according to the weighted average of the first probability and the second probability.
- the first fully connected network and the second fully connected network are different networks trained using the same training method, wherein the first fully connected network is trained using the first training set, and the second fully connected network
- the connection network is trained by using the second training set, and the ratio of positive and negative samples in the training images included in the first training set and the second training set are different.
- Fig. 5 shows a schematic block diagram of an image classification device according to an embodiment of the present application.
- the image classification device 500 may include a receiving unit 510, an image feature determination unit 520, a fusion unit 530, and a classification result generation unit 540.
- the receiving unit 510 may be configured to receive a target image and at least one reference image regarding the target image.
- the reference image may be the same type of image as the target image.
- the reference image may contain the same target object or other target objects of the same type as in the target image.
- the target image may be a medical image.
- the reference image may be a similar medical image of an organ on the other side of the same person.
- the organs mentioned here can be any organs with two or more numbers in the human body such as breast, eyes, lungs, and teeth.
- the target image mentioned here can also be any other type of image except medical images, as long as the reference image can contain the same type as the target image.
- the target audience can be.
- the target image may be a face image.
- the reference image may be a face image taken at other times (for example, in a different background, under different lighting, and at different ages).
- the target image can be any animal or plant except humans.
- the target image may include multiple images for the same target object.
- the target image may include images of the target object obtained from at least two angles.
- the target image may include a mammography detection image of the human breast taken at a CC (Craniocaudal) perspective and a mammography detection image of the human breast taken at a Mediolateral-Oblique (MLO) perspective.
- MLO Mediolateral-Oblique
- the target image may also include target objects acquired by at least two different devices.
- the target image may include an image of the human breast acquired by an X-ray device and an image of the human breast acquired by an MRI device. It is understandable that when the target image includes other types of target objects, those skilled in the art can also arbitrarily set the manner of shooting the target object to obtain other target images obtained by different devices. For example, it is also possible to obtain a target image containing the target object through at least two cameras with different parameters.
- the reference image may include multiple reference images, respectively corresponding to each of the multiple target images. Each reference image is obtained in the same way as the corresponding target image.
- the image feature determining unit 520 may be configured to determine the first image feature of the target image and the second image feature of the reference image. For example, the first image feature of the target image and the second image feature of the reference image can be determined in the same manner.
- a first neural network including at least one convolution layer may be used to perform convolution processing on the target image to obtain the first image feature. Further, the first neural network may be used to perform convolution processing on the reference image to obtain a second image feature. In other words, a neural network sharing parameters can be used to process the target image and the reference image.
- the first image feature and the second image feature may each include a plurality of image features.
- the first neural network may be any neural network that can obtain image features from an image.
- the first neural network may be any network including at least one convolutional layer, such as any one of the Inception series network (such as Googlenet, etc.), the VGG series network, the Resnet series network, etc., or at least a part of any one of the foregoing networks.
- At least one of the color feature, texture feature, shape feature, spatial relationship feature and other features in the target image can also be extracted as the first image feature. Further, the same method can be used to extract the features in the reference image as the second image feature.
- the first image feature includes a first target image feature and a second target image feature.
- the first target image feature is obtained by convolution processing the target image using a first neural network including at least one convolutional layer
- the second target image feature is obtained by using a second neural network to perform convolution processing on the target image. It is obtained by convolution processing.
- the first target image feature and the second target image feature may each include a plurality of image features.
- the second image feature includes a first target reference image feature and a second target reference image feature, wherein the first target reference image feature uses a first neural network including at least one convolutional layer to convolve the reference image
- the feature of the second target reference image is obtained by performing convolution processing on the reference image by using the second neural network.
- the first neural network and the second neural network are different networks trained using the same training method, the first neural network is trained using the aforementioned first training set, and the second neural network is trained using the Obtained from the aforementioned second training set.
- the fusion unit 530 may be configured to fuse the first image feature and the second image feature to determine the image feature to be classified.
- the first image feature and the second image feature may be spliced to determine the image feature to be classified.
- fusing the first image feature and the second image feature to determine the image feature to be classified may include splicing the first image feature and the second image feature to determine the image feature to be classified.
- the image feature to be classified includes a first image feature to be classified and a second image feature to be classified.
- the first image feature to be classified may be determined by splicing the first target image feature and the first target reference image feature
- the second image feature to be classified may be determined by splicing the second target image feature.
- the image feature and the second target reference image feature are determined.
- the classification result generating unit 540 may be configured to determine the probability that the target image belongs to a preset category by using the features of the image to be classified generated by the fusion unit 530.
- the first fully connected network may be used to process the features of the image to be classified to obtain the probability that the target image belongs to a preset category.
- the first fully connected network may be configured such that the first fully connected network outputs a vector with multiple dimensions, and each element in the vector represents the confidence score that the target image and the reference image belong to a preset category.
- the probability that the target image and the reference image belong to a preset category may be determined according to the confidence scores for multiple dimensions output by the first fully connected network.
- the softmax function can be used to normalize the two confidence scores for the left breast to obtain the medical image of the left breast that belongs to the health category. The probability and the probability that the medical image of the left breast belongs to the disease category. Similarly, the softmax function can be used to obtain the probability that the medical image of the right breast belongs to the healthy category and the probability that the medical image of the right breast belongs to the disease category.
- the target image when the probability that the target image belongs to the preset category is greater than the preset probability threshold (such as 0.5), the target image can be considered to belong to the preset category.
- the preset probability threshold such as 0.5
- the classification result generating unit 540 may also use the first fully connected network to process the first image feature to be classified to obtain the first probability that the target image belongs to the preset category.
- a second fully connected network may be used to process the features of the second image to be classified to obtain a second probability that the target image belongs to a preset category.
- the probability that the target image belongs to a preset category can be determined by fusing the first probability and the second probability. For example, the probability that the target image belongs to a preset category may be determined according to a weighted average of the first probability and the second probability.
- the target image can be classified according to the image information of the reference image.
- the image information of the reference image and the target image can be merged in the image classification process, and the target image can be determined as belonging to the preset category according to the image characteristics of the image information of the fused target image and the reference image. Probability, so as to achieve a more accurate classification of the target image.
- the target image and the reference image are medical images, the accuracy of disease screening is improved.
- it can also overcome the problem of unbalanced training data proportion in related technologies, and further improve the accuracy of image classification and the accuracy of disease screening.
- Fig. 6 shows a schematic block diagram of a medical electronic device according to an embodiment of the present application.
- the medical electronic device 600 may include an image acquisition unit 610, an image feature determination unit 620, a fusion unit 630, and a classification result generation unit 640.
- the image acquisition unit 610 may be used to acquire a target image and a reference image related to the target image.
- the medical images mentioned here can be, for example, medical images collected by CT, MRI, ultrasound, X-ray, radionuclide imaging (such as SPECT, PET), etc., or can be displays such as electrocardiogram, electroencephalogram, optical photography, etc. Images of human body physiological information.
- the image feature determination unit 620, the fusion unit 630, and the classification result generation unit 640 may be implemented as the image feature determination unit 520, the fusion unit 530, and the classification result generation unit 540 shown in FIG. 5, which will not be repeated here.
- the medical electronic equipment provided in this application may be any medical imaging equipment such as CT, MRI, ultrasound, X-ray equipment.
- the image acquisition unit 610 may be implemented as the imaging unit of the above-mentioned medical imaging device, and the image feature determination unit 620, the fusion unit 630, and the classification result generation unit 640 may be implemented by an internal processing unit (for example, a processor) of the medical imaging device.
- FIG. 7 shows the architecture of the computing device.
- the computing device 700 may include a bus 710, one or at least two CPUs 720, a read only memory (ROM) 730, a random access memory (RAM) 740, a communication port 750 connected to a network, input/output Components 760, hard disk 770, etc.
- the storage device in the computing device 700 such as a ROM 730 or a hard disk 770, can store various data or files used in the processing and/or communication of the method for detecting targets in a video provided by this application, and programs executed by the CPU. instruction.
- the computing device 700 may also include a user interface 780.
- the architecture shown in FIG. 7 is only exemplary. When implementing different devices, one or at least two components in the computing device shown in FIG. 7 may be omitted according to actual needs.
- the embodiments of the present application can also be implemented as a computer-readable storage medium.
- the computer-readable storage medium stores computer-readable instructions.
- the computer-readable instructions are executed by the processor, the method according to the embodiments of the present application described with reference to the above drawings can be executed.
- the computer-readable storage medium includes, but is not limited to, for example, volatile memory and/or non-volatile memory.
- the volatile memory may include random access memory (RAM) and/or cache memory (cache), for example.
- the non-volatile memory may include, for example, read-only memory (ROM), hard disk, flash memory, etc.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Artificial Intelligence (AREA)
- Data Mining & Analysis (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- General Engineering & Computer Science (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- Medical Informatics (AREA)
- Molecular Biology (AREA)
- Biomedical Technology (AREA)
- Multimedia (AREA)
- Mathematical Physics (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Databases & Information Systems (AREA)
- Quality & Reliability (AREA)
- Radiology & Medical Imaging (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Biodiversity & Conservation Biology (AREA)
- Probability & Statistics with Applications (AREA)
- Image Analysis (AREA)
- Apparatus For Radiation Diagnosis (AREA)
Abstract
Description
Claims (15)
- 一种图像分类方法,由电子设备执行,所述方法包括:接收目标图像以及关于所述目标图像的参考图像,其中所述目标图像是医学图像;采用相同的方式确定所述目标图像的第一图像特征和所述参考图像的第二图像特征;融合所述第一图像特征和所述第二图像特征确定待分类图像特征;以及利用所述待分类图像特征确定所述目标图像属于预设类别的概率,该步骤包括:利用所述待分类图像特征,获得多个维度的向量,该向量中的元素分别表示所述目标图像和所述参考图像属于预设类别的置信分数,根据所述目标图像属于预设类别的置信分数,确定所述目标图像属于预设类别的概率。
- 根据权利要求1所述的图像分类方法,其中,确定所述目标图像的第一图像特征和所述参考图像的第二图像特征包括:利用包括至少一个卷积层的第一神经网络对所述目标图像进行卷积处理以得到第一图像特征;以及利用所述第一神经网络对所述参考图像进行卷积处理以得到第二图像特征。
- 根据权利要求1所述的方法,其中,融合所述第一图像特征和所述第二图像特征确定待分类图像特征包括:拼接所述第一图像特征和所述第二图像特征,以确定所述待分类图像特征。
- 根据权利要求2所述的图像分类方法,其中,利用所述待分类图像特征确定所述目标图像属于预设类别的概率包括:利用第一全连接网络对所述待分类图像特征进行处理,以得到所述目标图像属于预设类别的概率。
- 根据权利要求1所述的图像分类方法,其中,所述第一图像特征包括第一目标图像特征和第二目标图像特征,所述第一目标图像特征是利用包括至少一个卷积层的第一神经网络对所述目标图像进行卷积处理得到的,所述第二目标图像特征是利用第二神经网络对所述目标图像进行卷积处理得到的,所述第二图像特征包括第一目标参考图像特征以及第二目标参考图像特 征,其中所述第一目标参考图像特征是利用所述包括至少一个卷积层的第一神经网络对所述参考图像进行卷积处理得到的,所述第二目标参考图像特征是利用所述第二神经网络对所述参考图像进行卷积处理得到的,所述第一神经网络与所述第二神经网络是采用相同训练方法训练的不同网络,其中所述第一神经网络是利用第一训练集训练得到的,所述第二神经网络是利用第二训练集训练得到的,第一训练集和第二训练集的包括的训练图像中的正负样本比例是不同的。
- 根据权利要求5所述的图像分类方法,其中所述待分类图像特征包括第一待分类图像特征和第二待分类图像特征,其中所述第一待分类图像特征是通过拼接所述第一目标图像特征和所述第一目标参考图像特征确定的,所述第二待分类图像特征是通过拼接所述第二目标图像特征和所述第二目标参考图像特征确定的。
- 根据权利要求6所述的图像分类方法,其中,利用所述待分类图像特征确定所述目标图像的图像分类结果包括:利用第一全连接网络对所述第一待分类图像特征进行处理,以得到所述目标图像属于预设类别的第一概率;利用第二全连接网络对所述第二待分类图像特征进行处理,以得到所述目标图像属于预设类别的第二概率;以及根据所述第一概率和所述第二概率的加权平均值确定所述目标图像属于预设类别的概率,其中,所述第一全连接网络与所述第二全连接网络是采用相同训练方法训练的不同网络,其中所述第一全连接网络是利用第一训练集训练得到的,所述第二全连接网络是利用第二训练集训练得到的,第一训练集和第二训练集的包括的训练图像中的正负样本比例是不同的。
- 根据权利要求4或7所述的图像分类方法,其中所述第一神经网络和所述第一全连接网络是通过以下方式训练的:确定所述第一神经网络的第一训练集,其中所述第一训练集包括第一训练图像;确定所述第一训练图像的第一参考训练图像;利用所述第一神经网络对所述第一训练图像和所述第一参考训练图像分别进行卷积处理以得到第一训练图像特征和第二训练图像特征;根据所述第一训练图像特征和所述第二训练图像特征确定待分类训练图像特征;利用第一全连接网络对对所述待分类训练图像特征进行处理以确定所述第一训练图像属于预设类别的概率;调整所述第一神经网络和所述第一全连接网络的参数使得所述第一训练图像属于预设类别的概率与所述第一训练图像所属的真实类别之间的损失最小。
- 一种图像分类装置,包括:接收单元,配置成接收目标图像以及关于所述目标图像的参考图像,其中所述目标图像是医学图像;图像特征确定单元,配置成采用相同的方式确定所述目标图像的第一图像特征和所述参考图像的第二图像特征;融合单元,配置成融合所述第一图像特征和所述第二图像特征确定待分类图像特征;以及分类结果生成单元,配置成利用所述待分类图像特征确定所述目标图像属于预设类别的概率,其中,分类结果生成单元用于利用所述待分类图像特征,获得多个维度的向量,该向量中的元素分别表示所述目标图像和所述参考图像属于预设类别的置信分数,根据所述目标图像属于预设类别的置信分数,确定所述目标图像属于预设类别的概率。
- 根据权利要求9所述的图像分类装置,其中,图像特征确定单元进一步配置成:利用包括至少一个卷积层的第一神经网络对所述目标图像进行卷积处理以得到第一图像特征;以及利用所述第一神经网络对所述参考图像进行卷积处理以得到第二图像特征。
- 一种图像处理方法,由电子设备执行,所述方法包括:接收目标图像;利用第一神经网络确定所述目标图像的第一目标图像特征;利用第二神经网络确定所述目标图像的第二目标图像特征;根据所述第一目标图像特征和所述第二目标图像特征确定所述目标图像的第一图像处理结果和第二图像处理结果;融合所述第一图像处理结果和所述第二图像处理结果以确定所述目标图像的图像处理结果,其中,所述第一神经网络与所述第二神经网络是采用相同训练方法训练的不同网络,所述第一神经网络是利用第一训练集训练得到的,所述第二神经网络是利用第二训练集训练得到的,第一训练集和第二训练集的包括的训练图像中的正负样本比例是不同的。
- 如权利要求11所述的图像处理方法,其中所述图像处理结果包括图像分类结果、图像分割结果、目标检测结果中的至少一个。
- 一种医疗电子设备,包括:图像采集单元,配置成采集目标图像以及关于所述目标图像的参考图像,其中所述目标图像是医学图像;图像特征确定单元,配置成采用相同的方式确定所述目标图像的第一图像特征和所述参考图像的第二图像特征;融合单元,配置成融合所述第一图像特征和所述第二图像特征确定待分类图像特征;以及分类结果生成单元,配置成利用所述待分类图像特征确定所述目标图像属于预设类别的概率,其中,分类结果生成单元用于利用所述待分类图像特征,获得多个维度的向量,该向量中的元素分别表示所述目标图像和所述参考图像属于预设类别的置信分数,根据所述目标图像属于预设类别的置信分数,确定所述目标图像属于预设类别的概率。
- 一种图像分类设备,所述设备包括存储器和处理器,其中所述存储器中存有指令,当利用所述处理器执行所述指令时,使得所述处理器执行如权利要求1-8和11-12中任一项所述的图像分类方法。
- 一种计算机可读存储介质,其上存储有指令,所述指令在被处理器执行时,使得所述处理器执行如权利要求1-8和11-12中任一项所述的图像分类方法。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP20831068.0A EP3992851A4 (en) | 2019-06-28 | 2020-06-28 | METHOD, APPARATUS AND DEVICE FOR IMAGE CLASSIFICATION, STORAGE MEDIA AND MEDICAL ELECTRONIC DEVICE |
KR1020217028280A KR102605969B1 (ko) | 2019-06-28 | 2020-06-28 | 이미지 분류 방법, 장치 및 디바이스, 저장 매체, 그리고 의료 전자 디바이스 |
JP2021548679A JP7297081B2 (ja) | 2019-06-28 | 2020-06-28 | 画像分類方法、画像分類装置、医療用電子機器、画像分類機器、及びコンピュータプログラム |
US17/459,920 US11900647B2 (en) | 2019-06-28 | 2021-08-27 | Image classification method, apparatus, and device, storage medium, and medical electronic device |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910573560.9A CN110276411B (zh) | 2019-06-28 | 2019-06-28 | 图像分类方法、装置、设备、存储介质和医疗电子设备 |
CN201910573560.9 | 2019-06-28 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/459,920 Continuation US11900647B2 (en) | 2019-06-28 | 2021-08-27 | Image classification method, apparatus, and device, storage medium, and medical electronic device |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020259666A1 true WO2020259666A1 (zh) | 2020-12-30 |
Family
ID=67963631
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/098407 WO2020259666A1 (zh) | 2019-06-28 | 2020-06-28 | 图像分类方法、装置、设备、存储介质和医疗电子设备 |
Country Status (6)
Country | Link |
---|---|
US (1) | US11900647B2 (zh) |
EP (1) | EP3992851A4 (zh) |
JP (1) | JP7297081B2 (zh) |
KR (1) | KR102605969B1 (zh) |
CN (1) | CN110276411B (zh) |
WO (1) | WO2020259666A1 (zh) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114139031A (zh) * | 2021-10-28 | 2022-03-04 | 马上消费金融股份有限公司 | 数据分类方法、装置、电子设备及存储介质 |
CN114582469A (zh) * | 2022-03-11 | 2022-06-03 | 无锡祥生医疗科技股份有限公司 | 医疗影像分类方法、装置、设备及存储介质 |
EP4350575A4 (en) * | 2021-06-30 | 2024-09-18 | Huawei Tech Co Ltd | IMAGE CLASSIFICATION METHOD AND ASSOCIATED DEVICE |
Families Citing this family (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110276411B (zh) * | 2019-06-28 | 2022-11-18 | 腾讯科技(深圳)有限公司 | 图像分类方法、装置、设备、存储介质和医疗电子设备 |
CN110827294A (zh) * | 2019-10-31 | 2020-02-21 | 北京推想科技有限公司 | 网络模型训练方法及装置、病灶区域确定方法及装置 |
CN111028310B (zh) * | 2019-12-31 | 2023-10-03 | 上海联影医疗科技股份有限公司 | 乳腺断层扫描的扫描参数确定方法、装置、终端及介质 |
CN111311578B (zh) * | 2020-02-17 | 2024-05-03 | 腾讯科技(深圳)有限公司 | 基于人工智能的对象分类方法以及装置、医学影像设备 |
CN111598131B (zh) * | 2020-04-17 | 2023-08-25 | 北京百度网讯科技有限公司 | 图像处理方法、装置、电子设备及存储介质 |
CN112149748B (zh) * | 2020-09-28 | 2024-05-21 | 商汤集团有限公司 | 图像分类方法及装置、电子设备和存储介质 |
CN112138394B (zh) * | 2020-10-16 | 2022-05-03 | 腾讯科技(深圳)有限公司 | 图像处理方法、装置、电子设备及计算机可读存储介质 |
CN113313063A (zh) * | 2021-06-21 | 2021-08-27 | 暨南大学 | 麦穗检测方法、电子装置和存储介质 |
CN113283552A (zh) * | 2021-07-22 | 2021-08-20 | 深圳百胜扬工业电子商务平台发展有限公司 | 图像的分类方法、装置、存储介质及电子设备 |
CN113569953A (zh) * | 2021-07-29 | 2021-10-29 | 中国工商银行股份有限公司 | 分类模型的训练方法、装置及电子设备 |
KR102415616B1 (ko) * | 2021-10-18 | 2022-07-01 | 라이트하우스(주) | 예술품의 이미지 표준화 기반 교육 및 거래 서비스 제공 방법, 장치 및 시스템 |
CN114220063B (zh) * | 2021-11-17 | 2023-04-07 | 浙江大华技术股份有限公司 | 目标检测方法及装置 |
CN115130539A (zh) * | 2022-04-21 | 2022-09-30 | 腾讯科技(深圳)有限公司 | 分类模型训练、数据分类方法、装置和计算机设备 |
CN114820592B (zh) * | 2022-06-06 | 2023-04-07 | 北京医准智能科技有限公司 | 图像处理装置、电子设备及介质 |
CN117036894B (zh) * | 2023-10-09 | 2024-03-26 | 之江实验室 | 基于深度学习的多模态数据分类方法、装置及计算机设备 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020067857A1 (en) * | 2000-12-04 | 2002-06-06 | Hartmann Alexander J. | System and method for classification of images and videos |
CN106650550A (zh) * | 2015-10-28 | 2017-05-10 | 中通服公众信息产业股份有限公司 | 一种融合车标与车头图像特征的车辆型号识别方法及系统 |
CN109146848A (zh) * | 2018-07-23 | 2019-01-04 | 东北大学 | 一种融合多模态乳腺图像的计算机辅助参考系统及方法 |
CN109447065A (zh) * | 2018-10-16 | 2019-03-08 | 杭州依图医疗技术有限公司 | 一种乳腺影像识别的方法及装置 |
CN110276411A (zh) * | 2019-06-28 | 2019-09-24 | 腾讯科技(深圳)有限公司 | 图像分类方法、装置、设备、存储介质和医疗电子设备 |
CN110321920A (zh) * | 2019-05-08 | 2019-10-11 | 腾讯科技(深圳)有限公司 | 图像分类方法、装置、计算机可读存储介质和计算机设备 |
Family Cites Families (26)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6778705B2 (en) * | 2001-02-27 | 2004-08-17 | Koninklijke Philips Electronics N.V. | Classification of objects through model ensembles |
JP5652250B2 (ja) * | 2011-02-24 | 2015-01-14 | 富士ゼロックス株式会社 | 画像処理プログラム及び画像処理装置 |
US8498491B1 (en) * | 2011-08-10 | 2013-07-30 | Google Inc. | Estimating age using multiple classifiers |
US10055551B2 (en) * | 2013-10-10 | 2018-08-21 | Board Of Regents Of The University Of Texas System | Systems and methods for quantitative analysis of histopathology images using multiclassifier ensemble schemes |
CN104573708A (zh) * | 2014-12-19 | 2015-04-29 | 天津大学 | 组合降采样极限学习机 |
CN106156807B (zh) * | 2015-04-02 | 2020-06-02 | 华中科技大学 | 卷积神经网络模型的训练方法及装置 |
CN106874279B (zh) * | 2015-12-11 | 2021-01-15 | 腾讯科技(深圳)有限公司 | 生成应用类别标签的方法及装置 |
US9779492B1 (en) * | 2016-03-15 | 2017-10-03 | International Business Machines Corporation | Retinal image quality assessment, error identification and automatic quality correction |
CN106096670B (zh) * | 2016-06-17 | 2019-07-30 | 深圳市商汤科技有限公司 | 级联卷积神经网络训练和图像检测方法、装置及系统 |
CN106682435B (zh) * | 2016-12-31 | 2021-01-29 | 西安百利信息科技有限公司 | 一种多模型融合自动检测医学图像中病变的系统及方法 |
CN108771530B (zh) * | 2017-05-04 | 2021-03-30 | 深圳硅基智能科技有限公司 | 基于深度神经网络的眼底病变筛查系统 |
CN107688823B (zh) * | 2017-07-20 | 2018-12-04 | 北京三快在线科技有限公司 | 一种图像特征获取方法及装置,电子设备 |
US20190034734A1 (en) * | 2017-07-28 | 2019-01-31 | Qualcomm Incorporated | Object classification using machine learning and object tracking |
JP2019028887A (ja) * | 2017-08-02 | 2019-02-21 | 国立研究開発法人産業技術総合研究所 | 画像処理方法 |
CN109426858B (zh) * | 2017-08-29 | 2021-04-06 | 京东方科技集团股份有限公司 | 神经网络、训练方法、图像处理方法及图像处理装置 |
CN107665352A (zh) * | 2017-09-07 | 2018-02-06 | 浙江工业大学 | 一种基于多通道残差网络的珍珠分类方法 |
CN107680088A (zh) * | 2017-09-30 | 2018-02-09 | 百度在线网络技术(北京)有限公司 | 用于分析医学影像的方法和装置 |
US20190130188A1 (en) * | 2017-10-26 | 2019-05-02 | Qualcomm Incorporated | Object classification in a video analytics system |
US20190130191A1 (en) * | 2017-10-30 | 2019-05-02 | Qualcomm Incorporated | Bounding box smoothing for object tracking in a video analytics system |
CN109934369A (zh) * | 2017-12-15 | 2019-06-25 | 北京京东尚科信息技术有限公司 | 用于信息推送的方法及装置 |
CN109461495B (zh) * | 2018-11-01 | 2023-04-14 | 腾讯科技(深圳)有限公司 | 一种医学图像的识别方法、模型训练的方法及服务器 |
WO2020102584A2 (en) * | 2018-11-14 | 2020-05-22 | Intuitive Surgical Operations, Inc. | Convolutional neural networks for efficient tissue segmentation |
CN110516620B (zh) * | 2019-08-29 | 2023-07-28 | 腾讯科技(深圳)有限公司 | 目标跟踪方法、装置、存储介质及电子设备 |
US11544495B2 (en) * | 2020-07-10 | 2023-01-03 | Adobe Inc. | Attributionally robust training for weakly supervised localization and segmentation |
US11776129B2 (en) * | 2020-12-16 | 2023-10-03 | Qualcomm Incorporated | Semantic refinement of image regions |
WO2023164145A1 (en) * | 2022-02-25 | 2023-08-31 | Regeneron Pharmaceuticals, Inc. | Size exclusion chromatography for characterizing host cell proteins |
-
2019
- 2019-06-28 CN CN201910573560.9A patent/CN110276411B/zh active Active
-
2020
- 2020-06-28 KR KR1020217028280A patent/KR102605969B1/ko active IP Right Grant
- 2020-06-28 WO PCT/CN2020/098407 patent/WO2020259666A1/zh active Application Filing
- 2020-06-28 EP EP20831068.0A patent/EP3992851A4/en active Pending
- 2020-06-28 JP JP2021548679A patent/JP7297081B2/ja active Active
-
2021
- 2021-08-27 US US17/459,920 patent/US11900647B2/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020067857A1 (en) * | 2000-12-04 | 2002-06-06 | Hartmann Alexander J. | System and method for classification of images and videos |
CN106650550A (zh) * | 2015-10-28 | 2017-05-10 | 中通服公众信息产业股份有限公司 | 一种融合车标与车头图像特征的车辆型号识别方法及系统 |
CN109146848A (zh) * | 2018-07-23 | 2019-01-04 | 东北大学 | 一种融合多模态乳腺图像的计算机辅助参考系统及方法 |
CN109447065A (zh) * | 2018-10-16 | 2019-03-08 | 杭州依图医疗技术有限公司 | 一种乳腺影像识别的方法及装置 |
CN110321920A (zh) * | 2019-05-08 | 2019-10-11 | 腾讯科技(深圳)有限公司 | 图像分类方法、装置、计算机可读存储介质和计算机设备 |
CN110276411A (zh) * | 2019-06-28 | 2019-09-24 | 腾讯科技(深圳)有限公司 | 图像分类方法、装置、设备、存储介质和医疗电子设备 |
Non-Patent Citations (1)
Title |
---|
See also references of EP3992851A4 |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP4350575A4 (en) * | 2021-06-30 | 2024-09-18 | Huawei Tech Co Ltd | IMAGE CLASSIFICATION METHOD AND ASSOCIATED DEVICE |
CN114139031A (zh) * | 2021-10-28 | 2022-03-04 | 马上消费金融股份有限公司 | 数据分类方法、装置、电子设备及存储介质 |
CN114139031B (zh) * | 2021-10-28 | 2024-03-19 | 马上消费金融股份有限公司 | 数据分类方法、装置、电子设备及存储介质 |
CN114582469A (zh) * | 2022-03-11 | 2022-06-03 | 无锡祥生医疗科技股份有限公司 | 医疗影像分类方法、装置、设备及存储介质 |
Also Published As
Publication number | Publication date |
---|---|
EP3992851A1 (en) | 2022-05-04 |
CN110276411A (zh) | 2019-09-24 |
CN110276411B (zh) | 2022-11-18 |
KR20210119526A (ko) | 2021-10-05 |
JP7297081B2 (ja) | 2023-06-23 |
JP2022537866A (ja) | 2022-08-31 |
KR102605969B1 (ko) | 2023-11-23 |
US20210390695A1 (en) | 2021-12-16 |
EP3992851A4 (en) | 2022-08-10 |
US11900647B2 (en) | 2024-02-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2020259666A1 (zh) | 图像分类方法、装置、设备、存储介质和医疗电子设备 | |
Han et al. | Learning more with less: Conditional PGGAN-based data augmentation for brain metastases detection using highly-rough annotation on MR images | |
WO2020238734A1 (zh) | 图像分割模型的训练方法、装置、计算机设备和存储介质 | |
US12118739B2 (en) | Medical image processing method, apparatus, and device, medium, and endoscope | |
Pang et al. | CTumorGAN: a unified framework for automatic computed tomography tumor segmentation | |
WO2020125498A1 (zh) | 心脏磁共振图像分割方法、装置、终端设备及存储介质 | |
CN107563434B (zh) | 一种基于三维卷积神经网络的脑部mri图像分类方法、装置 | |
WO2020248898A1 (zh) | 图像处理方法、装置、设备和存储介质 | |
TW202125415A (zh) | 三維目標檢測及模型的訓練方法、設備、儲存媒體 | |
CN110458829A (zh) | 基于人工智能的图像质控方法、装置、设备及存储介质 | |
WO2021159811A1 (zh) | 青光眼辅助诊断装置、方法及存储介质 | |
CN113989293A (zh) | 图像分割方法及相关模型的训练方法和装置、设备 | |
Cai et al. | Identifying architectural distortion in mammogram images via a se-densenet model and twice transfer learning | |
Wen et al. | Pulmonary nodule detection based on convolutional block attention module | |
Guo et al. | Blind image quality assessment for pathological microscopic image under screen and immersion scenarios | |
Wen et al. | A-PSPNet: A novel segmentation method of renal ultrasound image | |
Saednia et al. | A cascaded deep learning framework for segmentation of nuclei in digital histology images | |
CN118262121A (zh) | 图像特征提取方法、装置、电子设备及存储介质 | |
CN104766068A (zh) | 一种多规则融合的随机游走舌像提取方法 | |
Sibille et al. | Whole-body tumor segmentation of 18f-fdg pet/ct using a cascaded and ensembled convolutional neural networks | |
CN114972297A (zh) | 口腔健康监测方法及装置 | |
CN113808130A (zh) | 肿瘤图像智能分类方法、装置和医学图像处理设备 | |
Tawfeeq et al. | Predication of Most Significant Features in Medical Image by Utilized CNN and Heatmap. | |
CN110415239B (zh) | 图像处理方法、装置、设备、医疗电子设备以及介质 | |
CN114708973B (zh) | 一种用于对人体健康进行评估的设备和存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20831068 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2021548679 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20217028280 Country of ref document: KR Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2020831068 Country of ref document: EP |