WO2022095258A1 - Image object classification method and apparatus, device, storage medium and program - Google Patents

Image object classification method and apparatus, device, storage medium and program Download PDF

Info

Publication number
WO2022095258A1
WO2022095258A1 PCT/CN2020/139913 CN2020139913W WO2022095258A1 WO 2022095258 A1 WO2022095258 A1 WO 2022095258A1 CN 2020139913 W CN2020139913 W CN 2020139913W WO 2022095258 A1 WO2022095258 A1 WO 2022095258A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
classified
target object
feature information
initial
Prior art date
Application number
PCT/CN2020/139913
Other languages
French (fr)
Chinese (zh)
Inventor
朱雅靖
陈翼男
罗祥德
任家敏
Original Assignee
上海商汤智能科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 上海商汤智能科技有限公司 filed Critical 上海商汤智能科技有限公司
Publication of WO2022095258A1 publication Critical patent/WO2022095258A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features

Definitions

  • the present disclosure relates to the technical field of image processing, and in particular, to an image object classification method, apparatus, device, storage medium and program.
  • the scan image categories often include time-related pre-contrast scan, early arterial phase, late arterial phase, portal venous phase, and delayed phase, etc.
  • the scan image category can also include scan parameters related to the T1-weighted inverse imaging, T1-weighted in-phase imaging, T2-weighted imaging, diffusion-weighted imaging, and surface diffusion coefficient imaging, etc.
  • the doctor in the process of disease diagnosis and treatment, the doctor usually needs to repeatedly check the signs of the target object such as the tumor on the medical image, which makes the determination of the type of the tumor overly dependent on the professional level of the doctor, and the problem of low efficiency in determining the type of the tumor. .
  • the present disclosure provides at least an image object classification method, apparatus, device, storage medium and program.
  • the embodiment of the present disclosure provides an image object classification method, and the image object classification method includes:
  • the at least one image to be classified is a medical image belonging to at least one scanned image category
  • the at least one image to be classified is subjected to target classification to obtain the type of the target object.
  • the classification model is used to classify the at least one image to be classified, and the type of the target object is obtained. Because the classification model is used to classify the image to be classified, intelligence is realized. Target classification, and no manual target classification is required, which can reduce the dependence on manual work and improve the efficiency of target classification.
  • performing target classification on the at least one image to be classified to obtain the type of the target object includes:
  • the final feature information is classified to obtain the type of the target object.
  • the final feature information can be classified to obtain the type of the target object, so the target classification is realized by using the feature information of the target object.
  • the method before the target classification is performed on the at least one image to be classified and the type of the target object is obtained, the method further includes:
  • the final area to perform several layers of feature extraction on the at least one image to be classified, correspondingly obtain several sets of initial feature information; wherein, in the feature extraction process, the weight of the image to be classified corresponding to the final area is high the weight of other regions in the image to be classified; and/or, the features corresponding to the final region in the initial feature information are more abundant than the features of other regions.
  • the weight of the corresponding final region in the image to be classified is higher than the weights of other regions in the image to be classified, so that the classification model tends to extract features with more details for the final region; and /or, the features corresponding to the final region in the initial feature information are richer than those of other regions; thus, the classification model can learn the feature information of the target object itself better by using the initial feature information of the image to be classified, to a certain extent Reduce the impact of noise interference around the target object on target classification.
  • obtaining the final area of the target object based on the initial area corresponding to the target object in the image to be classified includes:
  • the union of the initial regions corresponding to the target object in the at least one image to be classified is obtained as the final region of the target object.
  • the final area of the target object is the union of the initial areas of the target object in the image to be classified
  • the final area is greater than or equal to any initial area, ensuring that the final area of the target object can contain different targets in the to-be-classified image.
  • the object corresponds to the area, so that the feature information of the target object can be paid attention to as much as possible when the feature extraction of the image to be classified is performed.
  • the at least one image to be classified includes a first image to be classified without an initial region of the target object and a second image to be classified with an initial region of the target object labeled; Before obtaining the final area of the target object based on the initial area corresponding to the target object in the image to be classified, the method further includes:
  • the classification model to detect that the first image to be classified is not marked with the initial area of the target object, and based on the initial area of the target object marked on the second to-be-classified image and the second to-be-classified image The registration relationship between the classified image and the first to-be-classified image determines the initial area of the target object on the first to-be-classified image.
  • the classification model can be used to determine the initial area of the target object for the first to-be-classified image that is not labeled with the initial area of the target object, so as to complete the labeling, so that the to-be-classified images all include the initial area.
  • the method before obtaining the final feature information based on at least one set of initial feature information in the several sets of initial feature information, the method further includes:
  • obtaining final feature information based on at least one set of initial feature information in the several groups of initial feature information including:
  • the at least one set of initial feature information is fused to obtain the final feature information.
  • each group of initial feature information is uniformly converted into a preset dimension, which facilitates subsequent acquisition of final feature information.
  • the weight of at least one set of initial feature information can be used to fuse the initial feature information of different sizes extracted from at least one layer of features to obtain the final feature information.
  • the initial feature information of small size may be compressed to remove important features.
  • the weight of each set of the initial feature information is determined during the training process of the classification model.
  • the weight of the initial feature information for fusion is determined through the iterative training of the classification model, so that the final feature information obtained by using the weight fusion can better reflect the characteristics of the target object and further improve the classification performance.
  • the preset dimension is one dimension.
  • each group of initial feature information can be converted into one-dimensional, data unification is realized, and subsequent fusion is facilitated.
  • the classification model uses an ArcFace loss function during the training process to determine the loss value of the classification model; and/or, the batch sample data selected for each training of the classification model is generated by using the data
  • the number of different target types selected by the processor from the sample data set is a preset proportion of sample data.
  • the ArcFace loss function to determine the loss value of the classification model can aggregate the feature information of the same target objects and keep the feature information of different types of target objects away, thereby improving the classification performance of the target objects.
  • the data generator is used to select sample data from the sample data set, and the sample data with a preset ratio of different target types is used as the batch sample data, so that the target types of the batch sample data for training the classification model are more balanced.
  • the acquiring at least one image to be classified including the target object includes:
  • the to-be-classified images containing the target object are respectively extracted from a plurality of original medical images.
  • the image to be classified is obtained, and the image to be classified can be extracted from the original medical image.
  • the image size of the subsequent classification can be reduced, and some backgrounds in the original medical image can be avoided to a certain extent. Therefore, the consumption of processing resources for subsequent classification can be reduced, and the classification performance can be improved.
  • the image to be classified containing the target object is extracted from multiple original medical images, including:
  • the image data in the to-be-extracted area is extracted from the original medical image to obtain the to-be-classified image.
  • the initial area is the area containing the target object, and the initial area of the target object is expanded according to a preset ratio, so that the obtained area to be extracted contains both the target object and some background information around the target object, so that the area to be extracted can be extracted.
  • the image to be classified can include the target object and some background information.
  • the method before the image to be classified including the target object is extracted from multiple original medical images, the method further includes at least one of the following steps:
  • the initial area of the target object is not marked in the first original medical image, and the initial area of the target object marked on the second original medical image and the second original medical image and the first original medical image are used.
  • the registration relationship is determined, and the initial area of the target object on the first original medical image is determined.
  • the original medical image can be preprocessed before the image to be classified is extracted from the original medical image, and the unclassified image can be unified. Image parameters of the image to improve the quality of the image to be classified.
  • the original medical image and the image to be classified are two-dimensional images; or, the original medical image is a three-dimensional image, and the image to be classified is a two-dimensional image or a three-dimensional image.
  • the image to be classified is extracted from the original medical image. If the original medical image is a two-dimensional image, the image to be classified is a two-dimensional image; and if the original medical image is a three-dimensional image, the image to be classified is a two-dimensional image.
  • the dimensions can be two-dimensional or three-dimensional.
  • the original medical image is a three-dimensional image
  • the image to be classified is a two-dimensional image obtained by extracting a layer where the target object has the largest area in the original medical image.
  • the layer where the maximum area of the target object is located in the original medical image can be extracted as the image to be classified, so that the extraction range of the target object in the image to be classified is larger. , contains more information about the target object and improves the classification accuracy of the target object.
  • the embodiment of the present disclosure also provides an image object classification device, and the image object classification device includes:
  • an image acquisition module configured to acquire at least one image to be classified including the target object, wherein the at least one image to be classified is a medical image belonging to at least one scanned image category;
  • the target classification module is configured to use a classification model to perform target classification on the at least one image to be classified to obtain the type of the target object.
  • the target classification module is configured to:
  • the final feature information is classified to obtain the type of the target object.
  • the target classification module is configured to:
  • the target classification module configured as:
  • the final area to perform several layers of feature extraction on the at least one image to be classified, correspondingly obtain several sets of initial feature information; wherein, in the feature extraction process, the weight of the image to be classified corresponding to the final area is high the weight of other regions in the image to be classified; and/or, the features corresponding to the final region in the initial feature information are more abundant than the features of other regions.
  • the target classification module is configured to:
  • the union of the initial regions corresponding to the target object in the at least one image to be classified is obtained as the final region of the target object.
  • the target classification module is configured to:
  • the classification model to detect that the first image to be classified is not marked with the initial area of the target object, and based on the initial area of the target object marked on the second to-be-classified image and the second to-be-classified image The registration relationship between the classified image and the first to-be-classified image determines the initial area of the target object on the first to-be-classified image.
  • the target classification module is configured to:
  • the target classification module configured as:
  • the at least one set of initial feature information is fused to obtain the final feature information.
  • the weight of each set of the initial feature information is determined during the training process of the classification model.
  • the preset dimension is one dimension.
  • the classification model uses an ArcFace loss function during the training process to determine the loss value of the classification model; and/or, the batch sample data selected for each training of the classification model is generated by using the data
  • the number of different target types selected by the processor from the sample data set is a preset proportion of sample data.
  • the image acquisition module is configured to:
  • the to-be-classified images containing the target object are respectively extracted from a plurality of original medical images.
  • the image acquisition module is configured to:
  • the image data in the to-be-extracted area is extracted from the original medical image to obtain the to-be-classified image.
  • the image acquisition module is configured to:
  • the initial area of the target object is not marked in the first original medical image, and the initial area of the target object marked on the second original medical image and the second original medical image and the first original medical image are used.
  • the registration relationship is determined, and the initial area of the target object on the first original medical image is determined.
  • the original medical image and the image to be classified are two-dimensional images; or, the original medical image is a three-dimensional image, and the image to be classified is a two-dimensional image or a three-dimensional image.
  • the original medical image is a three-dimensional image
  • the image to be classified is a two-dimensional image obtained by extracting a layer where the target object has the largest area in the original medical image.
  • An embodiment of the present disclosure also provides an electronic device, including a memory and a processor coupled to each other, where the processor is configured to execute program instructions stored in the memory, so as to implement the image object classification method provided in any of the previous embodiments .
  • the embodiments of the present disclosure also provide a computer-readable storage medium, which stores program instructions, and the program instructions are executed by a processor as the image object classification method provided in any of the previous embodiments.
  • An embodiment of the present disclosure also provides a computer program, the computer program includes computer-readable codes, and when the computer-readable codes are executed in an electronic device, the processor of the electronic device executes any of the preceding implementations The image object classification method described in the example.
  • an image target classification method based on artificial intelligence technology is proposed to achieve intelligent target classification. Because the classification model is used to classify the images to be classified, it not only makes the target classification process simpler, reduces the dependence on doctors, improves the speed and accuracy of target classification, but also combines artificial intelligence technology to achieve target classification, so as to assist doctors in intelligent diseases diagnosis and treatment.
  • FIG. 1 is a schematic flowchart of an image target classification method provided by an embodiment of the present disclosure
  • FIG. 2 is a schematic diagram of a system architecture to which the image object classification method according to the embodiment of the present disclosure can be applied;
  • FIG. 3 is a schematic flowchart of obtaining at least one image to be classified according to an embodiment of the present disclosure
  • FIG. 4 is a schematic flowchart of a target classification for at least one image to be classified according to an embodiment of the present disclosure
  • FIG. 5 is a schematic diagram of a network architecture used by a classification model in the image target classification method according to an embodiment of the present disclosure
  • FIG. 6 is a schematic diagram of a framework of an image object classification apparatus 60 provided by an embodiment of the present disclosure.
  • FIG. 7 is a schematic frame diagram of an electronic device 70 provided by an embodiment of the present disclosure.
  • FIG. 8 is a schematic diagram of a framework of a computer-readable storage medium 80 provided by an embodiment of the present disclosure.
  • three-dimensional imaging technology based on CT and MR plays a crucial role in medical imaging diagnosis, and is one of the main imaging examination methods for diagnosing, for example, liver diseases.
  • the scanning sequence of CT examination mainly includes the plain scan phase, the dynamic enhancement phase, the arterial phase, the portal venous phase and the delayed phase.
  • the plain scan period is generally used to observe changes in the liver surface, whether there are fatty liver, liver fibrosis, liver cirrhosis and other diseases.
  • phase images with dynamic enhancement can show the specific image features of the lesion.
  • HCC Hepatocellular Carcinoma
  • HCC mainly occurs in patients with chronic liver disease and liver cirrhosis, and the corresponding changes in liver surface morphology can be observed from the plain scan period. Or the same density as the liver parenchyma; after enhanced scanning, HCC in each phase showed: marked enhancement or inhomogeneous enhancement in arterial phase, accompanied by low-density capsule; contrast agent outflow in portal venous phase, showing enhanced capsule at the same time ; the delayed phase presents a delayed enhanced envelope. Therefore, in a feasible implementation manner, it can be determined whether the target tumor is HCC by identifying the imaging features exhibited by the images in multiple stages. Compared with the single-phase image, the judgment accuracy is higher, because the image characteristics of the small liver metastases with rich blood supply in the plain and arterial phase are similar to those of small HCC. classification task, which can further improve the accuracy of image classification.
  • medical image analysis generally has problems such as less labeled data, complex and difficult tasks, and at the same time, in order to better characterize lesions, it is necessary to analyze the correlation between sequences.
  • the existence of these problems limits the complexity and depth of deep learning networks to a certain extent, and some other strategies need to be introduced to solve the task of medical image analysis.
  • the image features of the tumor itself are the main basis for judging its type, and there may be various noises around the target tumor, which will mislead the deep learning network to learn some wrong features; liver tumors vary in size and small.
  • the network needs to be able to take into account the characteristics of liver tumors, and improve the identification ability of small tumors on the basis of ensuring high-precision classification and identification of large tumors; limited to the resolution of CT scan images,
  • the imaging features of liver tumors are not necessarily obvious.
  • the present disclosure provides at least one image target classification method, which uses a classification model to classify images to be classified, which not only makes the target classification process simpler, reduces the dependence on doctors, and improves the speed and accuracy of target classification, And combined with artificial intelligence technology to achieve target classification, in order to assist doctors in intelligent disease diagnosis and treatment.
  • FIG. 1 is a schematic flowchart of an image object classification method provided by an embodiment of the present disclosure. Specifically, the following steps can be included:
  • Step S11 Acquire at least one image to be classified that includes the target object.
  • At least one image to be classified is a medical image belonging to at least one scanned image category.
  • the images to be classified may be medical images, including but not limited to CT images and MR images, which are not limited herein.
  • the images to be classified may all be CT images, may all be MR images, and may also be partly CT images and partly MR images, which are not specifically limited herein.
  • CT images and MR images are multi-phase images or multi-sequence imaging. Each phase image or sequence shows different image information of the area where the target object is located or other areas. Combined effectively, the nature of the lesions can be more precisely defined.
  • the images to be classified may be obtained by scanning the abdomen, chest and other regions.
  • the image to be classified obtained by scanning the abdomen may include tissues and organs such as liver, spleen, and kidney
  • the image to be classified obtained by scanning the chest may include tissues and organs such as the heart and lung.
  • the images to be classified may be scanned according to the actual application. Images, not limited here.
  • the target object may be, but is not limited to, a liver tumor and other objects that need to be classified using the image object classification method of the embodiment of the present disclosure.
  • the at least one image to be classified may be a medical image belonging to at least one category of scanned images. Medical images of different scanned image categories can be used to display different characteristic information of target objects, thus improving the accuracy of image target classification.
  • the class of scanned images may also be referred to as the above-described images and/or sequences.
  • the images of the different scanned image categories may be timing-dependent and/or scan-parameter-dependent images.
  • the scan image category may include time-series-related pre-contrast scan, early arterial phase, late arterial phase, portal venous phase, and delayed phase, etc.; alternatively, the scan image category may also include T1-weighted inverse imaging, T1 Weighted in-phase imaging, T2-weighted imaging, diffusion-weighted imaging, and surface diffusion coefficient imaging, etc.
  • the early arterial stage can indicate that the portal vein has not been enhanced
  • the late arterial stage can indicate that the portal vein has been enhanced
  • the portal venous phase can indicate that the portal vein has been sufficiently enhanced and the liver blood vessels have been enhanced by forward blood flow.
  • the delay period can indicate that the portal vein and arteries are in an enhanced state and weaker than the portal venous phase
  • the liver parenchyma is in an enhanced state and weaker than the portal venous phase
  • other scanning image categories will not be listed one by one here.
  • Step S12 Using the classification model, perform target classification on at least one image to be classified to obtain the type of the target object.
  • the classification model is used to classify the at least one image to be classified, so as to obtain the type of the target object.
  • the classification model performs target classification on at least one image to be classified, obtains probabilities that the target objects belong to different types, and uses the types that satisfy the preset probability conditions as the types of the target objects.
  • the preset probability conditions include but are not limited to the maximum probability value and the like.
  • the probability that the target objects belong to different types can be obtained by training the classification model.
  • the batch sample data selected for each training of the classification model is sample data with a preset proportion of the number of different target types selected from the sample data set by the data generator. Since the data generator randomly selects sample data containing equal proportions of different target types as batch sample data, so as to avoid unbalanced classification performance due to too few sample data of a certain target type, the classification model is used for at least one image to be classified.
  • the target classification is obtained by training a large number of batch sample data, which can improve the classification performance of the classification model.
  • Using the classification model to obtain the type of the target object can assist the doctor in determining the type of the target object, save the doctor's time for reviewing the images to be classified, and thus can speed up the output of the report.
  • target classification is performed on at least one image to be classified, and when the type of the target object is obtained, several layers of feature extraction are performed on at least one image to be classified, and several sets of initial feature information are correspondingly obtained; At least one set of initial feature information in the feature information is used to obtain final feature information; the final feature information is classified to obtain the type of the target object.
  • the number of layers for feature extraction may be one layer, two layers or even more layers.
  • which layers to perform feature extraction on can be obtained through artificial settings, or can be determined through a large number of experiments when training a classification model, which is not specifically limited here.
  • a layer of feature extraction is performed on at least one image to be classified, and a set of initial feature information is correspondingly obtained.
  • Multi-layer feature extraction is performed on at least one image to be classified, and multiple sets of initial feature information are correspondingly obtained, wherein the multi-layer feature extraction may be continuous or discontinuous.
  • the initial feature information may be a feature map of the target object, reflecting the feature information of the target object in the image to be classified.
  • the classification model is a deep learning network
  • the deep learning network may include an encoder (encoder) or its variants, Resnet or its variants, and may be a neural network (Visual Geometry Group Network, VGG) 16 or its variants. , or other network model structures for classification.
  • the classification model performs feature extraction on at least one image to be classified through a convolution layer, and different convolution layers correspond to different layers of feature extraction to obtain different groups of initial feature information.
  • the classification model is used to classify the at least one image to be classified, and the type of the target object is obtained. Therefore, an image target classification method based on artificial intelligence technology is proposed, which realizes the Intelligent target classification, and no manual target classification is required, which can reduce the dependence on manual work and improve the efficiency of target classification.
  • At least one image to be classified including liver tumors is acquired, and a classification model is used to perform target classification on at least one image to be classified, so as to obtain the type of liver tumor, and no manual classification is required.
  • the images are classified, and the classification model can be used to realize the classification of liver tumors, so that the doctor can obtain the type of liver tumors.
  • FIG. 2 is a schematic diagram of a system architecture to which the image object classification method according to an embodiment of the present disclosure can be applied; as shown in FIG. 2 , the system architecture includes an image acquisition terminal 201 , a network 202 and an object classification terminal 203 .
  • the image acquisition terminal 201 and the target classification terminal 203 establish a communication connection through the network 202, and the image acquisition terminal 201 reports at least one image to be classified containing the target object to the target classification terminal 203 through the network 202, and the target classification The terminal 203 responds to the received at least one image to be classified, and uses the classification model to perform target classification on the at least one image to be classified to obtain the type of the target object.
  • the target classification terminal 203 uploads the type of the target object to the network 202 and sends it to the image acquisition terminal 201 through the network 202 .
  • the image acquisition terminal 201 may include an image acquisition device, and the target classification terminal 203 may include a vision processing device or a remote server with visual information processing capability.
  • Network 202 may employ wired or wireless connections.
  • the image acquisition terminal 201 can be connected to the visual processing device through a wired connection, such as data communication through a bus; when the target classification terminal 203 is a remote server, the image acquisition terminal 201 can perform data interaction with a remote server through a wireless network.
  • the image acquisition terminal 201 may be a vision processing device with an image acquisition module, which is specifically implemented as a host with a camera.
  • the image object classification method according to the embodiment of the present disclosure may be executed by the image acquisition terminal 201 , and the above-mentioned system architecture may not include the network 202 and the object classification terminal 203 .
  • FIG. 3 is a schematic flowchart of acquiring at least one image to be classified according to an embodiment of the present disclosure. Specifically, the following steps can be included:
  • Step S111 Resampling the original medical image to a preset resolution.
  • the size of the preset resolution can be customized, and the preset resolution corresponding to the target object can be set according to different target objects, so as to unify the resolution of the original medical image to the resolution with the best image effect.
  • Step S112 Adjust the pixel value range in the original medical image.
  • the brightness and color of the original medical image are made easier to display the target object.
  • the categories of the original medical images include, but are not limited to, CT images, MR images, and other images that can reflect the feature information of the target object, which are not limited herein.
  • the original medical image is a CT image
  • the original medical image can be unified to a preset window width and window level;
  • the gray value corresponding to a preset ratio (for example, 99.9%) under the grayscale cumulative distribution function can be used as the normalized preprocessing clamp value, so that the contrast of the MR image data can be enhanced and the subsequent image target can be improved. Classification accuracy.
  • Step S113 Normalize the original medical image.
  • the raw medical images may be normalized.
  • the normalization process includes, but is not limited to, normalizing the intensity or pixel values of the original medical image to a preset range (eg, a range of 0 to 1).
  • Step S114 Detecting the initial area of the first original medical image that is not marked with the target object, using the initial area of the target object marked on the second original medical image and the registration relationship between the second original medical image and the first original medical image, An initial region of the target object on the first original medical image is determined.
  • not all original medical images may be marked with the initial area of the target object. Therefore, in order to use more images to be classified including the target object to perform image object classification and improve the accuracy of image object classification, The initial area of the original medical image can be filled. After detecting the initial area of the first original medical image that is not marked with the target object, use the initial area of the target object marked on the second original medical image and the registration relationship between the second original medical image and the first original medical image to determine the first An initial region of the target object on the original medical image.
  • the above step of determining the initial area of the target object on the first original medical image may be performed by using a registration network.
  • the image target classification method may include several steps from the above steps S111 to S114.
  • the above steps S111 to S114 are only exemplary descriptions. In the disclosed embodiment, several steps can be selected to preprocess the original medical image as required, that is, the number of the above steps S111 to S114 can be arbitrarily selected, which is not specifically limited herein.
  • the original medical image can be preprocessed before the image to be classified is extracted from the original medical image, and the Image parameters to improve the quality of images to be classified.
  • the images to be classified including the target object can be extracted from the multiple original medical images respectively. For details, refer to steps S115 and S116 below.
  • Step S115 Determine the initial area of the target object in the original medical image, and expand the initial area according to a preset ratio to obtain the area to be extracted.
  • the characteristics of the target object itself are the main basis for judging its type, and there may be a variety of noise interference around the target object, which will mislead the classification of the target object.
  • the target object as a liver tumor as an example, the background of chronic liver disease or cirrhosis, other types of tumors, and blood vessels close to the liver tumor will affect the classification accuracy of the target object.
  • the area to be extracted is used as the area to be extracted, so that the area to be extracted contains the target object.
  • the initial area of the target object in the original medical image is determined.
  • the initial area may be expanded according to a preset ratio to obtain the area to be extracted.
  • the initial region is used to delineate the position of the target object in the original medical image.
  • an image segmentation technique can be used to determine the boundary contour of the target object in the original medical image, and mark the boundary contour to form an initial area.
  • Step S116 Extract the image data in the area to be extracted from the original medical image to obtain the image to be classified.
  • the image data is extracted from the original medical image by using the area to be extracted, and the obtained image to be classified includes the target object.
  • the original medical image can be a two-dimensional image or a three-dimensional image.
  • the image to be classified is a two-dimensional image.
  • the image to be classified may be a three-dimensional image, or the image to be classified may be a two-dimensional image.
  • the two-dimensional image of the layer where the target object has the largest area may be, but not limited to, the image to be classified.
  • the original medical image is a three-dimensional image
  • the image to be classified is a two-dimensional image obtained by extracting the layer where the maximum area of the target object is located in the original medical image, so that the layer where the maximum area of the target object is located in the original medical image can be extracted.
  • the extraction range of the target object in the to-be-classified image is larger and contains more information of the target object, thereby improving the classification accuracy of the target object.
  • the initial area of the target object in the original medical image is determined, and the initial area is expanded according to a preset ratio to obtain the area to be extracted; then the area to be extracted is extracted from the original medical image.
  • image data to obtain an image to be classified The initial area is the area containing the target object, and the initial area of the target object is expanded according to a preset ratio, so that the obtained area to be extracted contains both the target object and some background information around the target object, so that the image in the area to be extracted can be extracted.
  • the image to be classified can include the target object and some background information.
  • the to-be-classified images containing the target object are extracted from multiple original medical images respectively, so as to realize the acquisition of the to-be-classified images, and the to-be-classified images can be extracted from the original medical images.
  • the subsequent classification is reduced. It can avoid some background noise in the original medical image to a certain extent, so it can reduce the processing resource consumption of subsequent classification and improve the classification performance.
  • FIG. 4 is a schematic flowchart of a target classification for at least one image to be classified according to an embodiment of the present disclosure. Specifically, the following steps can be included:
  • Step S121 extracting several layers of features on at least one image to be classified, and correspondingly obtaining several sets of initial feature information.
  • the size of each group of initial feature information is different.
  • the number of layers for feature extraction may be one layer, two layers or even more layers.
  • Feature extraction can be implemented by convolutional layers, and each convolutional layer performs feature extraction on at least one image to be classified to obtain initial feature information.
  • which layers to perform feature extraction on can be obtained through manual settings, or can be determined through a large number of experiments when training a classification model, which is not specifically limited here.
  • a layer of feature extraction is performed on at least one image to be classified, and a set of initial feature information is correspondingly obtained, wherein the layer of feature extraction can be any layer, such as but not limited to the initial feature information obtained by extracting the last layer of features. as the basis for subsequent target classification.
  • Multi-layer feature extraction is performed on at least one image to be classified, and multiple sets of initial feature information are correspondingly obtained, wherein the multi-layer feature extraction may be continuous or discontinuous.
  • the initial feature information may be a feature map of the target object, reflecting the feature information of the target object in the image to be classified.
  • the size of each set of initial feature information is different, wherein the size includes dimension and/or resolution, so that the multiple sets of initial feature information respectively reflect different feature information of the target object.
  • the classification model is a deep learning network
  • the included network model structure can be an encoder or its variant, Resnet or its variant, VGG16 or its variant, or other network model structures for classification.
  • the classification model performs feature extraction on at least one image to be classified through a convolution layer, and different convolution layers correspond to different layers of feature extraction to obtain different groups of initial feature information.
  • the target object in the image to be classified, there may be noise interference around the target object.
  • the target object as a liver tumor as an example, the background of chronic liver disease or cirrhosis, other types of tumors, and blood vessels close to the liver tumor will affect the classification accuracy of the target object. Therefore, before using the classification model to perform target classification on at least one image to be classified to obtain the type of the target object, the final area of the target object can be obtained based on the initial area corresponding to the target object in the image to be classified.
  • an initial area can be used as the final area of the target object, or the final area of the target object can be obtained by combining the initial areas corresponding to the target object in at least one image to be classified.
  • the union of the initial regions corresponding to the target object in the image to be classified is regarded as the final region of the target object, which is not limited here.
  • the initial feature information of the image to be classified (such as the global features of the image to be classified, etc.) can be extracted.
  • the weight of the corresponding final area in the image to be classified is higher than the weight of other areas in the image to be classified, which makes the classification model tend to
  • the final region extracts features with richer details, so that the initial feature information output by the classification model corresponding to the final region can be as rich in features as possible; and/or, the features corresponding to the final region in the initial feature information are more abundant than the features of other regions.
  • the classification model is guided to pay more attention to the target object in the final area, so that the classification model can learn the target object.
  • the feature information of itself can reduce the influence of noise interference around the target object on the target classification.
  • a union of the initial regions corresponding to the target object in at least one image to be classified is obtained as the target object. Therefore, several layers of feature extraction can be performed on at least one image to be classified by using the final region, and several sets of initial feature information can be obtained correspondingly.
  • the final area of the target object is the union of the initial areas of the target object in the image to be classified, the final area is greater than or equal to any initial area, ensuring that the final area of the target object can contain the corresponding areas of the target objects in different images to be classified, Therefore, when the feature extraction of the image to be classified is performed, the feature information of the target object can be paid attention to as much as possible.
  • the at least one image to be classified includes a first image to be classified without an initial area of the target object and a second image to be classified with an initial area of the target object marked;
  • the classification model can also be used to detect the initial area of the first image to be classified that is not marked with the target object, and based on the initial area of the target object marked on the second image to be classified and the first image to be classified 2.
  • the registration relationship between the image to be classified and the first image to be classified determines the initial area of the target object on the first image to be classified.
  • the classification model can be used to determine the initial area of the target object for the first to-be-classified image that is not labeled with the initial area of the target object, so as to complete the labeling, so that the to-be-classified images all include the initial area.
  • a final area map including the final area of the target object may be generated, and the final area map and the image to be classified may be input into the classification model, so that the target classification is performed on at least one image to be classified by using the classification model,
  • the final region of the target object included in the final region map to perform several layers of feature extraction on at least one image to be classified can guide the network to pay more attention to the learning of the features of the final region and avoid the network to a certain extent. A lot of wrong feature information is learned, and the interference of noise around the target object on feature extraction is reduced. It can be understood that, before inputting the final area map and the image to be classified into the classification model, the sizes of the final area image and the image to be classified may be adjusted to a uniform size.
  • Step S122 Obtain final feature information based on at least one set of initial feature information in several sets of initial feature information.
  • Several layers of feature extraction are performed on at least one image to be classified, and after several sets of initial feature information are obtained, the final feature information can be obtained based on at least one set of initial feature information in several sets of initial feature information, and the selected initial feature information The information is different, and the final feature information obtained is different.
  • the number of groups of initial feature information and the parameter information such as the convolution layer corresponding to the classification model may be manually set, or may be determined during the training process of the classification model, which is not limited here. Fusion of multiple sets of initial feature information can improve the performance of the classification model and the accuracy of target classification, but the fusion of too much initial feature information will cause overfitting problems.
  • each set of initial feature information is different in dimension and resolution, and reflects different feature information of the target object, at least one set of initial feature information can be fused to obtain the final feature information.
  • the layer high-dimensional feature map is used as the final feature information, after multiple convolutions, some important feature information may be compressed, especially the target object with small area and blurred image features is missed.
  • Feature information the initial feature information obtained in different feature extraction stages can be spliced together to improve the accuracy of image target classification.
  • the weight of at least one set of initial feature information is used to fuse at least one set of initial feature information to obtain final feature information.
  • the weight of each set of initial feature information may be manually set, or may be determined during the training process of the classification model, which is not limited here. For example, first initialize the weight of each group of initial feature information, and continuously update the weight during the training process of the classification model. The above steps of updating weights are continuously repeated by using the training classification model, so that the training classification model continuously learns and updates the weight of each group of initial feature information, and obtains the trained classification model and the weight of each group of initial feature information.
  • the weights of each initial set of initial feature information may be the same or different, and the sum of the weights of each set of initial feature information is 1.
  • the weight of the initial feature information for fusion is determined, so that the final feature information obtained by using the weight fusion can better reflect the characteristics of the target object and further improve the classification performance.
  • the weights of different groups of initial feature information may be the same or different, and the sum of the weights of each group of initial feature information is 1.
  • the weight of the initial feature information can be used to fuse the initial feature information of different sizes extracted from at least one layer of features to obtain the final feature information, considering the initial feature information of smaller size.
  • a feature fusion network can be used to obtain final feature information based on at least one set of initial feature information in several sets of initial feature information, and initial feature information of multiple sizes can be spliced together as the final feature of the classification task.
  • each initial feature information is given a weight, and the weight is continuously updated during the model training process after initialization, so as to integrate multiple initial feature information to obtain a better feature representation of the target object, thereby improving the accuracy of target classification. performance.
  • each set of initial feature information may be converted into a preset dimension to facilitate subsequent acquisition of final feature information.
  • a feature extraction network is used to convert each set of initial feature information into a preset dimension.
  • the preset dimension can be set as required, for example, but not limited to, the preset dimension is one dimension.
  • Step S123 Classify the final feature information to obtain the type of the target object.
  • the final feature information carries the features of the target object, so that the final feature information is classified to obtain the type of the target object.
  • the classification model When determining the type of the target object, including but not limited to, the classification model performs target classification on at least one image to be classified, obtains the probability that the target object belongs to different types, and takes the type that satisfies the preset probability condition as the type of the target object.
  • the preset probability conditions include but are not limited to the maximum probability value and the like.
  • the classification model uses the ArcFace loss function to determine the loss value of the classification model during the training process, and the ArcFace loss function is used to shorten the distance of similar target objects and shorten the distance of heterogeneous target objects, thereby increasing the confusion of target objects. classification ability.
  • the ArcFace loss function is simple and easy to use, and can be well applied to the network structure of the classification model without being combined with other loss functions. At the same time, the overfitting problem is reduced to a certain extent, thereby improving the classification performance of the target object.
  • the training result of the classification model can be the cosine of the angle between the weight of the first fully connected layer and the feature entering the first fully connected layer. value.
  • the dot product between the features entering the first fully-connected layer of the classification model and the weights of the first fully-connected layer can be equal to the normalized cosine distance between the feature and the weight, so that the angular cosine function can be used to calculate the normalized
  • the target object as a liver tumor as an example, considering that the characteristic information of the liver tumor itself is the main basis for judging its type, but the size of the liver tumor varies, ranging from less than 0.5 cm to more than 20 cm, plus the size of the tumor outside the target object. Influencing factors, such as the low resolution of the image to be classified, other types of tumors around the liver tumor, blood vessels with similar characteristics to the target object, chronic liver disease or liver cirrhosis background, etc.
  • the ArcFace loss function can learn better The feature representation of liver tumors can realize the aggregation of similar tumors and the distance of heterogeneous tumors, and can effectively improve the classification performance of tumors.
  • the effect of using the ArcFace loss function to determine the loss value of the classification model in the training process of the classification model is similar, and no examples will be given here.
  • the ArcFace loss function is a loss function that uses margin to expand the distance between different classes.
  • the predicted value is the cosine of the angle between the weight of the first fully connected layer and the feature entering the first fully connected layer.
  • the principle and operation process are as follows: first, the dot product between the feature entering the first fully connected layer and the weight of the first fully connected layer is equal to the cosine distance after the normalization of the feature and the weight, and secondly, using the angular cosine function (arc-cosine function) to calculate the angle between normalized features and normalized weights; then, add an additional angular margin (additive angular margin) to the target angle, and then obtain through the cosine function
  • the logit of the target then rescales all logits with a fixed feature norm, and the subsequent steps are exactly the same as in the softmax loss.
  • the writing order of each step does not mean a strict execution order but constitutes any limitation on the implementation process, and the specific execution order of each step should be based on its function and possible Internal logic is determined.
  • At least one image to be classified is preprocessed, and a corresponding two-dimensional or three-dimensional multi-phase image tumor sub-image block is extracted, that is, a multi-phase image tumor patch image, and corresponding
  • the mask image, the mask patch image is fed together into a deep learning classification network.
  • FIG. 5 it is a schematic diagram of the network architecture used by the classification model in the image target classification method according to the embodiment of the present disclosure; wherein, the batch data input to the classification model 501 randomly includes data of different types of tumors in equal proportions, including stage Like 1, phase like 2, ..., phase like m, and multi-phase like the union of lesion masks.
  • 502 is the CNN backbone network, that is, CNN backbone, which can be the encoder of U-Net or its variant, Resnet or its variant, VGG16 or its variant, or other CNN structures for classification;
  • 503 is Feature Block, which includes Adaptive average pooling, FC and Relu; the previously obtained feature map is subjected to adaptive average pooling, full connection and Relu activation to obtain a one-dimensional feature; at the same time, each Feature Block corresponds to A feature_1.
  • Feature Fusion a feature fusion layer
  • Feature Fusion splices multiple one-dimensional features, each feature has a corresponding weight coefficient, and the coefficient can be learned; including: the weight coefficient of feature_1_ 1, weight coefficient_2 of feature_2, ..., weight coefficient_n of feature_n.
  • the feature map from any convolutional layer in the CNN backbone needs to enter the feature block and feature fusion layer, which can be determined by experimenting in the training process. In the experiment process of this scheme, it is found that the introduction of feature fusion layers can improve the performance of the model; however, fusing too many feature maps will cause over-fitting problems, especially the feature maps obtained by fusing the front convolutional layers.
  • 505 is a fully connected layer (Fully Connected), that is, the fused features are sent to FC, and converted into classification probability values of each tumor category through softmax.
  • 506 is the predicted probability value.
  • FIG. 6 is a schematic frame diagram of an image object classification apparatus 60 provided by an embodiment of the present disclosure.
  • the image object classification device 60 includes an image acquisition module 61 and an object classification module 62 .
  • the image acquisition module 61 is configured to: acquire at least one image to be classified including the target object, wherein at least one image to be classified is a medical image belonging to at least one scanned image category;
  • the target classification module 62 is configured to: use the classification The model performs target classification on at least one image to be classified to obtain the type of the target object.
  • the target classification module 62 is configured to: perform several layers of feature extraction on at least one image to be classified, and correspondingly obtain several sets of initial feature information; wherein, the size of each set of initial feature information is different; At least one group of initial feature information in the group of initial feature information is obtained to obtain final feature information; the final feature information is classified to obtain the type of the target object.
  • the target classification module 62 is configured to: obtain the final area of the target object based on the initial area corresponding to the target object in the image to be classified; correspondingly, the target classification module 62 is configured to: use the final area to At least one image to be classified is subjected to several layers of feature extraction, corresponding to several sets of initial feature information; wherein, in the feature extraction process, the weight of the corresponding final area in the image to be classified is higher than the weight of other areas in the image to be classified; and/ Or, the features corresponding to the final region in the initial feature information are more abundant than the features of other regions.
  • the target classification module 62 is configured to obtain a union of initial regions corresponding to the target object in at least one image to be classified, as the final region of the target object.
  • the target classification module 62 is configured to: use the classification model to detect the initial area of the first image to be classified that is not marked with the target object, and based on the initial area of the target object marked on the second image to be classified and the registration relationship between the second to-be-classified image and the first to-be-classified image to determine the initial area of the target object on the first to-be-classified image.
  • the target classification module 62 is configured to: convert each set of initial feature information into a preset dimension; and/or, the target classification module 62 is configured to: use the weight of at least one set of initial feature information to classify At least one set of initial feature information is fused to obtain final feature information.
  • the weight of each set of initial feature information is determined during the training process of the classification model.
  • the preset dimension is one dimension.
  • the classification model adopts the ArcFace loss function during the training process to determine the loss value of the classification model; and/or, the batch sample data selected for each training of the classification model is selected from the sample data set using a data generator
  • the number of different target types is a preset ratio of sample data.
  • the image acquisition module 61 is configured to: extract images to be classified including the target object from a plurality of original medical images, respectively.
  • the image acquisition module 61 is configured to: determine the initial area of the target object in the original medical image, expand the initial area according to a preset ratio to obtain the area to be extracted; extract the area to be extracted from the original medical image image data to obtain the image to be classified.
  • the image acquisition module 61 is configured to: resample the original medical image to a preset resolution; adjust the range of pixel values in the original medical image; normalize the original medical image; detect The first original medical image is not marked with the initial area of the target object, and the first original medical image is determined by using the initial area of the target object marked on the second original medical image and the registration relationship between the second original medical image and the first original medical image. The initial area of the target object on the image.
  • the original medical image and the image to be classified are two-dimensional images; or, the original medical image is a three-dimensional image, and the image to be classified is a two-dimensional image or a three-dimensional image.
  • the object classification module 62 uses the classification model to perform object classification on the at least one image to be classified to obtain the target object Therefore, an image target classification method based on artificial intelligence technology is proposed to achieve intelligent target classification. Because the classification model is used to classify the images to be classified, it not only makes the target classification process simpler, reduces the dependence on doctors, and improves the speed of target classification, but also combines artificial intelligence technology to achieve target classification, so as to assist doctors in intelligent disease diagnosis and treatment.
  • FIG. 7 is a schematic diagram of a framework of an embodiment of an electronic device 70 of the present disclosure.
  • the electronic device 70 includes a memory 71 and a processor 72 coupled to each other, and the processor 72 is configured to execute program instructions stored in the memory 71 to implement the steps of any of the image object classification method embodiments described above.
  • the electronic device 70 may include, but is not limited to, a microcomputer and a server.
  • the electronic device 70 may also include mobile devices such as a notebook computer and a tablet computer, which are not limited herein.
  • the processor 72 is configured to control itself and the memory 71 to implement the steps of any of the image object classification method embodiments described above.
  • the processor 72 may also be referred to as a central processing unit (Central Processing Unit, CPU).
  • the processor 72 may be an integrated circuit chip with signal processing capability.
  • the processor 72 may also be a general-purpose processor, a digital signal processor (Digital Signal Processor, DSP), an application-specific integrated circuit (Application Specific Integrated Circuit, ASIC), a Field-Programmable Gate Array (Field-Programmable Gate Array, FPGA) or other Programmable logic devices, discrete gate or transistor logic devices, discrete hardware components.
  • a general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
  • the processor 72 may be jointly implemented by an integrated circuit chip.
  • the electronic device 70 after acquiring at least one image to be classified including a target object, uses a classification model to classify the at least one image to be classified, and obtains the type of the target object. Therefore, an artificial intelligence technology-based method is proposed.
  • Image object classification method to achieve intelligent object classification. Because the classification model is used to classify the images to be classified, it not only makes the target classification process simpler, reduces the dependence on doctors, and improves the speed of target classification, but also combines artificial intelligence technology to achieve target classification, so as to assist doctors in intelligent disease diagnosis and treatment.
  • FIG. 8 is a schematic diagram of a framework of an embodiment of a computer-readable storage medium 80 of the present disclosure.
  • the computer-readable storage medium 80 stores program instructions 801 that can be executed by the processor, and the program instructions 801 are used to implement the steps of any of the foregoing image object classification method embodiments.
  • the classification model is used to classify the at least one image to be classified to obtain the type of the target object.
  • the image target classification method of intelligent technology realizes intelligent target classification. Because the classification model is used to classify the images to be classified, it not only makes the target classification process simpler, reduces the dependence on doctors, and improves the speed of target classification, but also combines artificial intelligence technology to achieve target classification, so as to assist doctors in intelligent disease diagnosis and treatment.
  • An embodiment of the present disclosure further provides a computer program, where the computer program includes a computer-readable code, and when the computer-readable code is run in an electronic device, the processor of the electronic device executes the program for implementing any of the foregoing embodiments.
  • Image object classification methods can be specifically implemented by hardware, software or a combination thereof.
  • the computer program product is embodied as a computer storage medium, and in some embodiments of the present disclosure, the computer program product is embodied as a software product, such as a software development kit (Software Development Kit, SDK), etc. .
  • the functions or modules included in the image object classification apparatus provided in the embodiments of the present disclosure may be used to execute the methods described in the above method embodiments, and the specific implementation may refer to the above method embodiments. It is concise and will not be repeated here.
  • the disclosed method and apparatus may be implemented in other manners.
  • the device implementations described above are only illustrative.
  • the division of modules or units is only a logical function division. In actual implementation, there may be other divisions.
  • units or components may be combined or integrated. to another system, or some features can be ignored, or not implemented.
  • the shown or discussed mutual coupling or direct coupling or communication connection may be through some interfaces, indirect coupling or communication connection of devices or units, which may be in electrical, mechanical or other forms.
  • each functional unit in each embodiment of the present disclosure may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit.
  • the above-mentioned integrated units may be implemented in the form of hardware, or may be implemented in the form of software functional units.
  • the integrated unit if implemented as a software functional unit and sold or used as a stand-alone product, may be stored in a computer-readable storage medium.
  • the technical solutions of the present disclosure can be embodied in the form of software products in essence, or the part that contributes to the prior art, or all or part of the technical solutions, and the computer software product is stored in a storage medium , including several instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) or a processor (processor) to execute all or part of the steps of the methods of the various embodiments of the present disclosure.
  • the aforementioned storage medium includes: U disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk and other media that can store program codes .
  • the present disclosure provides an image object classification method, device, device, storage medium and program, wherein the image object classification method includes: acquiring at least one image to be classified including a target object, wherein the at least one image to be classified is a medical image belonging to at least one type of scanned image; using a classification model, perform target classification on the at least one image to be classified to obtain the type of the target object.

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

Disclosed in embodiments are an image object classification method and apparatus, a device, a storage medium and a program. The image object classification method comprises: acquiring at least one image to be classified that includes a target object, wherein the at least one image is a medical image belonging to at least one type of scan images; and performing object classification on the at least one image by using a classification model to obtain the type of the target object. The described solution can be applied to a medical image including at least one phase of a tumor, so as to determine the type of the tumor in the medical image, that is, can realize intelligent object classification and improve object classification efficiency.

Description

图像目标分类方法、装置、设备、存储介质及程序Image object classification method, device, equipment, storage medium and program
相关申请的交叉引用CROSS-REFERENCE TO RELATED APPLICATIONS
本公开基于申请号为202011212261.1、申请日为2020年11月03日的中国专利申请提出,并要求该中国专利申请的优先权,该中国专利申请的全部内容在此引入本公开作为参考。The present disclosure is based on a Chinese patent application with application number 202011212261.1 and an application date of November 03, 2020, and claims the priority of the Chinese patent application, the entire contents of which are incorporated herein by reference.
技术领域technical field
本公开涉及图像处理技术领域,特别是涉及一种图像目标分类方法、装置、设备、存储介质及程序。The present disclosure relates to the technical field of image processing, and in particular, to an image object classification method, apparatus, device, storage medium and program.
背景技术Background technique
计算机断层扫描(Computed Tomography,CT)和核磁共振(Magnetic Resonance)等医学图像在临床具有重要意义。以与肝脏相关的临床为例,扫描图像类别往往包括与时序有关的造影前平扫、动脉早期、动脉晚期、门脉期以及延迟期等,此外,扫描图像类别还可以包含与扫描参数有关的T1加权反相成像、T1加权同相成像、T2加权成像、扩散加权成像以及表面扩散系数成像等。通过对医学图像的判别,有助于临床医生对疾病的了解。Medical images such as Computed Tomography (CT) and Magnetic Resonance (Magnetic Resonance) are of great clinical significance. Taking the clinical practice related to the liver as an example, the scan image categories often include time-related pre-contrast scan, early arterial phase, late arterial phase, portal venous phase, and delayed phase, etc. In addition, the scan image category can also include scan parameters related to the T1-weighted inverse imaging, T1-weighted in-phase imaging, T2-weighted imaging, diffusion-weighted imaging, and surface diffusion coefficient imaging, etc. Through the identification of medical images, it is helpful for clinicians to understand the disease.
相关技术中,在疾病诊疗过程中,通常需要医生反复查看肿瘤等目标对象在医学图像上的征象,这样使得确定肿瘤所属类型过度依赖于医生的专业水平,同时存在确定肿瘤所属类型效率低下的问题。In the related art, in the process of disease diagnosis and treatment, the doctor usually needs to repeatedly check the signs of the target object such as the tumor on the medical image, which makes the determination of the type of the tumor overly dependent on the professional level of the doctor, and the problem of low efficiency in determining the type of the tumor. .
发明内容SUMMARY OF THE INVENTION
本公开至少提供一种图像目标分类方法、装置、设备、存储介质及程序。The present disclosure provides at least an image object classification method, apparatus, device, storage medium and program.
本公开实施例提供了一种图像目标分类方法,该图像目标分类方法包括:The embodiment of the present disclosure provides an image object classification method, and the image object classification method includes:
获取包含目标对象的至少一张待分类图像,其中,所述至少一张待分类图像为属于至少一种扫描图像类别的医学图像;acquiring at least one image to be classified containing the target object, wherein the at least one image to be classified is a medical image belonging to at least one scanned image category;
利用分类模型,对所述至少一张待分类图像进行目标分类,得到所述目标对象的类型。如此,获取包含目标对象的至少一张待分类图像后,利用分类模型对至少一张待分类图像进行目标分类,得到目标对象的类型,由于利用分类模型对待分类图像进行目标分类,实现了智能化目标分类,且无需人工进行目标分类,可减小对人工依赖,提高目标分类效率。Using a classification model, the at least one image to be classified is subjected to target classification to obtain the type of the target object. In this way, after obtaining at least one image to be classified including the target object, the classification model is used to classify the at least one image to be classified, and the type of the target object is obtained. Because the classification model is used to classify the image to be classified, intelligence is realized. Target classification, and no manual target classification is required, which can reduce the dependence on manual work and improve the efficiency of target classification.
在本公开的一些实施例中,所述对所述至少一张待分类图像进行目标分类,得到所述目标对象的类型,包括:In some embodiments of the present disclosure, performing target classification on the at least one image to be classified to obtain the type of the target object includes:
对所述至少一张待分类图像进行若干层特征提取,对应得到若干组初始特征信息;其中,每组所述初始特征信息的尺寸不同;Performing several layers of feature extraction on the at least one image to be classified, correspondingly obtaining several groups of initial feature information; wherein, the size of each group of the initial feature information is different;
基于所述若干组初始特征信息中的至少一组初始特征信息,得到最终特征信息;Obtain final feature information based on at least one set of initial feature information in the several sets of initial feature information;
对所述最终特征信息进行分类,得到所述目标对象的类型。The final feature information is classified to obtain the type of the target object.
如此,通过特征提取得到初始特征信息,从而基于初始特征信息得到最终特征信息后,则可对最终特征信息进行分类,得到目标对象的类型,故实现了利用目标对象的特征信息进行目标分类。In this way, after the initial feature information is obtained through feature extraction, and the final feature information is obtained based on the initial feature information, the final feature information can be classified to obtain the type of the target object, so the target classification is realized by using the feature information of the target object.
在本公开的一些实施例中,所述对所述至少一张待分类图像进行目标分类,得到所述目标对象的类型之前,所述方法还包括:In some embodiments of the present disclosure, before the target classification is performed on the at least one image to be classified and the type of the target object is obtained, the method further includes:
基于所述待分类图像中所述目标对象对应的初始区域,得到所述目标对象的最终区域;obtaining the final area of the target object based on the initial area corresponding to the target object in the image to be classified;
相应地,所述对所述至少一张待分类图像进行若干层特征提取,对应得到若干组初始特征信息,包括:Correspondingly, several layers of feature extraction are performed on the at least one image to be classified, and several sets of initial feature information are correspondingly obtained, including:
利用所述最终区域对所述至少一张待分类图像进行若干层特征提取,对应得到若干组初始特征信息;其中,在特征提取过程中,所述待分类图像中对应所述最终区域的权重高于所述待分类图像中其他区域的权重;和/或,所述初始特征信息中对应所述最终区域的特征比其他区域的特征更丰富。Using the final area to perform several layers of feature extraction on the at least one image to be classified, correspondingly obtain several sets of initial feature information; wherein, in the feature extraction process, the weight of the image to be classified corresponding to the final area is high the weight of other regions in the image to be classified; and/or, the features corresponding to the final region in the initial feature information are more abundant than the features of other regions.
如此,利用最终区域对待分类图像进行特征提取时,待分类图像中对应最终区域的权重高于待分类图像中其他区域的权重,故使得分类模型趋向于对最终区域提取细节更丰富的特征;和/或,初始特征信息中对应最终区域的特征比其他区域的特征更丰富;由此使得分类模型利用待分类图像的初始特征信息,能够更能学习到目标对象本身的特征信息,在一定程度上减小目标对象周围噪声干扰对目标分类的影响。In this way, when using the final region to perform feature extraction on the image to be classified, the weight of the corresponding final region in the image to be classified is higher than the weights of other regions in the image to be classified, so that the classification model tends to extract features with more details for the final region; and /or, the features corresponding to the final region in the initial feature information are richer than those of other regions; thus, the classification model can learn the feature information of the target object itself better by using the initial feature information of the image to be classified, to a certain extent Reduce the impact of noise interference around the target object on target classification.
在本公开的一些实施例中,所述基于所述待分类图像中所述目标对象对应的初始区域,得到所述目标对象的最终区域,包括:In some embodiments of the present disclosure, obtaining the final area of the target object based on the initial area corresponding to the target object in the image to be classified includes:
获取所述至少一张待分类图像中所述目标对象对应的初始区域的并集,以作为所述目标对象的最终区域。The union of the initial regions corresponding to the target object in the at least one image to be classified is obtained as the final region of the target object.
如此,在目标对象的最终区域是待分类图像中目标对象的初始区域的并集的情况下,最终区域大于或等于任意一个初始区域,保证目标对象的最终区域能够包含不同待分类图像中的目标对象对应区域,从而在对待分类图像进行特征提取时,能够尽可能关注目标对象的特征信息。In this way, when the final area of the target object is the union of the initial areas of the target object in the image to be classified, the final area is greater than or equal to any initial area, ensuring that the final area of the target object can contain different targets in the to-be-classified image. The object corresponds to the area, so that the feature information of the target object can be paid attention to as much as possible when the feature extraction of the image to be classified is performed.
在本公开的一些实施例中,所述至少一张待分类图像包括未标注所述目标对象的初始区域的第一待分类图像和标注所述目标对象的初始区域的第二待分类图像;所述基于所述待分类图像中所述目标对象对应的初始区域,得到所述目标对象的最终区域之前,所述方法还包括:In some embodiments of the present disclosure, the at least one image to be classified includes a first image to be classified without an initial region of the target object and a second image to be classified with an initial region of the target object labeled; Before obtaining the final area of the target object based on the initial area corresponding to the target object in the image to be classified, the method further includes:
利用所述分类模型检测到所述第一待分类图像未标注有所述目标对象的初始区域,并基于所述第二待分类图像上标注的所述目标对象的初始区域以及所述第二待分类图像与所述第一待分类图像的配准关系,确定所述第一待分类图像上所述目标对象的初始区域。Using the classification model to detect that the first image to be classified is not marked with the initial area of the target object, and based on the initial area of the target object marked on the second to-be-classified image and the second to-be-classified image The registration relationship between the classified image and the first to-be-classified image determines the initial area of the target object on the first to-be-classified image.
如此,可以利用分类模型为未标注目标对象初始区域的第一待分类图像确定目标对象的初始区域,从而补齐标注,使得待分类图像中均包括初始区域。In this way, the classification model can be used to determine the initial area of the target object for the first to-be-classified image that is not labeled with the initial area of the target object, so as to complete the labeling, so that the to-be-classified images all include the initial area.
在本公开的一些实施例中,所述基于所述若干组初始特征信息中的至少一组初始特征信息,得到最终特征信息之前,所述方法还包括:In some embodiments of the present disclosure, before obtaining the final feature information based on at least one set of initial feature information in the several sets of initial feature information, the method further includes:
将每组所述初始特征信息转换为预设维度;converting each group of the initial feature information into a preset dimension;
和/或,所述基于所述若干组初始特征信息中的至少一组初始特征信息,得到最终特征信息,包括:And/or, obtaining final feature information based on at least one set of initial feature information in the several groups of initial feature information, including:
利用所述至少一组初始特征信息的权重,将所述至少一组初始特征信息进行融合,得到所述最终特征信息。Using the weight of the at least one set of initial feature information, the at least one set of initial feature information is fused to obtain the final feature information.
如此,将每组初始特征信息统一转换为预设维度,方便后续最终特征信息的获取。另外,由于每组初始特征信息均反映了目标对象的特征,可以利用至少一组初始特征信息的权重,将至少一层特征提取的不同尺寸的初始特征信息进行融合,得到最终特征信息,考虑较小尺寸的初始特征信息可能被压缩掉重要特征,通过综合不同尺寸的特征信息,能够得到较为综合和有用的最终特征信息,进而提高后续分类性能。In this way, each group of initial feature information is uniformly converted into a preset dimension, which facilitates subsequent acquisition of final feature information. In addition, since each set of initial feature information reflects the characteristics of the target object, the weight of at least one set of initial feature information can be used to fuse the initial feature information of different sizes extracted from at least one layer of features to obtain the final feature information. The initial feature information of small size may be compressed to remove important features. By synthesizing feature information of different sizes, more comprehensive and useful final feature information can be obtained, thereby improving the subsequent classification performance.
在本公开的一些实施例中,每组所述初始特征信息的权重是在所述分类模型训练过程确定的。In some embodiments of the present disclosure, the weight of each set of the initial feature information is determined during the training process of the classification model.
如此,通过分类模型的迭代训练,来确定用于融合的初始特征信息的权重,以使得利用该权重融合得到的最终特征信息更能反映目标对象特征,进一步提高分类性能。In this way, the weight of the initial feature information for fusion is determined through the iterative training of the classification model, so that the final feature information obtained by using the weight fusion can better reflect the characteristics of the target object and further improve the classification performance.
在本公开的一些实施例中,所述预设维度为一维。In some embodiments of the present disclosure, the preset dimension is one dimension.
如此,可将每组初始特征信息转换为一维,实现数据统一化,而且便于后续融合。In this way, each group of initial feature information can be converted into one-dimensional, data unification is realized, and subsequent fusion is facilitated.
在本公开的一些实施例中,所述分类模型在训练过程中采用ArcFace损失函数确定所述分类模型的损失值;和/或,所述分类模型每次训练选择的批样本数据是利用数据生成器从样本数据集中选择的不同目标类型的数量为预设比例的样本数据。In some embodiments of the present disclosure, the classification model uses an ArcFace loss function during the training process to determine the loss value of the classification model; and/or, the batch sample data selected for each training of the classification model is generated by using the data The number of different target types selected by the processor from the sample data set is a preset proportion of sample data.
如此,采用ArcFace损失函数确定分类模型的损失值,可使得同类目标对象的特征信息聚合、不同类目标对象的特征信息远离,进而提高目标对象的分类性能。另外,利用数据生成器从样本数据集中选择样本数据,将不同目标类型的数量为预设比例的样本数据作为批样本数据,使得训练分类模型的批样本数据的目标类型更均衡。In this way, using the ArcFace loss function to determine the loss value of the classification model can aggregate the feature information of the same target objects and keep the feature information of different types of target objects away, thereby improving the classification performance of the target objects. In addition, the data generator is used to select sample data from the sample data set, and the sample data with a preset ratio of different target types is used as the batch sample data, so that the target types of the batch sample data for training the classification model are more balanced.
在本公开的一些实施例中,所述获取包含目标对象的至少一张待分类图像,包括:In some embodiments of the present disclosure, the acquiring at least one image to be classified including the target object includes:
分别从多张原始医学图像提取得到包含所述目标对象的待分类图像。The to-be-classified images containing the target object are respectively extracted from a plurality of original medical images.
如此,实现待分类图像的获取,而且待分类图像可从原始医学图像中提取得到,相比直接采用原始医学图像,减少后续分类的图像尺寸,而且可一定程度上避免原始医学图像中的一些背景噪声,故可减少后续分类的处理资源损耗,且提高分类性能。In this way, the image to be classified is obtained, and the image to be classified can be extracted from the original medical image. Compared with directly using the original medical image, the image size of the subsequent classification can be reduced, and some backgrounds in the original medical image can be avoided to a certain extent. Therefore, the consumption of processing resources for subsequent classification can be reduced, and the classification performance can be improved.
在本公开的一些实施例中,所述分别从多张原始医学图像提取得到包含所述目标对象的待分类图像,包括:In some embodiments of the present disclosure, the image to be classified containing the target object is extracted from multiple original medical images, including:
确定所述原始医学图像中所述目标对象的初始区域,按照所述预设比例扩大所述初始区域,得到待提取区域;determining the initial area of the target object in the original medical image, and expanding the initial area according to the preset ratio to obtain the area to be extracted;
从所述原始医学图像中提取所述待提取区域中的图像数据,得到所述待分类图像。The image data in the to-be-extracted area is extracted from the original medical image to obtain the to-be-classified image.
如此,初始区域是包含目标对象的区域,而按照预设比例扩大目标对象的初始区域,使得得到的待提取区域既包含目标对象,又包含目标对象周围的部分背景信息,以便将待提取区域中的图像数据提取作为待分类图像后,待分类图像能够囊括目标对象和部分背景信息。In this way, the initial area is the area containing the target object, and the initial area of the target object is expanded according to a preset ratio, so that the obtained area to be extracted contains both the target object and some background information around the target object, so that the area to be extracted can be extracted. After the image data is extracted as the image to be classified, the image to be classified can include the target object and some background information.
在本公开的一些实施例中,所述分别从多张原始医学图像提取得到包含所述目标对象的待分类图像之前,所述方法还包括以下至少一个步骤:In some embodiments of the present disclosure, before the image to be classified including the target object is extracted from multiple original medical images, the method further includes at least one of the following steps:
将所述原始医学图像重采样至预设分辨率;resampling the original medical image to a preset resolution;
调整所述原始医学图像中的像素值范围;adjusting the range of pixel values in the original medical image;
将所述原始医学图像进行归一化处理;normalizing the original medical image;
检测到第一原始医学图像未标注有所述目标对象的初始区域,利用第二原始医学图像上标注的所述目标对象的初始区域以及所述第二原始医学图像与所述第一原始医学图像的配准关系,确定所述第一原始医学图像上所述目标对象的初始区域。It is detected that the initial area of the target object is not marked in the first original medical image, and the initial area of the target object marked on the second original medical image and the second original medical image and the first original medical image are used. The registration relationship is determined, and the initial area of the target object on the first original medical image is determined.
如此,通过统一分辨率、调整像素值范围、归一化处理、以及确定目标对象的初始区域等操作,可在从原始医学图像提取待分类图像之前,对原始医学图像进行预处理,统一待分类图像的图像参数,提高待分类图像的质量。In this way, by unifying the resolution, adjusting the pixel value range, normalizing, and determining the initial area of the target object, the original medical image can be preprocessed before the image to be classified is extracted from the original medical image, and the unclassified image can be unified. Image parameters of the image to improve the quality of the image to be classified.
在本公开的一些实施例中,所述原始医学图像和所述待分类图像为二维图像;或者,所述原始医学图像为三维图像,所述待分类图像为二维图像或三维图像。In some embodiments of the present disclosure, the original medical image and the image to be classified are two-dimensional images; or, the original medical image is a three-dimensional image, and the image to be classified is a two-dimensional image or a three-dimensional image.
如此,待分类图像是从原始医学图像中提取得到的,在原始医学图像为二维图像的情况下,待分类图像为二维图像;而在原始医学图像为三维图像的情况下,待分类图像的维度可以为二维或三维。In this way, the image to be classified is extracted from the original medical image. If the original medical image is a two-dimensional image, the image to be classified is a two-dimensional image; and if the original medical image is a three-dimensional image, the image to be classified is a two-dimensional image. The dimensions can be two-dimensional or three-dimensional.
在本公开的一些实施例中,所述原始医学图像为三维图像,所述待分类图像为对所述原始医学图像中所述目标对象最大面积所在层提取得到的二维图像。In some embodiments of the present disclosure, the original medical image is a three-dimensional image, and the image to be classified is a two-dimensional image obtained by extracting a layer where the target object has the largest area in the original medical image.
如此,在原始医学图像为三维图像、待分类图形为二维图像的情况下,可以提取原始医学图像中目标对象最大面积所在层作为待分类图像,使得待分类图像中目标对象的提取范围较大,包含目标对象的信息更多,提高目标对象的分类精度。In this way, when the original medical image is a three-dimensional image and the graphic to be classified is a two-dimensional image, the layer where the maximum area of the target object is located in the original medical image can be extracted as the image to be classified, so that the extraction range of the target object in the image to be classified is larger. , contains more information about the target object and improves the classification accuracy of the target object.
以下装置、电子设备等的效果描述参见上述方法的说明,这里不再赘述。For descriptions of the effects of the following apparatuses, electronic devices, etc., reference may be made to the descriptions of the above-mentioned methods, which will not be repeated here.
本公开实施例还提供了一种图像目标分类装置,该图像目标分类装置包括:The embodiment of the present disclosure also provides an image object classification device, and the image object classification device includes:
图像获取模块,配置为获取包含目标对象的至少一张待分类图像,其中,所述至少一张待分类图像为属于至少一种扫描图像类别的医学图像;an image acquisition module, configured to acquire at least one image to be classified including the target object, wherein the at least one image to be classified is a medical image belonging to at least one scanned image category;
目标分类模块,配置为利用分类模型,对所述至少一张待分类图像进行目标分类,得到所述目标对象的类型。The target classification module is configured to use a classification model to perform target classification on the at least one image to be classified to obtain the type of the target object.
在本公开的一些实施例中,目标分类模块,配置为:In some embodiments of the present disclosure, the target classification module is configured to:
对所述至少一张待分类图像进行若干层特征提取,对应得到若干组初始特征信息;其中,每组所述初始特征信息的尺寸不同;Performing several layers of feature extraction on the at least one image to be classified, correspondingly obtaining several groups of initial feature information; wherein, the size of each group of the initial feature information is different;
基于所述若干组初始特征信息中的至少一组初始特征信息,得到最终特征信息;Obtain final feature information based on at least one set of initial feature information in the several sets of initial feature information;
对所述最终特征信息进行分类,得到所述目标对象的类型。The final feature information is classified to obtain the type of the target object.
在本公开的一些实施例中,目标分类模块,配置为:In some embodiments of the present disclosure, the target classification module is configured to:
基于所述待分类图像中所述目标对象对应的初始区域,得到所述目标对象的最终区域;obtaining the final area of the target object based on the initial area corresponding to the target object in the image to be classified;
相应地,目标分类模块,配置为:Accordingly, the target classification module, configured as:
利用所述最终区域对所述至少一张待分类图像进行若干层特征提取,对应得到若干组初始特征信息;其中,在特征提取过程中,所述待分类图像中对应所述最终区域的权重高于所述待分类图像中其他区域的权重;和/或,所述初始特征信息中对应所述最终区域的特征比其他区域的特征更丰富。Using the final area to perform several layers of feature extraction on the at least one image to be classified, correspondingly obtain several sets of initial feature information; wherein, in the feature extraction process, the weight of the image to be classified corresponding to the final area is high the weight of other regions in the image to be classified; and/or, the features corresponding to the final region in the initial feature information are more abundant than the features of other regions.
在本公开的一些实施例中,目标分类模块,配置为:In some embodiments of the present disclosure, the target classification module is configured to:
获取所述至少一张待分类图像中所述目标对象对应的初始区域的并集,以作为所述目标对象的最终区域。The union of the initial regions corresponding to the target object in the at least one image to be classified is obtained as the final region of the target object.
在本公开的一些实施例中,目标分类模块,配置为:In some embodiments of the present disclosure, the target classification module is configured to:
利用所述分类模型检测到所述第一待分类图像未标注有所述目标对象的初始区域,并基于所述第二待分类图像上标注的所述目标对象的初始区域以及所述第二待分类图像与所述第一待分类图像的配准关系,确定所述第一待分类图像上所述目标对象的初始区域。Using the classification model to detect that the first image to be classified is not marked with the initial area of the target object, and based on the initial area of the target object marked on the second to-be-classified image and the second to-be-classified image The registration relationship between the classified image and the first to-be-classified image determines the initial area of the target object on the first to-be-classified image.
在本公开的一些实施例中,目标分类模块,配置为:In some embodiments of the present disclosure, the target classification module is configured to:
将每组所述初始特征信息转换为预设维度;converting each group of the initial feature information into a preset dimension;
和/或,目标分类模块,配置为:and/or, the target classification module, configured as:
利用所述至少一组初始特征信息的权重,将所述至少一组初始特征信息进行融合,得到所述最终特征信息。Using the weight of the at least one set of initial feature information, the at least one set of initial feature information is fused to obtain the final feature information.
在本公开的一些实施例中,每组所述初始特征信息的权重是在所述分类模型训练过程确定的。In some embodiments of the present disclosure, the weight of each set of the initial feature information is determined during the training process of the classification model.
在本公开的一些实施例中,所述预设维度为一维。In some embodiments of the present disclosure, the preset dimension is one dimension.
在本公开的一些实施例中,所述分类模型在训练过程中采用ArcFace损失函数确定所述分类模型的损失值;和/或,所述分类模型每次训练选择的批样本数据是利用数据生成器从样本数据集中选择的不同目标类型的数量为预设比例的样本数据。In some embodiments of the present disclosure, the classification model uses an ArcFace loss function during the training process to determine the loss value of the classification model; and/or, the batch sample data selected for each training of the classification model is generated by using the data The number of different target types selected by the processor from the sample data set is a preset proportion of sample data.
在本公开的一些实施例中,图像获取模块,配置为:In some embodiments of the present disclosure, the image acquisition module is configured to:
分别从多张原始医学图像提取得到包含所述目标对象的待分类图像。The to-be-classified images containing the target object are respectively extracted from a plurality of original medical images.
在本公开的一些实施例中,图像获取模块,配置为:In some embodiments of the present disclosure, the image acquisition module is configured to:
确定所述原始医学图像中所述目标对象的初始区域,按照所述预设比例扩大所述初始区域,得到待提取区域;determining the initial area of the target object in the original medical image, and expanding the initial area according to the preset ratio to obtain the area to be extracted;
从所述原始医学图像中提取所述待提取区域中的图像数据,得到所述待分类图像。The image data in the to-be-extracted area is extracted from the original medical image to obtain the to-be-classified image.
在本公开的一些实施例中,图像获取模块,配置为:In some embodiments of the present disclosure, the image acquisition module is configured to:
将所述原始医学图像重采样至预设分辨率;resampling the original medical image to a preset resolution;
调整所述原始医学图像中的像素值范围;adjusting the range of pixel values in the original medical image;
将所述原始医学图像进行归一化处理;normalizing the original medical image;
检测到第一原始医学图像未标注有所述目标对象的初始区域,利用第二原始医学图像上标注的所述目标对象的初始区域以及所述第二原始医学图像与所述第一原始医学图像的配准关系,确定所述第一原始医学图像上所述目标对象的初始区域。It is detected that the initial area of the target object is not marked in the first original medical image, and the initial area of the target object marked on the second original medical image and the second original medical image and the first original medical image are used. The registration relationship is determined, and the initial area of the target object on the first original medical image is determined.
在本公开的一些实施例中,所述原始医学图像和所述待分类图像为二维图像;或者,所述原始医学图像为三维图像,所述待分类图像为二维图像或三维图像。In some embodiments of the present disclosure, the original medical image and the image to be classified are two-dimensional images; or, the original medical image is a three-dimensional image, and the image to be classified is a two-dimensional image or a three-dimensional image.
在本公开的一些实施例中,所述原始医学图像为三维图像,所述待分类图像为对所述原始医学图像中所述目标对象最大面积所在层提取得到的二维图像。In some embodiments of the present disclosure, the original medical image is a three-dimensional image, and the image to be classified is a two-dimensional image obtained by extracting a layer where the target object has the largest area in the original medical image.
本公开实施例还提供了一种电子设备,包括相互耦接的存储器和处理器,所述处理器用于执行所述存储器中存储的程序指令,以实现如前任一实施例提供的图像目标分类方法。An embodiment of the present disclosure also provides an electronic device, including a memory and a processor coupled to each other, where the processor is configured to execute program instructions stored in the memory, so as to implement the image object classification method provided in any of the previous embodiments .
本公开实施例还提供了一种计算机可读存储介质,其上存储有程序指令,程序指令被处理器执行如前任一实施例提供的图像目标分类方法。The embodiments of the present disclosure also provide a computer-readable storage medium, which stores program instructions, and the program instructions are executed by a processor as the image object classification method provided in any of the previous embodiments.
本公开实施例还提供了一种计算机程序,所述计算机程序包括计算机可读代码,在所述计算机可读代码在电子设备中运行的情况下,所述电子设备的处理器执行如前任一实施例所述的图像目标分类方法。An embodiment of the present disclosure also provides a computer program, the computer program includes computer-readable codes, and when the computer-readable codes are executed in an electronic device, the processor of the electronic device executes any of the preceding implementations The image object classification method described in the example.
本公开实施例提供的一种图像目标分类方法、装置、设备、存储介质及程序,获取包含目标对象的至少一张待分类图像后,利用分类模型对至少一张待分类图像进行目标分类,得到目标对象的类型,因此提出基于人工智能技术的图像目标分类方法,实现智能化目标分类。由于利用分类模型对待分类图像进行目标分类,不仅使得目标分类过程更加简单,减小对医生的依赖,提高目标分类速度和准确性,而且结合人工智能技术实现目标分类,以便辅助医生进行智能化疾病诊疗。In an image object classification method, device, device, storage medium, and program provided by the embodiments of the present disclosure, after acquiring at least one image to be classified that includes a target object, the classification model is used to classify the at least one image to be classified, and the result is obtained Therefore, an image target classification method based on artificial intelligence technology is proposed to achieve intelligent target classification. Because the classification model is used to classify the images to be classified, it not only makes the target classification process simpler, reduces the dependence on doctors, improves the speed and accuracy of target classification, but also combines artificial intelligence technology to achieve target classification, so as to assist doctors in intelligent diseases diagnosis and treatment.
应当理解的是,以上的一般描述和后文的细节描述仅是示例性和解释性的,而非限制本公开。It is to be understood that the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the present disclosure.
附图说明Description of drawings
此处的附图被并入说明书中并构成本说明书的一部分,这些附图示出了符合本公开的实施例,并与说明书一起用于说明本公开的技术方案。The accompanying drawings, which are incorporated into and constitute a part of this specification, illustrate embodiments consistent with the present disclosure, and together with the description, serve to explain the technical solutions of the present disclosure.
图1是本公开实施例提供的一种图像目标分类方法的流程示意图;FIG. 1 is a schematic flowchart of an image target classification method provided by an embodiment of the present disclosure;
图2是可以应用本公开实施例的图像目标分类方法的一种系统架构示意图;FIG. 2 is a schematic diagram of a system architecture to which the image object classification method according to the embodiment of the present disclosure can be applied;
图3是本公开实施例提供的一种获取至少一张待分类图像的流程示意图;FIG. 3 is a schematic flowchart of obtaining at least one image to be classified according to an embodiment of the present disclosure;
图4是本公开实施例提供的一种对至少一张待分类图像进行目标分类的流程示意图;4 is a schematic flowchart of a target classification for at least one image to be classified according to an embodiment of the present disclosure;
图5是本公开实施例的图像目标分类方法中分类模型所使用的网络架构示意图;5 is a schematic diagram of a network architecture used by a classification model in the image target classification method according to an embodiment of the present disclosure;
图6是本公开实施例提供的一种图像目标分类装置60的框架示意图;FIG. 6 is a schematic diagram of a framework of an image object classification apparatus 60 provided by an embodiment of the present disclosure;
图7是本公开实施例提供的一种电子设备70的框架示意图;FIG. 7 is a schematic frame diagram of an electronic device 70 provided by an embodiment of the present disclosure;
图8是本公开实施例提供的一种计算机可读存储介质80的框架示意图。FIG. 8 is a schematic diagram of a framework of a computer-readable storage medium 80 provided by an embodiment of the present disclosure.
具体实施方式Detailed ways
下面结合说明书附图,对本公开实施例的方案进行详细说明。The solutions of the embodiments of the present disclosure will be described in detail below with reference to the accompanying drawings.
以下描述中,为了说明而不是为了限定,提出了诸如特定系统结构、接口、技术之类的具体细节,以便透彻理解本公开。In the following description, for purposes of explanation and not limitation, specific details are set forth, such as specific system structures, interfaces, techniques, etc., in order to provide a thorough understanding of the present disclosure.
本文中术语“和/或”,仅仅是一种描述关联对象的关联关系,表示可以存在三种关系,例如,A和/或B,可以表示:单独存在A,同时存在A和B,单独存在B这三种情况。另外,本文中字符“/”,一般表示前后关联对象是一种“或”的关系。此外,本文 中的“多”表示两个或者多于两个。另外,本文中术语“至少一种”表示多种中的任意一种或多种中的至少两种的任意组合,例如,包括A、B、C中的至少一种,可以表示包括从A、B和C构成的集合中选择的任意一个或多张元素。The term "and/or" in this article is only an association relationship to describe the associated objects, indicating that there can be three kinds of relationships, for example, A and/or B, it can mean that A exists alone, A and B exist at the same time, and A and B exist independently B these three cases. In addition, the character "/" in this document generally indicates that the related objects are an "or" relationship. Also, "multiple" as used herein means two or more than two. In addition, the term "at least one" herein refers to any combination of any one of the plurality or at least two of the plurality, for example, including at least one of A, B, and C, and may mean including from A, B, and C. Any one or more elements selected from the set of B and C.
相关技术中,基于CT和MR的三维成像技术在医学影像学诊断中起到至关重要的作用,是诊断例如肝脏疾病的主要影像检查方法之一。以肝脏肿瘤的诊断为例,CT检查的扫描序列主要包括平扫期、动态增强期、动脉期、门静脉期和延迟期。其中平扫期一般用来观察肝表面的变化,是否存在脂肪肝,肝纤维化,肝硬化等病变。动态增强的几个期像可显示病变的具体图像特征。以肝细胞癌(Hepatocellular Carcinoma HCC)为例,HCC主要发生在慢性肝病和肝硬化患者,可以从平扫期观察到相应的肝表面形态的变化,同时该肿瘤在平扫期一般变现为低密度或与肝实质等密度;在增强扫描之后,HCC在各期像上表现为:动脉期明显强化或不均匀强化,并伴有低密度包膜;门静脉期对比剂流出,同时显示增强的包膜;延迟期则呈现延迟增强的包膜。因此在一种可行的实施方式中,可通过识别目标肿瘤在多个期像所表现出的影像特征,判断它是否是HCC。相较于通过单一期像做出的判断准确率更高,因为富血供的小肝转移瘤,在平扫期和动脉期的图像特征和小HCC的特征类似,针对多期像的图像做分类任务,可以进一步提升图像分类的准确率。In the related art, three-dimensional imaging technology based on CT and MR plays a crucial role in medical imaging diagnosis, and is one of the main imaging examination methods for diagnosing, for example, liver diseases. Taking the diagnosis of liver tumors as an example, the scanning sequence of CT examination mainly includes the plain scan phase, the dynamic enhancement phase, the arterial phase, the portal venous phase and the delayed phase. The plain scan period is generally used to observe changes in the liver surface, whether there are fatty liver, liver fibrosis, liver cirrhosis and other diseases. Several phase images with dynamic enhancement can show the specific image features of the lesion. Taking Hepatocellular Carcinoma HCC as an example, HCC mainly occurs in patients with chronic liver disease and liver cirrhosis, and the corresponding changes in liver surface morphology can be observed from the plain scan period. Or the same density as the liver parenchyma; after enhanced scanning, HCC in each phase showed: marked enhancement or inhomogeneous enhancement in arterial phase, accompanied by low-density capsule; contrast agent outflow in portal venous phase, showing enhanced capsule at the same time ; the delayed phase presents a delayed enhanced envelope. Therefore, in a feasible implementation manner, it can be determined whether the target tumor is HCC by identifying the imaging features exhibited by the images in multiple stages. Compared with the single-phase image, the judgment accuracy is higher, because the image characteristics of the small liver metastases with rich blood supply in the plain and arterial phase are similar to those of small HCC. classification task, which can further improve the accuracy of image classification.
临床上对于肝肿瘤类型的诊断主要通过两种方式:一是影像科医生反复查看肿瘤在CT或MR多期像图像上的征象,进而在诊断报告里给出肿瘤的良恶性分型或者具体肿瘤分型,这个过程会花费医生一定的时间,反复对比序列间肿瘤的影像特征,可能会花费3至5分钟。二是采集肿瘤病灶标本进行病理诊断,样本处理复杂且消耗时长,可能会花2至3天的时间。为提高医生阅片效率,提供一种结合人工智能技术来实现肿瘤的辅助智能诊断。Clinically, there are two main ways to diagnose the type of liver tumor: First, the radiologist repeatedly checks the signs of the tumor on the CT or MR multi-phase images, and then gives the benign and malignant classification of the tumor or the specific tumor in the diagnosis report. Typing, a process that will take doctors a certain amount of time to repeatedly compare the imaging features of tumors between sequences, may take 3 to 5 minutes. The second is to collect tumor lesion samples for pathological diagnosis. The sample processing is complicated and time-consuming, and it may take 2 to 3 days. In order to improve the efficiency of doctors in reading images, an auxiliary intelligent diagnosis of tumors is provided that combines artificial intelligence technology.
其中,医学图像分析普遍存在标注数据少,任务复杂且困难等问题,同时为了更好地表征病变,需要分析序列之间的相关性。这些问题的存在一定程度上限制了深度学习网络的复杂性和深度,需要引入一些其他的策略来解决医学图像分析任务。以肝脏肿瘤分类问题为例,肿瘤本身的影像特征是判断其类型的主要依据,而目标肿瘤周边可能存在多种噪声,会误导深度学习网络学习到一些错误的特征;肝脏肿瘤大小不一,小则0.5cm以下,大则20cm以上,需要网络能够考虑到肝脏肿瘤的这个特征,在确保高精准的大肿瘤分类识别的基础上,提高小肿瘤的识别能力;局限于CT扫描图像的分辨率,肝肿瘤的影像特征不一定很明显。肝肿瘤分类任务中存在很多困难,需要引入一定的策略来学习更好的特征表示,从而实现同类样本聚集,异类样本远离。Among them, medical image analysis generally has problems such as less labeled data, complex and difficult tasks, and at the same time, in order to better characterize lesions, it is necessary to analyze the correlation between sequences. The existence of these problems limits the complexity and depth of deep learning networks to a certain extent, and some other strategies need to be introduced to solve the task of medical image analysis. Taking the problem of liver tumor classification as an example, the image features of the tumor itself are the main basis for judging its type, and there may be various noises around the target tumor, which will mislead the deep learning network to learn some wrong features; liver tumors vary in size and small. If it is less than 0.5cm and larger than 20cm, the network needs to be able to take into account the characteristics of liver tumors, and improve the identification ability of small tumors on the basis of ensuring high-precision classification and identification of large tumors; limited to the resolution of CT scan images, The imaging features of liver tumors are not necessarily obvious. There are many difficulties in the task of liver tumor classification, and it is necessary to introduce certain strategies to learn better feature representations, so as to achieve the aggregation of similar samples and the distance of heterogeneous samples.
基于上述研究,本公开至少提供一种图像目标分类方法,该方法利用分类模型对待分类图像进行目标分类,不仅使得目标分类过程更加简单,减小对医生的依赖,提高目标分类速度和准确性,而且结合人工智能技术实现目标分类,以便辅助医生进行智能化疾病诊疗。Based on the above research, the present disclosure provides at least one image target classification method, which uses a classification model to classify images to be classified, which not only makes the target classification process simpler, reduces the dependence on doctors, and improves the speed and accuracy of target classification, And combined with artificial intelligence technology to achieve target classification, in order to assist doctors in intelligent disease diagnosis and treatment.
请参阅图1,图1是本公开实施例提供的在一种图像目标分类方法的流程示意图。具体而言,可以包括如下步骤:Please refer to FIG. 1. FIG. 1 is a schematic flowchart of an image object classification method provided by an embodiment of the present disclosure. Specifically, the following steps can be included:
步骤S11:获取包含目标对象的至少一张待分类图像。Step S11: Acquire at least one image to be classified that includes the target object.
其中,至少一张待分类图像为属于至少一种扫描图像类别的医学图像。Wherein, at least one image to be classified is a medical image belonging to at least one scanned image category.
本公开实施例中,待分类图像可以为医学图像,包括但不限于CT图像、MR图像,在此不做限定。待分类图像可以均为CT图像,可以均为MR图像,还可以一部分为CT图像、一部分为MR图像,在此不作具体限定。在医学影像学诊断中,CT图像、MR图像是多期像或多序列成像,每个期像或序列显示出目标对象所在区域或其他区域的不同影像信息,多个期像或序列的特征进行有效地结合,能够更精准地明确病变性质。In this embodiment of the present disclosure, the images to be classified may be medical images, including but not limited to CT images and MR images, which are not limited herein. The images to be classified may all be CT images, may all be MR images, and may also be partly CT images and partly MR images, which are not specifically limited herein. In medical imaging diagnosis, CT images and MR images are multi-phase images or multi-sequence imaging. Each phase image or sequence shows different image information of the area where the target object is located or other areas. Combined effectively, the nature of the lesions can be more precisely defined.
待分类图像可以是对腹部、胸部等区域进行扫描得到的。例如,对腹部进行扫描得到的待分类图像可以包括肝脏、脾脏、肾脏等组织器官,对胸部进行扫描得到的待分类图像可以包括心脏、肺等组织器官,具体可以根据实际应用情况扫描得到待分类图像,在此不做限定。目标对象可以但不限于是肝脏肿瘤等需要利用本公开实施例的图像目标分类方法进行分类的对象。The images to be classified may be obtained by scanning the abdomen, chest and other regions. For example, the image to be classified obtained by scanning the abdomen may include tissues and organs such as liver, spleen, and kidney, and the image to be classified obtained by scanning the chest may include tissues and organs such as the heart and lung. Specifically, the images to be classified may be scanned according to the actual application. Images, not limited here. The target object may be, but is not limited to, a liver tumor and other objects that need to be classified using the image object classification method of the embodiment of the present disclosure.
至少一张待分类图像可以为属于至少一种扫描图像类别的医学图像。不同扫描图像类别的医学图像可用于显示目标对象不同的特征信息,因此可提高图像目标分类的精准度。在一些公开实施例中,扫描图像类别也可以称为上述描述的期像和/或序列。不同扫描图像类别的图像可以是与时序有关和/或与扫描参数有关的图像。例如,扫描图像类别可以包括与时序有关的造影前平扫、动脉早期、动脉晚期、门脉期以及延迟期等;或者,扫描图像类别还可以包括与扫描参数有关的T1加权反相成像、T1加权同相成像、T2加权成像、扩散加权成像以及表面扩散系数成像等。The at least one image to be classified may be a medical image belonging to at least one category of scanned images. Medical images of different scanned image categories can be used to display different characteristic information of target objects, thus improving the accuracy of image target classification. In some disclosed embodiments, the class of scanned images may also be referred to as the above-described images and/or sequences. The images of the different scanned image categories may be timing-dependent and/or scan-parameter-dependent images. For example, the scan image category may include time-series-related pre-contrast scan, early arterial phase, late arterial phase, portal venous phase, and delayed phase, etc.; alternatively, the scan image category may also include T1-weighted inverse imaging, T1 Weighted in-phase imaging, T2-weighted imaging, diffusion-weighted imaging, and surface diffusion coefficient imaging, etc.
以肝脏为例,动脉早期可以表示门静脉尚未增强,动脉晚期可以表示门静脉已被增强,门脉期可以表示门静脉已充分增强且肝脏血管已被前向性血流增强、肝脏软细胞组织在标记物下已达到峰值,延迟期可以表示门脉和动脉处于增强状态并弱于门脉期、且肝脏软细胞组织处于增强状态并弱于门脉期,其他扫描图像类别在此不再一一举例。在待分类图像为对其他脏器扫描得到的医学图像时,可以以此类推,在此不再一一举例。Taking the liver as an example, the early arterial stage can indicate that the portal vein has not been enhanced, the late arterial stage can indicate that the portal vein has been enhanced, and the portal venous phase can indicate that the portal vein has been sufficiently enhanced and the liver blood vessels have been enhanced by forward blood flow. The delay period can indicate that the portal vein and arteries are in an enhanced state and weaker than the portal venous phase, and the liver parenchyma is in an enhanced state and weaker than the portal venous phase, and other scanning image categories will not be listed one by one here. When the image to be classified is a medical image obtained by scanning other organs, it can be deduced by analogy, and examples will not be given here.
步骤S12:利用分类模型,对至少一张待分类图像进行目标分类,得到目标对象的类型。Step S12: Using the classification model, perform target classification on at least one image to be classified to obtain the type of the target object.
获取到包含目标对象的至少一张待分类图像后,利用分类模型对至少一张待分类图像进行目标分类,即可得到目标对象的类型。After acquiring at least one image to be classified including the target object, the classification model is used to classify the at least one image to be classified, so as to obtain the type of the target object.
在一公开实施例中,分类模型对至少一张待分类图像进行目标分类,得到目标对象属于不同类型的概率,将满足预设概率条件的类型作为目标对象的类型。预设概率条件包括但不限于概率值最大等。目标对象属于不同类型的概率可以是分类模型训练得到的。分类模型每次训练选择的批样本数据是利用数据生成器从样本数据集中选择的不同目标类型的数量为预设比例的样本数据。由于数据生成器随机选择包含等比例的不同目标类型的样本数据作为批样本数据,以免因某目标类型的样本数据出现太少而导致分类性能不均衡,因此,分类模型对至少一张待分类图像进行目标分类是通过大量批样本数据训练得到的,可以提高分类模型的分类性能。利用分类模型得到目标对象的类型,可辅助医生对目标对象的类型的确定,节省医生审阅待分类图像的时间,进而能够加快报告的输出。In a disclosed embodiment, the classification model performs target classification on at least one image to be classified, obtains probabilities that the target objects belong to different types, and uses the types that satisfy the preset probability conditions as the types of the target objects. The preset probability conditions include but are not limited to the maximum probability value and the like. The probability that the target objects belong to different types can be obtained by training the classification model. The batch sample data selected for each training of the classification model is sample data with a preset proportion of the number of different target types selected from the sample data set by the data generator. Since the data generator randomly selects sample data containing equal proportions of different target types as batch sample data, so as to avoid unbalanced classification performance due to too few sample data of a certain target type, the classification model is used for at least one image to be classified. The target classification is obtained by training a large number of batch sample data, which can improve the classification performance of the classification model. Using the classification model to obtain the type of the target object can assist the doctor in determining the type of the target object, save the doctor's time for reviewing the images to be classified, and thus can speed up the output of the report.
在一公开实施例中,对至少一张待分类图像进行目标分类,得到目标对象的类型时,对至少一张待分类图像进行若干层特征提取,对应得到若干组初始特征信息;基于若干组初始特征信息中的至少一组初始特征信息,得到最终特征信息;对最终特征信息进行分类,得到目标对象的类型。In a disclosed embodiment, target classification is performed on at least one image to be classified, and when the type of the target object is obtained, several layers of feature extraction are performed on at least one image to be classified, and several sets of initial feature information are correspondingly obtained; At least one set of initial feature information in the feature information is used to obtain final feature information; the final feature information is classified to obtain the type of the target object.
对至少一张待分类图像进行特征提取时,特征提取的层数可以为一层、两层甚至更多层。对至少一张待分类图像进行特征提取时,具体对哪些层进行特征提取可以通过人为设置获取得到、也可以是在训练分类模型时通过大量实验确定的,在此不作具体限定。对至少一张待分类图像进行一层特征提取,则对应得到一组初始特征信息。对至少一张待分类图像进行多层特征提取,则对应得到多组初始特征信息,其中,多层特征提取可以是连续的,也可以是间断的。初始特征信息可以为目标对象的特征图,反映目标对象在待分类图像中的特征信息。在一公开实施例中,分类模型为深度学习网络,该深度学习网络可包括编码器(encoder)或其变种、Resnet或者其变种,可以是神经网络(Visual Geometry Group Network,VGG)16或者其变种,也可以是其他的用于分类的网络模型 结构。分类模型通过卷积层对至少一张待分类图像进行特征提取,不同卷积层对应不同层特征提取,得到不同组初始特征信息。When performing feature extraction on at least one image to be classified, the number of layers for feature extraction may be one layer, two layers or even more layers. When performing feature extraction on at least one image to be classified, which layers to perform feature extraction on can be obtained through artificial settings, or can be determined through a large number of experiments when training a classification model, which is not specifically limited here. A layer of feature extraction is performed on at least one image to be classified, and a set of initial feature information is correspondingly obtained. Multi-layer feature extraction is performed on at least one image to be classified, and multiple sets of initial feature information are correspondingly obtained, wherein the multi-layer feature extraction may be continuous or discontinuous. The initial feature information may be a feature map of the target object, reflecting the feature information of the target object in the image to be classified. In a disclosed embodiment, the classification model is a deep learning network, and the deep learning network may include an encoder (encoder) or its variants, Resnet or its variants, and may be a neural network (Visual Geometry Group Network, VGG) 16 or its variants. , or other network model structures for classification. The classification model performs feature extraction on at least one image to be classified through a convolution layer, and different convolution layers correspond to different layers of feature extraction to obtain different groups of initial feature information.
上述方案,获取包含目标对象的至少一张待分类图像后,利用分类模型对至少一张待分类图像进行目标分类,得到目标对象的类型,因此提出基于人工智能技术的图像目标分类方法,实现了智能化目标分类,且无需人工进行目标分类,可减小对人工依赖,提高目标分类效率。In the above scheme, after obtaining at least one image to be classified including the target object, the classification model is used to classify the at least one image to be classified, and the type of the target object is obtained. Therefore, an image target classification method based on artificial intelligence technology is proposed, which realizes the Intelligent target classification, and no manual target classification is required, which can reduce the dependence on manual work and improve the efficiency of target classification.
在一应用实施例中,为实现肝脏肿瘤的分类,获取包括肝脏肿瘤的至少一张待分类图像,利用分类模型对至少一张待分类图像进行目标分类,得到肝脏肿瘤的类型,无需人工对待分类图像进行分类,利用分类模型即可实现肝脏肿瘤的分类,以便医生获取到肝脏肿瘤的类型。In an application embodiment, in order to realize the classification of liver tumors, at least one image to be classified including liver tumors is acquired, and a classification model is used to perform target classification on at least one image to be classified, so as to obtain the type of liver tumor, and no manual classification is required. The images are classified, and the classification model can be used to realize the classification of liver tumors, so that the doctor can obtain the type of liver tumors.
图2为可以应用本公开实施例的图像目标分类方法的一种系统架构示意图;如图2所示,该系统架构中包括:图像获取终端201、网络202和目标分类终端203。为实现支撑一个示例性应用,图像获取终端201和目标分类终端203通过网络202建立通信连接,图像获取终端201通过网络202向目标分类终端203上报包含目标对象的至少一张待分类图像,目标分类终端203响应于接收到的至少一张待分类图像,并利用分类模型,对至少一张待分类图像进行目标分类,得到目标对象的类型。最后,目标分类终端203将目标对象的类型上传至网络202,并通过网络202发送给图像获取终端201。FIG. 2 is a schematic diagram of a system architecture to which the image object classification method according to an embodiment of the present disclosure can be applied; as shown in FIG. 2 , the system architecture includes an image acquisition terminal 201 , a network 202 and an object classification terminal 203 . In order to support an exemplary application, the image acquisition terminal 201 and the target classification terminal 203 establish a communication connection through the network 202, and the image acquisition terminal 201 reports at least one image to be classified containing the target object to the target classification terminal 203 through the network 202, and the target classification The terminal 203 responds to the received at least one image to be classified, and uses the classification model to perform target classification on the at least one image to be classified to obtain the type of the target object. Finally, the target classification terminal 203 uploads the type of the target object to the network 202 and sends it to the image acquisition terminal 201 through the network 202 .
作为示例,图像获取终端201可以包括图像采集设备,目标分类终端203可以包括具有视觉信息处理能力的视觉处理设备或远程服务器。网络202可以采用有线或无线连接方式。其中,当目标分类终端203为视觉处理设备时,图像获取终端201可以通过有线连接的方式与视觉处理设备通信连接,例如通过总线进行数据通信;当目标分类终端203为远程服务器时,图像获取终端201可以通过无线网络与远程服务器进行数据交互。As an example, the image acquisition terminal 201 may include an image acquisition device, and the target classification terminal 203 may include a vision processing device or a remote server with visual information processing capability. Network 202 may employ wired or wireless connections. Wherein, when the target classification terminal 203 is a visual processing device, the image acquisition terminal 201 can be connected to the visual processing device through a wired connection, such as data communication through a bus; when the target classification terminal 203 is a remote server, the image acquisition terminal 201 can perform data interaction with a remote server through a wireless network.
或者,在一些场景中,图像获取终端201可以是带有图像采集模组的视觉处理设备,具体实现为带有摄像头的主机。这时,本公开实施例的图像目标分类方法可以由图像获取终端201执行,上述系统架构可以不包含网络202和目标分类终端203。Alternatively, in some scenarios, the image acquisition terminal 201 may be a vision processing device with an image acquisition module, which is specifically implemented as a host with a camera. At this time, the image object classification method according to the embodiment of the present disclosure may be executed by the image acquisition terminal 201 , and the above-mentioned system architecture may not include the network 202 and the object classification terminal 203 .
为了使至少一张待分类图像更加统一,在从原始医学图像中提取得到待分类图像之前,可以对原始医学图像进行图像预处理,进而分别从多张原始医学图像提取得到包含目标对象的待分类图像,以获取包含目标对象的至少一张待分类图像。请参阅图3,图3是本公开实施例提供的一种获取至少一张待分类图像的流程示意图。具体而言,可以包括如下步骤:In order to make the at least one image to be classified more uniform, before the image to be classified is extracted from the original medical image, image preprocessing can be performed on the original medical image, and then the to-be-classified image containing the target object can be extracted from multiple original medical images respectively. image to obtain at least one image to be classified that contains the target object. Please refer to FIG. 3 , which is a schematic flowchart of acquiring at least one image to be classified according to an embodiment of the present disclosure. Specifically, the following steps can be included:
步骤S111:将原始医学图像重采样至预设分辨率。Step S111: Resampling the original medical image to a preset resolution.
预设分辨率的大小可自定义设置,可根据不同目标对象设置与目标对象对应的预设分辨率,从而将原始医学图像的分辨率统一至图像效果最佳的分辨率。The size of the preset resolution can be customized, and the preset resolution corresponding to the target object can be set according to different target objects, so as to unify the resolution of the original medical image to the resolution with the best image effect.
步骤S112:调整原始医学图像中的像素值范围。Step S112: Adjust the pixel value range in the original medical image.
通过调整原始医学图像的像素值范围,使原始医学图像的亮度和颜色更容易显示目标对象。原始医学图像的类别包括但不限于包括但不限于CT图像、MR图像等能够反映目标对象特征信息的图像,在此不做限定。若原始医学图像为CT图像,则可统一原始医学图像至预设窗宽窗位;若原始医学图像为MR图像,由于MR图像像素分布的动态范围变化较大,在本公开的一个实施场景中,可以采用灰度累积分布函数下的预设比例(例如,99.9%)对应的灰度值作为归一化的预处理钳位值,从而能够加强MR图像数据的对比度,有利于提升后续图像目标分类的准确性。By adjusting the pixel value range of the original medical image, the brightness and color of the original medical image are made easier to display the target object. The categories of the original medical images include, but are not limited to, CT images, MR images, and other images that can reflect the feature information of the target object, which are not limited herein. If the original medical image is a CT image, the original medical image can be unified to a preset window width and window level; if the original medical image is an MR image, since the dynamic range of the pixel distribution of the MR image changes greatly, in an implementation scenario of the present disclosure , the gray value corresponding to a preset ratio (for example, 99.9%) under the grayscale cumulative distribution function can be used as the normalized preprocessing clamp value, so that the contrast of the MR image data can be enhanced and the subsequent image target can be improved. Classification accuracy.
步骤S113:将原始医学图像进行归一化处理。Step S113: Normalize the original medical image.
在一公开实施例中,可以对原始医学图像进行归一化处理。归一化处理包括但不限于将原始医学图像的强度或者像素值归一化到预设范围(例如,0至1的范围)。In a disclosed embodiment, the raw medical images may be normalized. The normalization process includes, but is not limited to, normalizing the intensity or pixel values of the original medical image to a preset range (eg, a range of 0 to 1).
步骤S114:检测到第一原始医学图像未标注有目标对象的初始区域,利用第二原始医学图像上标注的目标对象的初始区域以及第二原始医学图像与第一原始医学图像的配准关系,确定第一原始医学图像上目标对象的初始区域。Step S114: Detecting the initial area of the first original medical image that is not marked with the target object, using the initial area of the target object marked on the second original medical image and the registration relationship between the second original medical image and the first original medical image, An initial region of the target object on the first original medical image is determined.
在本公开的一些实施例中,可能并非所有原始医学图像均标注有目标对象的初始区域,因此,为了利用更多包含目标对象的待分类图像进行图像目标分类,提高图像目标分类的准确性,可补齐原始医学图像的初始区域。在检测到第一原始医学图像未标注有目标对象的初始区域,利用第二原始医学图像上标注的目标对象的初始区域以及第二原始医学图像与第一原始医学图像的配准关系,确定第一原始医学图像上目标对象的初始区域。在一公开实施例中,为了提升确定目标对象的初始区域的便利性,可以利用配准网络进行上述确定第一原始医学图像上目标对象的初始区域的步骤。In some embodiments of the present disclosure, not all original medical images may be marked with the initial area of the target object. Therefore, in order to use more images to be classified including the target object to perform image object classification and improve the accuracy of image object classification, The initial area of the original medical image can be filled. After detecting the initial area of the first original medical image that is not marked with the target object, use the initial area of the target object marked on the second original medical image and the registration relationship between the second original medical image and the first original medical image to determine the first An initial region of the target object on the original medical image. In a disclosed embodiment, in order to improve the convenience of determining the initial area of the target object, the above step of determining the initial area of the target object on the first original medical image may be performed by using a registration network.
在分别从多张原始医学图像提取得到包含目标对象的待分类图像之前,图像目标分类方法可包括上述步骤S111至步骤S114的若干个步骤,上述步骤S111至步骤S114仅是示例性说明,在一公开实施例中可根据需要选取若干个步骤对原始医学图像进行预处理,也即上述步骤S111至步骤S114的个数可任意选择,在此不作具体限定。通过统一分辨率、调整像素值范围、归一化处理、以及确定目标对象的初始区域等操作,可在从原始医学图像提取待分类图像之前,对原始医学图像进行预处理,统一待分类图像的图像参数,提高待分类图像的质量。Before the images to be classified containing the target object are extracted from multiple original medical images respectively, the image target classification method may include several steps from the above steps S111 to S114. The above steps S111 to S114 are only exemplary descriptions. In the disclosed embodiment, several steps can be selected to preprocess the original medical image as required, that is, the number of the above steps S111 to S114 can be arbitrarily selected, which is not specifically limited herein. By unifying the resolution, adjusting the pixel value range, normalizing, and determining the initial area of the target object, the original medical image can be preprocessed before the image to be classified is extracted from the original medical image, and the Image parameters to improve the quality of images to be classified.
在对原始医学图像进行预处理后,即可分别从多张原始医学图像提取得到包含目标对象的待分类图像,具体描述参阅后文步骤S115和步骤S116。After the original medical images are preprocessed, the images to be classified including the target object can be extracted from the multiple original medical images respectively. For details, refer to steps S115 and S116 below.
步骤S115:确定原始医学图像中目标对象的初始区域,按照预设比例扩大初始区域,得到待提取区域。Step S115: Determine the initial area of the target object in the original medical image, and expand the initial area according to a preset ratio to obtain the area to be extracted.
目标对象本身特征是判断其类型的主要依据,而目标对象周边可能存在多种噪声干扰,该噪声干扰会误导目标对象的分类。以目标对象为肝脏肿瘤为例,慢性肝病或者肝硬化背景、其他类型肿瘤、与肝脏肿瘤位置相近的血管等噪声干扰均会影响目标对象的分类精度,因此,确定原始医学图像中目标对象的初始区域,以作为待提取区域,使得待提取区域包含目标对象。在一公开实施例中,为将目标对象周围的背景信息作为目标分类的辅助信息,或者避免初始区域的确定误差,以提高待分类图像的获取精度,在确定原始医学图像中目标对象的初始区域后,可按照预设比例扩大初始区域,得到待提取区域。初始区域用于圈定目标对象在原始医学图像的位置。在一公开实施例中,可利用图像分割技术确定原始医学图像中目标对象的边界轮廓,标记边界轮廓形成初始区域。The characteristics of the target object itself are the main basis for judging its type, and there may be a variety of noise interference around the target object, which will mislead the classification of the target object. Taking the target object as a liver tumor as an example, the background of chronic liver disease or cirrhosis, other types of tumors, and blood vessels close to the liver tumor will affect the classification accuracy of the target object. The area to be extracted is used as the area to be extracted, so that the area to be extracted contains the target object. In a disclosed embodiment, in order to use the background information around the target object as auxiliary information for target classification, or to avoid the determination error of the initial area, so as to improve the acquisition accuracy of the image to be classified, the initial area of the target object in the original medical image is determined. Afterwards, the initial area may be expanded according to a preset ratio to obtain the area to be extracted. The initial region is used to delineate the position of the target object in the original medical image. In a disclosed embodiment, an image segmentation technique can be used to determine the boundary contour of the target object in the original medical image, and mark the boundary contour to form an initial area.
步骤S116:从原始医学图像中提取待提取区域中的图像数据,得到待分类图像。Step S116: Extract the image data in the area to be extracted from the original medical image to obtain the image to be classified.
利用待提取区域从原始医学图像中提取图像数据,得到的待分类图像则包括目标对象。The image data is extracted from the original medical image by using the area to be extracted, and the obtained image to be classified includes the target object.
原始医学图像可以为二维图像或者三维图像。在原始医学图像为二维图像的情况下,待分类图像为二维图像。在原始医学图像为三维图像的情况下,待分类图像可以为三维图像,又或者待分类图像可以为二维图像。在本公开的一些实施例中,由于三维图像由若干层二维图像组成,如在确定二维的待分类图像时,可以但不限于将目标对象面积最大所在层的二维图像作为待分类图像;将目标对象直径最大所在层的二维图像作为待分类图像;或者将所有二维图像中的中间层作为待分类图像;或者将所有二维图像中的任意一层作为待分类图像,在此不作具体限定。在一应用实施例中,原始医学图像为三维图像,而待分类图像为对原始医学图像中目标对象最大面积所在层提取得到的二维图像,从而可以提取原始医学图像中目标对象最大面积所在层作为待分类图像,使得待分类图像中目标对象的提取范围较大,包含目标对象的信息更多,进而可以提高目标对象的分类精度。通过上述方式,在对原始医学图像进行预处理后,确定原始医学图像中目标对象的初始区域,按照预设比例扩大初始区域,得到待提取区域;然后从原始医学图像中 提取待提取区域中的图像数据,得到待分类图像。初始区域是包含目标对象的区域,而按照预设比例扩大目标对象的初始区域,使得得到的待提取区域既包含目标对象,又包含目标对象周围的部分背景信息,以便将待提取区域中的图像数据提取作为待分类图像后,待分类图像能够囊括目标对象和部分背景信息。The original medical image can be a two-dimensional image or a three-dimensional image. In the case where the original medical image is a two-dimensional image, the image to be classified is a two-dimensional image. When the original medical image is a three-dimensional image, the image to be classified may be a three-dimensional image, or the image to be classified may be a two-dimensional image. In some embodiments of the present disclosure, since a three-dimensional image is composed of several layers of two-dimensional images, for example, when determining a two-dimensional image to be classified, the two-dimensional image of the layer where the target object has the largest area may be, but not limited to, the image to be classified. ; take the two-dimensional image of the layer where the target object diameter is the largest as the image to be classified; or take the middle layer in all two-dimensional images as the image to be classified; or take any layer in all the two-dimensional images as the image to be classified, here There is no specific limitation. In an application embodiment, the original medical image is a three-dimensional image, and the image to be classified is a two-dimensional image obtained by extracting the layer where the maximum area of the target object is located in the original medical image, so that the layer where the maximum area of the target object is located in the original medical image can be extracted. As an image to be classified, the extraction range of the target object in the to-be-classified image is larger and contains more information of the target object, thereby improving the classification accuracy of the target object. In the above manner, after the original medical image is preprocessed, the initial area of the target object in the original medical image is determined, and the initial area is expanded according to a preset ratio to obtain the area to be extracted; then the area to be extracted is extracted from the original medical image. image data to obtain an image to be classified. The initial area is the area containing the target object, and the initial area of the target object is expanded according to a preset ratio, so that the obtained area to be extracted contains both the target object and some background information around the target object, so that the image in the area to be extracted can be extracted. After the data is extracted as an image to be classified, the image to be classified can include the target object and some background information.
另外,分别从多张原始医学图像提取得到包含目标对象的待分类图像,实现待分类图像的获取,而且待分类图像可从原始医学图像中提取得到,相比直接采用原始医学图像,减少后续分类的图像尺寸,而且可一定程度上避免原始医学图像中的一些背景噪声,故可减少后续分类的处理资源损耗,且提高分类性能。In addition, the to-be-classified images containing the target object are extracted from multiple original medical images respectively, so as to realize the acquisition of the to-be-classified images, and the to-be-classified images can be extracted from the original medical images. Compared with directly using the original medical images, the subsequent classification is reduced. It can avoid some background noise in the original medical image to a certain extent, so it can reduce the processing resource consumption of subsequent classification and improve the classification performance.
本公开实施例中,提出利用人工智能技术的分类模型对至少一张待分类图像进行目标分类,可以大大提高确定目标对象的类型的效率。请参阅图4,图4是本公开实施例提供的一种对至少一张待分类图像进行目标分类的流程示意图。具体而言,可以包括如下步骤:In the embodiments of the present disclosure, it is proposed to use a classification model of artificial intelligence technology to classify at least one image to be classified, which can greatly improve the efficiency of determining the type of the target object. Please refer to FIG. 4 . FIG. 4 is a schematic flowchart of a target classification for at least one image to be classified according to an embodiment of the present disclosure. Specifically, the following steps can be included:
步骤S121:对至少一张待分类图像进行若干层特征提取,对应得到若干组初始特征信息。Step S121 : extracting several layers of features on at least one image to be classified, and correspondingly obtaining several sets of initial feature information.
其中,每组初始特征信息的尺寸不同。Among them, the size of each group of initial feature information is different.
对至少一张待分类图像进行特征提取时,特征提取的层数可以为一层、两层甚至更多层。特征提取可以由卷积层实现,每个卷积层分别对至少一张待分类图像进行特征提取,得初始特征信息。对至少一张待分类图像进行特征提取时,具体对哪些层进行特征提取可以通过人为设置获取得到、也可以为训练分类模型时通过大量实验确定的,在此不作具体限定。对至少一张待分类图像进行一层特征提取,则对应得到一组初始特征信息,其中,该一层特征提取可以是任意一层,例如但不限于将最后一层特征提取得到的初始特征信息作为后续目标分类的依据。对至少一张待分类图像进行多层特征提取,则对应得到多组初始特征信息,其中,多层特征提取可以是连续的,也可以是间断的。初始特征信息可以为目标对象的特征图,反映目标对象在待分类图像中的特征信息。每组初始特征信息的尺寸不同,其中,尺寸包括维度和/或分辨率,从而多组初始特征信息分别反映目标对象不同的特征信息。When performing feature extraction on at least one image to be classified, the number of layers for feature extraction may be one layer, two layers or even more layers. Feature extraction can be implemented by convolutional layers, and each convolutional layer performs feature extraction on at least one image to be classified to obtain initial feature information. When performing feature extraction on at least one image to be classified, which layers to perform feature extraction on can be obtained through manual settings, or can be determined through a large number of experiments when training a classification model, which is not specifically limited here. A layer of feature extraction is performed on at least one image to be classified, and a set of initial feature information is correspondingly obtained, wherein the layer of feature extraction can be any layer, such as but not limited to the initial feature information obtained by extracting the last layer of features. as the basis for subsequent target classification. Multi-layer feature extraction is performed on at least one image to be classified, and multiple sets of initial feature information are correspondingly obtained, wherein the multi-layer feature extraction may be continuous or discontinuous. The initial feature information may be a feature map of the target object, reflecting the feature information of the target object in the image to be classified. The size of each set of initial feature information is different, wherein the size includes dimension and/or resolution, so that the multiple sets of initial feature information respectively reflect different feature information of the target object.
在一公开实施例中,分类模型为深度学习网络,包括的网络模型结构可以是encoder或其变种、Resnet或者其变种,可以是VGG16或者其变种,也可以是其他的用于分类的网络模型结构。分类模型通过卷积层对至少一张待分类图像进行特征提取,不同卷积层对应不同层特征提取,得到不同组初始特征信息。In a disclosed embodiment, the classification model is a deep learning network, and the included network model structure can be an encoder or its variant, Resnet or its variant, VGG16 or its variant, or other network model structures for classification. . The classification model performs feature extraction on at least one image to be classified through a convolution layer, and different convolution layers correspond to different layers of feature extraction to obtain different groups of initial feature information.
待分类图像中,目标对象周围可能存在噪声干扰,以目标对象为肝脏肿瘤为例,慢性肝病或者肝硬化背景、其他类型肿瘤、与肝脏肿瘤位置相近的血管等噪声均会影响目标对象的分类精度,因此,在利用分类模型对至少一张待分类图像进行目标分类,得到目标对象的类型之前,可基于待分类图像中目标对象对应的初始区域,得到目标对象的最终区域。确定目标对象的最终区域时,可将某一初始区域作为目标对象的最终区域,或者综合至少一张待分类图像中目标对象对应的初始区域得到目标对象的最终区域,具体如,将至少一张待分类图像中目标对象对应的初始区域的并集,作为目标对象的最终区域,在此不作限定。为了使得分类模型能够学习到目标对象本身的一些重要特征,且在一定程度上减少周边噪声对目标对象的分类影响,可在提取待分类图像的初始特征信息(例如该待分类图像的全局特征等)时,加上目标对象的最终区域的监督,例如:在特征提取过程中,待分类图像中对应最终区域的权重高于待分类图像中其他区域的权重,由此使得让分类模型趋向于对最终区域提取细节更丰富的特征,进而使得分类模型输出的初始特征信息中对应最终区域能够尽量的特征更丰富;和/或,初始特征信息中对应最终区域的特征比其他区域的特征更丰富。在对待分类图像进行特征提取得到初始特征信息时,不仅提取待分类图像的全局特征,而且由于加入最终区域的监督机制,引导分类 模型更关注最终区域中的目标对象,以便分类模型学习到目标对象本身的特征信息,减小目标对象周围噪声干扰对目标分类的影响。In the image to be classified, there may be noise interference around the target object. Taking the target object as a liver tumor as an example, the background of chronic liver disease or cirrhosis, other types of tumors, and blood vessels close to the liver tumor will affect the classification accuracy of the target object. Therefore, before using the classification model to perform target classification on at least one image to be classified to obtain the type of the target object, the final area of the target object can be obtained based on the initial area corresponding to the target object in the image to be classified. When determining the final area of the target object, an initial area can be used as the final area of the target object, or the final area of the target object can be obtained by combining the initial areas corresponding to the target object in at least one image to be classified. The union of the initial regions corresponding to the target object in the image to be classified is regarded as the final region of the target object, which is not limited here. In order to enable the classification model to learn some important features of the target object itself, and to reduce the influence of surrounding noise on the classification of the target object to a certain extent, the initial feature information of the image to be classified (such as the global features of the image to be classified, etc.) can be extracted. ), plus the supervision of the final area of the target object, for example: in the feature extraction process, the weight of the corresponding final area in the image to be classified is higher than the weight of other areas in the image to be classified, which makes the classification model tend to The final region extracts features with richer details, so that the initial feature information output by the classification model corresponding to the final region can be as rich in features as possible; and/or, the features corresponding to the final region in the initial feature information are more abundant than the features of other regions. When the initial feature information is obtained from the feature extraction of the image to be classified, not only the global features of the image to be classified are extracted, but also due to the addition of the supervision mechanism of the final area, the classification model is guided to pay more attention to the target object in the final area, so that the classification model can learn the target object. The feature information of itself can reduce the influence of noise interference around the target object on the target classification.
在一公开实施例中,在基于待分类图像中目标对象对应的初始区域,得到目标对象的最终区域时,获取至少一张待分类图像中目标对象对应的初始区域的并集,以作为目标对象的最终区域,从而可利用最终区域对至少一张待分类图像进行若干层特征提取,对应得到若干组初始特征信息。由于目标对象的最终区域是待分类图像中目标对象的初始区域的并集,使得最终区域大于或等于任意一个初始区域,保证目标对象的最终区域能够包含不同待分类图像中的目标对象对应区域,从而在对待分类图像进行特征提取时,能够尽可能关注目标对象特征信息。在一公开实施例中,至少一张待分类图像包括未标注目标对象的初始区域的第一待分类图像和标注目标对象的初始区域的第二待分类图像;在基于待分类图像中目标对象对应的初始区域,得到目标对象的最终区域之前,还可以利用分类模型检测到第一待分类图像未标注有目标对象的初始区域,并基于第二待分类图像上标注的目标对象的初始区域以及第二待分类图像与第一待分类图像的配准关系,确定第一待分类图像上目标对象的初始区域。因此,可以利用分类模型为未标注目标对象初始区域的第一待分类图像确定目标对象的初始区域,从而补齐标注,使得待分类图像中均包括初始区域。In a disclosed embodiment, when the final region of the target object is obtained based on the initial region corresponding to the target object in the image to be classified, a union of the initial regions corresponding to the target object in at least one image to be classified is obtained as the target object. Therefore, several layers of feature extraction can be performed on at least one image to be classified by using the final region, and several sets of initial feature information can be obtained correspondingly. Since the final area of the target object is the union of the initial areas of the target object in the image to be classified, the final area is greater than or equal to any initial area, ensuring that the final area of the target object can contain the corresponding areas of the target objects in different images to be classified, Therefore, when the feature extraction of the image to be classified is performed, the feature information of the target object can be paid attention to as much as possible. In a disclosed embodiment, the at least one image to be classified includes a first image to be classified without an initial area of the target object and a second image to be classified with an initial area of the target object marked; Before obtaining the final area of the target object, the classification model can also be used to detect the initial area of the first image to be classified that is not marked with the target object, and based on the initial area of the target object marked on the second image to be classified and the first image to be classified 2. The registration relationship between the image to be classified and the first image to be classified determines the initial area of the target object on the first image to be classified. Therefore, the classification model can be used to determine the initial area of the target object for the first to-be-classified image that is not labeled with the initial area of the target object, so as to complete the labeling, so that the to-be-classified images all include the initial area.
在一公开实施例中,可生成包括目标对象的最终区域的最终区域图,并将最终区域图与待分类图像一起输入分类模型,从而在利用分类模型对至少一张待分类图像进行目标分类,得到目标对象的类型时,利用最终区域图中包括的目标对象的最终区域对至少一张待分类图像进行若干层特征提取,能引导网络更关注最终区域的特征的学习,从一定程度上避免网络学习到很多错误的特征信息,减小目标对象周围噪声对特征提取的干扰。可以理解的是,在将最终区域图与待分类图像输入至分类模型之前,可将最终区域图像与待分类图像的尺寸调整为统一尺寸。In a disclosed embodiment, a final area map including the final area of the target object may be generated, and the final area map and the image to be classified may be input into the classification model, so that the target classification is performed on at least one image to be classified by using the classification model, When the type of the target object is obtained, using the final region of the target object included in the final region map to perform several layers of feature extraction on at least one image to be classified can guide the network to pay more attention to the learning of the features of the final region and avoid the network to a certain extent. A lot of wrong feature information is learned, and the interference of noise around the target object on feature extraction is reduced. It can be understood that, before inputting the final area map and the image to be classified into the classification model, the sizes of the final area image and the image to be classified may be adjusted to a uniform size.
步骤S122:基于若干组初始特征信息中的至少一组初始特征信息,得到最终特征信息。Step S122: Obtain final feature information based on at least one set of initial feature information in several sets of initial feature information.
对至少一张待分类图像进行若干层特征提取,对应得到若干组初始特征信息后,可以选择基于若干组初始特征信息中的至少一组初始特征信息,得到最终特征信息,且所选择的初始特征信息不同,得到的最终特征信息不同。初始特征信息的组数及其对应于分类模型的卷积层等参数信息可以是人工设置的,也可以是分类模型在训练过程中确定的,在此不作限定。融合多组初始特征信息可以提高分类模型的性能和目标分类的精度,但融合过多初始特征信息会引起过拟合问题,因此,合理调整进行融合的初始特征信息的组数,既能提高分类性能又能降低过拟合。由于每组初始特征信息在维度和分辨率等尺寸信息不同,分别反映目标对象的不同特征信息,可以将至少一组初始特征信息进行融合,得到最终特征信息,相较于现有的将最后一层高维特征图作为最终特征信息时,经过多个卷积后,一些重要的特征信息可能被压缩,尤其遗漏面积较小、图像特征模糊的目标对象,本公开实施例通过融合至少一组初始特征信息,可将不同特征提取阶段得到的初始特征信息拼接在一起,提高图像目标分类的精准度。Several layers of feature extraction are performed on at least one image to be classified, and after several sets of initial feature information are obtained, the final feature information can be obtained based on at least one set of initial feature information in several sets of initial feature information, and the selected initial feature information The information is different, and the final feature information obtained is different. The number of groups of initial feature information and the parameter information such as the convolution layer corresponding to the classification model may be manually set, or may be determined during the training process of the classification model, which is not limited here. Fusion of multiple sets of initial feature information can improve the performance of the classification model and the accuracy of target classification, but the fusion of too much initial feature information will cause overfitting problems. Therefore, rationally adjusting the number of groups of initial feature information to be fused can not only improve the classification performance can also reduce overfitting. Since each set of initial feature information is different in dimension and resolution, and reflects different feature information of the target object, at least one set of initial feature information can be fused to obtain the final feature information. When the layer high-dimensional feature map is used as the final feature information, after multiple convolutions, some important feature information may be compressed, especially the target object with small area and blurred image features is missed. Feature information, the initial feature information obtained in different feature extraction stages can be spliced together to improve the accuracy of image target classification.
在一公开实施例中,利用至少一组初始特征信息的权重,将至少一组初始特征信息进行融合,得到最终特征信息。每组初始特征信息的权重可以是人工设置的,也可以是在分类模型训练过程确定的,在此不作限定。例如,先初始化每组初始特征信息的权重,并在分类模型训练过程中不断更新该权重,具体如根据训练分类模型的训练结果与真实结果的比较结果,更新每组初始特征信息的权重,并利用训练分类模型不断重复上述更新权重的步骤,使得训练分类模型不断学习和更新每组初始特征信息的权重,得到训练好的分类模型及每组初始特征信息的权重。可以理解的,初始化的每组初始特征信息的权重可以相同或不同,且每组初始特征信息的权重之和为1。通过分类模型的迭代训练, 来确定用于融合的初始特征信息的权重,以使得利用该权重融合得到的最终特征信息更能反映目标对象特征,进一步提高分类性能。不同组初始特征信息的权重可以是相同或不同的,且每组初始特征信息的权重之和为1。由于利用多组初始特征信息得到最终特征信息时,可以利用初始特征信息的权重,将至少一层特征提取的不同尺寸的初始特征信息进行融合,得到最终特征信息,考虑较小尺寸的初始特征信息可能被压缩掉重要特征,通过综合不同尺寸的特征信息,能够得到较为综合和有用的最终特征信息,进而提高后续分类性能。在一公开实施例中,可以利用特征融合网络基于若干组初始特征信息中的至少一组初始特征信息,融合得到最终特征信息,可以把多个尺寸的初始特征信息拼接在一起作为分类任务的最终特征信息,同时给予每个初始特征信息一个权重,该权重经初始化后在模型训练过程中不断更新得到的,从而综合多个初始特征信息,得到更好的目标对象特征表示,进而提高目标分类的性能。In a disclosed embodiment, the weight of at least one set of initial feature information is used to fuse at least one set of initial feature information to obtain final feature information. The weight of each set of initial feature information may be manually set, or may be determined during the training process of the classification model, which is not limited here. For example, first initialize the weight of each group of initial feature information, and continuously update the weight during the training process of the classification model. The above steps of updating weights are continuously repeated by using the training classification model, so that the training classification model continuously learns and updates the weight of each group of initial feature information, and obtains the trained classification model and the weight of each group of initial feature information. It can be understood that the weights of each initial set of initial feature information may be the same or different, and the sum of the weights of each set of initial feature information is 1. Through the iterative training of the classification model, the weight of the initial feature information for fusion is determined, so that the final feature information obtained by using the weight fusion can better reflect the characteristics of the target object and further improve the classification performance. The weights of different groups of initial feature information may be the same or different, and the sum of the weights of each group of initial feature information is 1. When the final feature information is obtained by using multiple sets of initial feature information, the weight of the initial feature information can be used to fuse the initial feature information of different sizes extracted from at least one layer of features to obtain the final feature information, considering the initial feature information of smaller size. Important features may be compressed, and by synthesizing feature information of different sizes, more comprehensive and useful final feature information can be obtained, thereby improving the subsequent classification performance. In a disclosed embodiment, a feature fusion network can be used to obtain final feature information based on at least one set of initial feature information in several sets of initial feature information, and initial feature information of multiple sizes can be spliced together as the final feature of the classification task. At the same time, each initial feature information is given a weight, and the weight is continuously updated during the model training process after initialization, so as to integrate multiple initial feature information to obtain a better feature representation of the target object, thereby improving the accuracy of target classification. performance.
在一公开实施例中,在基于若干组初始特征信息中的至少一组初始特征信息,得到最终特征信息之前,可将每组初始特征信息转换为预设维度,方便后续最终特征信息的获取。例如一应用场景中,利用特征提取网络将每组初始特征信息转换为预设维度。预设维度可以根据需要设置,例如但不限于预设维度为一维。In a disclosed embodiment, before obtaining final feature information based on at least one set of initial feature information among several sets of initial feature information, each set of initial feature information may be converted into a preset dimension to facilitate subsequent acquisition of final feature information. For example, in an application scenario, a feature extraction network is used to convert each set of initial feature information into a preset dimension. The preset dimension can be set as required, for example, but not limited to, the preset dimension is one dimension.
步骤S123:对最终特征信息进行分类,得到目标对象的类型。Step S123: Classify the final feature information to obtain the type of the target object.
最终特征信息携带目标对象的特征,从而对最终特征信息进行分类,即可得到目标对象的类型。在确定目标对象的类型时,包括但不限于分类模型对至少一张待分类图像进行目标分类,得到目标对象属于不同类型的概率,将满足预设概率条件的类型作为目标对象的类型。预设概率条件包括但不限于概率值最大等。The final feature information carries the features of the target object, so that the final feature information is classified to obtain the type of the target object. When determining the type of the target object, including but not limited to, the classification model performs target classification on at least one image to be classified, obtains the probability that the target object belongs to different types, and takes the type that satisfies the preset probability condition as the type of the target object. The preset probability conditions include but are not limited to the maximum probability value and the like.
在一公开实施例中,分类模型在训练过程中采用ArcFace损失函数确定分类模型的损失值,通过ArcFace损失函数拉近同类目标对象的距离,拉远异类目标对象的距离,从而提高易混淆目标对象的分类能力。ArcFace损失函数简单易用,能很好地应用在分类模型的网络结构上,而不需要和其他损失函数相组合,同时在一定程度上减小过拟合问题,进而提高目标对象的分类性能。相较于softmax等损失函数,采用ArcFace损失函数确定分类模型的损失值时,分类模型的训练结果可以是第一个全连接层的权重与进入第一个全连接层的特征的夹角的余弦值。具体地,可将进入分类模型第一个全连接层的特征与第一个全连接层的权重之间的点积等于特征和权重归一化后的余弦距离,从而使用角余弦函数来计算归一化的特征和归一化的权重之间的目标角度,然后在目标角度上加上一个附加的角边距,再通过余弦函数得到目标的logit,再用一个固定的特征范数重新缩放所有logits,其后的相关步骤同softmax损失函数类似。以目标对象是肝脏肿瘤为例,考虑到肝脏肿瘤本身的特征信息是判断其类型的主要依据,但肝脏肿瘤大小不一,小则0.5cm以下,大则20cm以上,再加上目标对象之外的影响因素,例如待分类图像低分辨率,肝脏肿瘤周围其余类型肿瘤、与目标对象特征相似的血管、慢性肝病或者肝硬化背景等,本公开实施例中,ArcFace损失函数能学习到更好的肝脏肿瘤的特征表示,可实现同类肿瘤的聚合与异类肿瘤的远离,能有效地提高肿瘤的分类性能。其余目标对象的分类中,分类模型在训练过程中采用ArcFace损失函数确定分类模型的损失值的效果与之类似,在此不再一一举例。In a disclosed embodiment, the classification model uses the ArcFace loss function to determine the loss value of the classification model during the training process, and the ArcFace loss function is used to shorten the distance of similar target objects and shorten the distance of heterogeneous target objects, thereby increasing the confusion of target objects. classification ability. The ArcFace loss function is simple and easy to use, and can be well applied to the network structure of the classification model without being combined with other loss functions. At the same time, the overfitting problem is reduced to a certain extent, thereby improving the classification performance of the target object. Compared with loss functions such as softmax, when the ArcFace loss function is used to determine the loss value of the classification model, the training result of the classification model can be the cosine of the angle between the weight of the first fully connected layer and the feature entering the first fully connected layer. value. Specifically, the dot product between the features entering the first fully-connected layer of the classification model and the weights of the first fully-connected layer can be equal to the normalized cosine distance between the feature and the weight, so that the angular cosine function can be used to calculate the normalized The target angle between the normalized features and the normalized weights, then an additional angular margin is added to the target angle, and the logit of the target is obtained by the cosine function, and then all are rescaled with a fixed feature norm logits, and subsequent related steps are similar to the softmax loss function. Taking the target object as a liver tumor as an example, considering that the characteristic information of the liver tumor itself is the main basis for judging its type, but the size of the liver tumor varies, ranging from less than 0.5 cm to more than 20 cm, plus the size of the tumor outside the target object. Influencing factors, such as the low resolution of the image to be classified, other types of tumors around the liver tumor, blood vessels with similar characteristics to the target object, chronic liver disease or liver cirrhosis background, etc. In the embodiment of the present disclosure, the ArcFace loss function can learn better The feature representation of liver tumors can realize the aggregation of similar tumors and the distance of heterogeneous tumors, and can effectively improve the classification performance of tumors. In the classification of other target objects, the effect of using the ArcFace loss function to determine the loss value of the classification model in the training process of the classification model is similar, and no examples will be given here.
需要说明的是,ArcFace损失函数是一个利用margin来扩大不同类之间距离的损失函数,预测值是第一个全连接层的权重与进入第一个全连接层的特征的夹角的余弦值。其中,原理和操作流程如下:首先,进入第一个全连接层的特征与第一个全连接层的权重之间的点积等于特征和权重归一化后的余弦距离,其次,使用角余弦函数(arc-cosine function)来计算归一化的特征和归一化的权重之间的角度;然后,在目标角度上加上一个附加的角边距(additive angular margin),再通过余弦函数得到目标的logit,再用一个固定的特征范数重新缩放所有logits,随后的步骤与softmax loss中的步骤完全相同。It should be noted that the ArcFace loss function is a loss function that uses margin to expand the distance between different classes. The predicted value is the cosine of the angle between the weight of the first fully connected layer and the feature entering the first fully connected layer. . Among them, the principle and operation process are as follows: first, the dot product between the feature entering the first fully connected layer and the weight of the first fully connected layer is equal to the cosine distance after the normalization of the feature and the weight, and secondly, using the angular cosine function (arc-cosine function) to calculate the angle between normalized features and normalized weights; then, add an additional angular margin (additive angular margin) to the target angle, and then obtain through the cosine function The logit of the target, then rescales all logits with a fixed feature norm, and the subsequent steps are exactly the same as in the softmax loss.
通过上述方式,利用分类模型对至少一张待分类图像进行若干层特征提取,对应得到若干组初始特征信息;基于若干组初始特征信息中的至少一组初始特征信息,得到最终特征信息;对最终特征信息进行分类,得到目标对象的类型,实现了利用目标对象的特征信息进行目标分类。In the above manner, several layers of feature extraction are performed on at least one image to be classified by using the classification model, and several sets of initial feature information are correspondingly obtained; the final feature information is obtained based on at least one set of initial feature information in the several sets of initial feature information; the final feature information is obtained. The feature information is classified to obtain the type of the target object, which realizes the target classification using the feature information of the target object.
本领域技术人员可以理解,在具体实施方式的上述方法中,各步骤的撰写顺序并不意味着严格的执行顺序而对实施过程构成任何限定,各步骤的具体执行顺序应当以其功能和可能的内在逻辑确定。Those skilled in the art can understand that in the above method of the specific implementation, the writing order of each step does not mean a strict execution order but constitutes any limitation on the implementation process, and the specific execution order of each step should be based on its function and possible Internal logic is determined.
在本公开的一些实施例中,对至少一张待分类图像进行预处理,并提取到对应的二维或者三维的多期像肿瘤子图像块,也就是多期像肿瘤patch图像,以及对应的掩膜图像,即mask patch图像,将其一起输入到深度学习分类网络中。如图5所示,是本公开实施例的图像目标分类方法中分类模型所使用的网络架构示意图;其中,501输入至分类模型的批数据中随机包含等比例的不同类型肿瘤的数据,包括期像1,期像2,…,期像m以及多期像病灶mask的并集。502为CNN主干网络,即CNN backbone,其可以是U-Net的encoder或其变种,可以是Resnet或者其变种,可以是VGG16或者其变种,也可以是其他的用于分类的CNN结构;503为特征块(Feature Block),其中包含Adaptive average pooling、FC和Relu;其中前面得到的特征图经过自适应平均池化,全连接和Relu激活,从而得到一个一维的特征;同时每一Feature Block对应一特征_1。504为特征融合层(Feature Fusion),将多个一维特征进行拼接,每个特征有对应的权重系数,该系数是可学习的;其中包括:特征_1的权重系数_1,特征_2的权重系数_2,…,特征_n的权重系数_n。同时,CNN backbone中任一卷积层出来的特征图需要进入特征块和特征融合层,可以通过在训练过程中做试验来决定。本方案在实验过程中发现,引入特征融合层能提高模型性能;但是融合过多特征图会引起过拟合问题,尤其是融合靠前的卷积层出来的特征图。合理地调整进入特征融合结构的特征图数量,既能提高分类性能又能降低过拟合。505为全连接层(Fully Connected),即融合的特征送到FC,经softmax转换为各个肿瘤类别的分类概率值。506为预测的概率值。In some embodiments of the present disclosure, at least one image to be classified is preprocessed, and a corresponding two-dimensional or three-dimensional multi-phase image tumor sub-image block is extracted, that is, a multi-phase image tumor patch image, and corresponding The mask image, the mask patch image, is fed together into a deep learning classification network. As shown in FIG. 5 , it is a schematic diagram of the network architecture used by the classification model in the image target classification method according to the embodiment of the present disclosure; wherein, the batch data input to the classification model 501 randomly includes data of different types of tumors in equal proportions, including stage Like 1, phase like 2, ..., phase like m, and multi-phase like the union of lesion masks. 502 is the CNN backbone network, that is, CNN backbone, which can be the encoder of U-Net or its variant, Resnet or its variant, VGG16 or its variant, or other CNN structures for classification; 503 is Feature Block, which includes Adaptive average pooling, FC and Relu; the previously obtained feature map is subjected to adaptive average pooling, full connection and Relu activation to obtain a one-dimensional feature; at the same time, each Feature Block corresponds to A feature_1. 504 is a feature fusion layer (Feature Fusion), which splices multiple one-dimensional features, each feature has a corresponding weight coefficient, and the coefficient can be learned; including: the weight coefficient of feature_1_ 1, weight coefficient_2 of feature_2, ..., weight coefficient_n of feature_n. At the same time, the feature map from any convolutional layer in the CNN backbone needs to enter the feature block and feature fusion layer, which can be determined by experimenting in the training process. In the experiment process of this scheme, it is found that the introduction of feature fusion layers can improve the performance of the model; however, fusing too many feature maps will cause over-fitting problems, especially the feature maps obtained by fusing the front convolutional layers. Reasonably adjusting the number of feature maps entering the feature fusion structure can not only improve the classification performance but also reduce overfitting. 505 is a fully connected layer (Fully Connected), that is, the fused features are sent to FC, and converted into classification probability values of each tumor category through softmax. 506 is the predicted probability value.
请参阅图6,图6是本公开实施例提供的一种图像目标分类装置60的框架示意图。图像目标分类装置60包括图像获取模块61和目标分类模块62。图像获取模块61,配置为:获取包含目标对象的至少一张待分类图像,其中,至少一张待分类图像为属于至少一种扫描图像类别的医学图像;目标分类模块62,配置为:利用分类模型,对至少一张待分类图像进行目标分类,得到目标对象的类型。Please refer to FIG. 6 . FIG. 6 is a schematic frame diagram of an image object classification apparatus 60 provided by an embodiment of the present disclosure. The image object classification device 60 includes an image acquisition module 61 and an object classification module 62 . The image acquisition module 61 is configured to: acquire at least one image to be classified including the target object, wherein at least one image to be classified is a medical image belonging to at least one scanned image category; the target classification module 62 is configured to: use the classification The model performs target classification on at least one image to be classified to obtain the type of the target object.
在本公开的一些实施例中,目标分类模块62配置为:对至少一张待分类图像进行若干层特征提取,对应得到若干组初始特征信息;其中,每组初始特征信息的尺寸不同;基于若干组初始特征信息中的至少一组初始特征信息,得到最终特征信息;对最终特征信息进行分类,得到目标对象的类型。In some embodiments of the present disclosure, the target classification module 62 is configured to: perform several layers of feature extraction on at least one image to be classified, and correspondingly obtain several sets of initial feature information; wherein, the size of each set of initial feature information is different; At least one group of initial feature information in the group of initial feature information is obtained to obtain final feature information; the final feature information is classified to obtain the type of the target object.
在本公开的一些实施例中,目标分类模块62配置为:基于待分类图像中目标对象对应的初始区域,得到目标对象的最终区域;相应地,目标分类模块62,配置为:利用最终区域对至少一张待分类图像进行若干层特征提取,对应得到若干组初始特征信息;其中,在特征提取过程中,待分类图像中对应最终区域的权重高于待分类图像中其他区域的权重;和/或,初始特征信息中对应最终区域的特征比其他区域的特征更丰富。In some embodiments of the present disclosure, the target classification module 62 is configured to: obtain the final area of the target object based on the initial area corresponding to the target object in the image to be classified; correspondingly, the target classification module 62 is configured to: use the final area to At least one image to be classified is subjected to several layers of feature extraction, corresponding to several sets of initial feature information; wherein, in the feature extraction process, the weight of the corresponding final area in the image to be classified is higher than the weight of other areas in the image to be classified; and/ Or, the features corresponding to the final region in the initial feature information are more abundant than the features of other regions.
在本公开的一些实施例中,目标分类模块62配置为:获取至少一张待分类图像中目标对象对应的初始区域的并集,以作为目标对象的最终区域。In some embodiments of the present disclosure, the target classification module 62 is configured to obtain a union of initial regions corresponding to the target object in at least one image to be classified, as the final region of the target object.
在本公开的一些实施例中,目标分类模块62配置为:利用分类模型检测到第一待分类图像未标注有目标对象的初始区域,并基于第二待分类图像上标注的目标对象的初始区域以及第二待分类图像与第一待分类图像的配准关系,确定第一待分类图像上目标对象的初始区域。In some embodiments of the present disclosure, the target classification module 62 is configured to: use the classification model to detect the initial area of the first image to be classified that is not marked with the target object, and based on the initial area of the target object marked on the second image to be classified and the registration relationship between the second to-be-classified image and the first to-be-classified image to determine the initial area of the target object on the first to-be-classified image.
在本公开的一些实施例中,目标分类模块62配置为:将每组初始特征信息转换为预设维度;和/或,目标分类模块62配置为:利用至少一组初始特征信息的权重,将至少一组初始特征信息进行融合,得到最终特征信息。In some embodiments of the present disclosure, the target classification module 62 is configured to: convert each set of initial feature information into a preset dimension; and/or, the target classification module 62 is configured to: use the weight of at least one set of initial feature information to classify At least one set of initial feature information is fused to obtain final feature information.
在本公开的一些实施例中,每组初始特征信息的权重是在分类模型训练过程中确定的。In some embodiments of the present disclosure, the weight of each set of initial feature information is determined during the training process of the classification model.
在本公开的一些实施例中,预设维度为一维。In some embodiments of the present disclosure, the preset dimension is one dimension.
在本公开的一些实施例中,分类模型在训练过程中采用ArcFace损失函数确定分类模型的损失值;和/或,分类模型每次训练选择的批样本数据是利用数据生成器从样本数据集中选择的不同目标类型的数量为预设比例的样本数据。In some embodiments of the present disclosure, the classification model adopts the ArcFace loss function during the training process to determine the loss value of the classification model; and/or, the batch sample data selected for each training of the classification model is selected from the sample data set using a data generator The number of different target types is a preset ratio of sample data.
在本公开的一些实施例中,图像获取模块61配置为:分别从多张原始医学图像提取得到包含目标对象的待分类图像。In some embodiments of the present disclosure, the image acquisition module 61 is configured to: extract images to be classified including the target object from a plurality of original medical images, respectively.
在本公开的一些实施例中,图像获取模块61配置为:确定原始医学图像中目标对象的初始区域,按照预设比例扩大初始区域,得到待提取区域;从原始医学图像中提取待提取区域中的图像数据,得到待分类图像。In some embodiments of the present disclosure, the image acquisition module 61 is configured to: determine the initial area of the target object in the original medical image, expand the initial area according to a preset ratio to obtain the area to be extracted; extract the area to be extracted from the original medical image image data to obtain the image to be classified.
在本公开的一些实施例中,图像获取模块61配置为:将原始医学图像重采样至预设分辨率;调整原始医学图像中的像素值范围;将原始医学图像进行归一化处理;检测到第一原始医学图像未标注有目标对象的初始区域,利用第二原始医学图像上标注的目标对象的初始区域以及第二原始医学图像与第一原始医学图像的配准关系,确定第一原始医学图像上目标对象的初始区域。In some embodiments of the present disclosure, the image acquisition module 61 is configured to: resample the original medical image to a preset resolution; adjust the range of pixel values in the original medical image; normalize the original medical image; detect The first original medical image is not marked with the initial area of the target object, and the first original medical image is determined by using the initial area of the target object marked on the second original medical image and the registration relationship between the second original medical image and the first original medical image. The initial area of the target object on the image.
在本公开的一些实施例中,原始医学图像和待分类图像为二维图像;或者,原始医学图像为三维图像,待分类图像为二维图像或三维图像。In some embodiments of the present disclosure, the original medical image and the image to be classified are two-dimensional images; or, the original medical image is a three-dimensional image, and the image to be classified is a two-dimensional image or a three-dimensional image.
本公开实施例提供的图像目标分类装置60,图像获取模块61获取包含目标对象的至少一张待分类图像后,目标分类模块62利用分类模型对至少一张待分类图像进行目标分类,得到目标对象的类型,因此提出基于人工智能技术的图像目标分类方法,实现智能化目标分类。由于利用分类模型对待分类图像进行目标分类,不仅使得目标分类过程更加简单,减小对医生的依赖,提高目标分类速度,而且结合人工智能技术实现目标分类,以便辅助医生进行智能化疾病诊疗。In the image object classification device 60 provided by the embodiment of the present disclosure, after the image acquisition module 61 acquires at least one image to be classified including the target object, the object classification module 62 uses the classification model to perform object classification on the at least one image to be classified to obtain the target object Therefore, an image target classification method based on artificial intelligence technology is proposed to achieve intelligent target classification. Because the classification model is used to classify the images to be classified, it not only makes the target classification process simpler, reduces the dependence on doctors, and improves the speed of target classification, but also combines artificial intelligence technology to achieve target classification, so as to assist doctors in intelligent disease diagnosis and treatment.
请参阅图7,图7是本公开电子设备70一实施例的框架示意图。电子设备70包括相互耦接的存储器71和处理器72,处理器72用于执行存储器71中存储的程序指令,以实现上述任一图像目标分类方法实施例的步骤。在本公开的一些实施场景中,电子设备70可以包括但不限于:微型计算机、服务器,此外,电子设备70还可以包括笔记本电脑、平板电脑等移动设备,在此不做限定。Please refer to FIG. 7 , which is a schematic diagram of a framework of an embodiment of an electronic device 70 of the present disclosure. The electronic device 70 includes a memory 71 and a processor 72 coupled to each other, and the processor 72 is configured to execute program instructions stored in the memory 71 to implement the steps of any of the image object classification method embodiments described above. In some implementation scenarios of the present disclosure, the electronic device 70 may include, but is not limited to, a microcomputer and a server. In addition, the electronic device 70 may also include mobile devices such as a notebook computer and a tablet computer, which are not limited herein.
在本公开的一些实施例中,处理器72用于控制其自身以及存储器71以实现上述任一图像目标分类方法实施例的步骤。处理器72还可以称为中央处理单元(Central Processing Unit,CPU)。处理器72可能是一种集成电路芯片,具有信号的处理能力。处理器72还可以是通用处理器、数字信号处理器(Digital Signal Processor,DSP)、专用集成电路(Application Specific Integrated Circuit,ASIC)、现场可编程门阵列(Field-Programmable Gate Array,FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。另外,处理器72可以由集成电路芯片共同实现。In some embodiments of the present disclosure, the processor 72 is configured to control itself and the memory 71 to implement the steps of any of the image object classification method embodiments described above. The processor 72 may also be referred to as a central processing unit (Central Processing Unit, CPU). The processor 72 may be an integrated circuit chip with signal processing capability. The processor 72 may also be a general-purpose processor, a digital signal processor (Digital Signal Processor, DSP), an application-specific integrated circuit (Application Specific Integrated Circuit, ASIC), a Field-Programmable Gate Array (Field-Programmable Gate Array, FPGA) or other Programmable logic devices, discrete gate or transistor logic devices, discrete hardware components. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like. In addition, the processor 72 may be jointly implemented by an integrated circuit chip.
本公开实施例提供的电子设备70,获取包含目标对象的至少一张待分类图像后,利用分类模型对至少一张待分类图像进行目标分类,得到目标对象的类型,因此提出基于人工智能技术的图像目标分类方法,实现智能化目标分类。由于利用分类模型对待分类图像进行目标分类,不仅使得目标分类过程更加简单,减小对医生的依赖,提高目标分类速度,而且结合人工智能技术实现目标分类,以便辅助医生进行智能化疾病诊疗。The electronic device 70 provided in the embodiment of the present disclosure, after acquiring at least one image to be classified including a target object, uses a classification model to classify the at least one image to be classified, and obtains the type of the target object. Therefore, an artificial intelligence technology-based method is proposed. Image object classification method to achieve intelligent object classification. Because the classification model is used to classify the images to be classified, it not only makes the target classification process simpler, reduces the dependence on doctors, and improves the speed of target classification, but also combines artificial intelligence technology to achieve target classification, so as to assist doctors in intelligent disease diagnosis and treatment.
请参阅图8,图8是本公开计算机可读存储介质80一实施例的框架示意图。计算机可读存储介质80存储有能够被处理器运行的程序指令801,程序指令801用于实现上述任一图像目标分类方法实施例的步骤。Please refer to FIG. 8 , which is a schematic diagram of a framework of an embodiment of a computer-readable storage medium 80 of the present disclosure. The computer-readable storage medium 80 stores program instructions 801 that can be executed by the processor, and the program instructions 801 are used to implement the steps of any of the foregoing image object classification method embodiments.
本公开实施例提供的计算机可读存储介质80,获取包含目标对象的至少一张待分类图像后,利用分类模型对至少一张待分类图像进行目标分类,得到目标对象的类型,因此提出基于人工智能技术的图像目标分类方法,实现智能化目标分类。由于利用分类模型对待分类图像进行目标分类,不仅使得目标分类过程更加简单,减小对医生的依赖,提高目标分类速度,而且结合人工智能技术实现目标分类,以便辅助医生进行智能化疾病诊疗。In the computer-readable storage medium 80 provided by the embodiment of the present disclosure, after obtaining at least one image to be classified including the target object, the classification model is used to classify the at least one image to be classified to obtain the type of the target object. The image target classification method of intelligent technology realizes intelligent target classification. Because the classification model is used to classify the images to be classified, it not only makes the target classification process simpler, reduces the dependence on doctors, and improves the speed of target classification, but also combines artificial intelligence technology to achieve target classification, so as to assist doctors in intelligent disease diagnosis and treatment.
本公开实施例还提供一种计算机程序,该计算机程序包括计算机可读代码,该计算机可读代码在电子设备中运行的情况下,电子设备的处理器执行用于实现前述任一实施例提供的图像目标分类方法。该计算机程序产品可以具体通过硬件、软件或其结合的方式实现。在本公开的一些实施例中,计算机程序产品具体体现为计算机存储介质,在本公开的一些实施例中,计算机程序产品具体体现为软件产品,例如软件开发包(Software Development Kit,SDK)等等。An embodiment of the present disclosure further provides a computer program, where the computer program includes a computer-readable code, and when the computer-readable code is run in an electronic device, the processor of the electronic device executes the program for implementing any of the foregoing embodiments. Image object classification methods. The computer program product can be specifically implemented by hardware, software or a combination thereof. In some embodiments of the present disclosure, the computer program product is embodied as a computer storage medium, and in some embodiments of the present disclosure, the computer program product is embodied as a software product, such as a software development kit (Software Development Kit, SDK), etc. .
在一些实施例中,本公开实施例提供的图像目标分类装置具有的功能或包含的模块可以用于执行上文方法实施例描述的方法,其具体实现可以参照上文方法实施例的描述,为了简洁,这里不再赘述。In some embodiments, the functions or modules included in the image object classification apparatus provided in the embodiments of the present disclosure may be used to execute the methods described in the above method embodiments, and the specific implementation may refer to the above method embodiments. It is concise and will not be repeated here.
上文对各个实施例的描述倾向于强调各个实施例之间的不同之处,其相同或相似之处可以互相参考,为了简洁,本文不再赘述。The above descriptions of the various embodiments tend to emphasize the differences between the various embodiments, and the similarities or similarities can be referred to each other. For the sake of brevity, details are not repeated herein.
在本公开所提供的几个实施例中,应该理解到,所揭露的方法和装置,可以通过其它的方式实现。例如,以上所描述的装置实施方式仅仅是示意性的,例如,模块或单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,装置或单元的间接耦合或通信连接,可以是电性、机械或其它的形式。In the several embodiments provided in the present disclosure, it should be understood that the disclosed method and apparatus may be implemented in other manners. For example, the device implementations described above are only illustrative. For example, the division of modules or units is only a logical function division. In actual implementation, there may be other divisions. For example, units or components may be combined or integrated. to another system, or some features can be ignored, or not implemented. On the other hand, the shown or discussed mutual coupling or direct coupling or communication connection may be through some interfaces, indirect coupling or communication connection of devices or units, which may be in electrical, mechanical or other forms.
另外,在本公开各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。In addition, each functional unit in each embodiment of the present disclosure may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit. The above-mentioned integrated units may be implemented in the form of hardware, or may be implemented in the form of software functional units.
集成的单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本公开的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的全部或部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)或处理器(processor)执行本公开各个实施方式方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(Read-Only Memory,ROM)、随机存取存储器(Random Access Memory,RAM)、磁碟或者光盘等各种可以存储程序代码的介质。The integrated unit, if implemented as a software functional unit and sold or used as a stand-alone product, may be stored in a computer-readable storage medium. Based on this understanding, the technical solutions of the present disclosure can be embodied in the form of software products in essence, or the part that contributes to the prior art, or all or part of the technical solutions, and the computer software product is stored in a storage medium , including several instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) or a processor (processor) to execute all or part of the steps of the methods of the various embodiments of the present disclosure. The aforementioned storage medium includes: U disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk and other media that can store program codes .
工业实用性Industrial Applicability
本公开提供了一种图像目标分类方法、装置、设备、存储介质及程序,其中,图像目标分类方法包括:获取包含目标对象的至少一张待分类图像,其中,所述至少一张待分类图像为属于至少一种扫描图像类别的医学图像;利用分类模型,对所述至少一张待分类图像进行目标分类,得到目标对象的类型。The present disclosure provides an image object classification method, device, device, storage medium and program, wherein the image object classification method includes: acquiring at least one image to be classified including a target object, wherein the at least one image to be classified is a medical image belonging to at least one type of scanned image; using a classification model, perform target classification on the at least one image to be classified to obtain the type of the target object.

Claims (18)

  1. 一种图像目标分类方法,包括:An image object classification method, comprising:
    获取包含目标对象的至少一张待分类图像,其中,所述至少一张待分类图像为属于至少一种扫描图像类别的医学图像;acquiring at least one image to be classified containing the target object, wherein the at least one image to be classified is a medical image belonging to at least one scanned image category;
    利用分类模型,对所述至少一张待分类图像进行目标分类,得到所述目标对象的类型。Using a classification model, the at least one image to be classified is subjected to target classification to obtain the type of the target object.
  2. 根据权利要求1所述的方法,所述对所述至少一张待分类图像进行目标分类,得到所述目标对象的类型,包括:The method according to claim 1, wherein the target classification is performed on the at least one image to be classified to obtain the type of the target object, comprising:
    对所述至少一张待分类图像进行若干层特征提取,对应得到若干组初始特征信息;其中,每组所述初始特征信息的尺寸不同;Performing several layers of feature extraction on the at least one image to be classified, correspondingly obtaining several groups of initial feature information; wherein, the size of each group of the initial feature information is different;
    基于所述若干组初始特征信息中的至少一组初始特征信息,得到最终特征信息;Obtain final feature information based on at least one set of initial feature information in the several sets of initial feature information;
    对所述最终特征信息进行分类,得到所述目标对象的类型。The final feature information is classified to obtain the type of the target object.
  3. 根据权利要求1所述的方法,所述对所述至少一张待分类图像进行目标分类,得到所述目标对象的类型之前,所述方法还包括:The method according to claim 1, before the target classification is performed on the at least one image to be classified and the type of the target object is obtained, the method further comprises:
    基于所述待分类图像中所述目标对象对应的初始区域,得到所述目标对象的最终区域;obtaining the final area of the target object based on the initial area corresponding to the target object in the image to be classified;
    相应地,所述对所述至少一张待分类图像进行若干层特征提取,对应得到若干组初始特征信息,包括:Correspondingly, several layers of feature extraction are performed on the at least one image to be classified, and several sets of initial feature information are correspondingly obtained, including:
    利用所述最终区域对所述至少一张待分类图像进行若干层特征提取,对应得到若干组初始特征信息;其中,在特征提取过程中,所述待分类图像中对应所述最终区域的权重高于所述待分类图像中其他区域的权重;和/或,所述初始特征信息中对应所述最终区域的特征比其他区域的特征更丰富。Using the final area to perform several layers of feature extraction on the at least one image to be classified, correspondingly obtain several sets of initial feature information; wherein, in the feature extraction process, the weight of the image to be classified corresponding to the final area is high the weight of other regions in the image to be classified; and/or, the features corresponding to the final region in the initial feature information are more abundant than the features of other regions.
  4. 根据权利要求3所述的方法,所述基于所述待分类图像中所述目标对象对应的初始区域,得到所述目标对象的最终区域,包括:The method according to claim 3, wherein the obtaining of the final area of the target object based on the initial area corresponding to the target object in the image to be classified includes:
    获取所述至少一张待分类图像中所述目标对象对应的初始区域的并集,以作为所述目标对象的最终区域。The union of the initial regions corresponding to the target object in the at least one image to be classified is obtained as the final region of the target object.
  5. 根据权利要求3或4所述的方法,所述至少一张待分类图像包括未标注所述目标对象的初始区域的第一待分类图像和标注所述目标对象的初始区域的第二待分类图像;所述基于所述待分类图像中所述目标对象对应的初始区域,得到所述目标对象的最终区域之前,所述方法还包括:The method according to claim 3 or 4, wherein the at least one image to be classified comprises a first image to be classified without an initial area of the target object and a second image to be classified with an initial area of the target object marked ; Before obtaining the final area of the target object based on the initial area corresponding to the target object in the image to be classified, the method further includes:
    利用所述分类模型检测到所述第一待分类图像未标注有所述目标对象的初始区域,并基于所述第二待分类图像上标注的所述目标对象的初始区域以及所述第二待分类图像与所述第一待分类图像的配准关系,确定所述第一待分类图像上所述目标对象的初始区域。Using the classification model to detect that the first image to be classified is not marked with the initial area of the target object, and based on the initial area of the target object marked on the second to-be-classified image and the second to-be-classified image The registration relationship between the classified image and the first to-be-classified image determines the initial area of the target object on the first to-be-classified image.
  6. 根据权利要求2至5任一项所述的方法,所述基于所述若干组初始特征信息中的至少一组初始特征信息,得到最终特征信息之前,所述方法还包括:The method according to any one of claims 2 to 5, before obtaining the final feature information based on at least one set of initial feature information in the several sets of initial feature information, the method further comprises:
    将每组所述初始特征信息转换为预设维度;converting each group of the initial feature information into a preset dimension;
    和/或,所述基于所述若干组初始特征信息中的至少一组初始特征信息,得到最终特征信息,包括:And/or, obtaining final feature information based on at least one set of initial feature information in the several groups of initial feature information, including:
    利用所述至少一组初始特征信息的权重,将所述至少一组初始特征信息进行融合,得到所述最终特征信息。Using the weight of the at least one set of initial feature information, the at least one set of initial feature information is fused to obtain the final feature information.
  7. 根据权利要求6所述的方法,每组所述初始特征信息的权重是在所述分类模型训 练过程确定的。According to the method of claim 6, the weight of each set of the initial feature information is determined during the training process of the classification model.
  8. 根据权利要求6或7所述的方法,所述预设维度为一维。The method according to claim 6 or 7, wherein the preset dimension is one dimension.
  9. 根据权利要求1至8任一项所述的方法,所述分类模型在训练过程中采用ArcFace损失函数确定所述分类模型的损失值;和/或,所述分类模型每次训练选择的批样本数据是利用数据生成器从样本数据集中选择的不同目标类型的数量为预设比例的样本数据。The method according to any one of claims 1 to 8, wherein the classification model adopts an ArcFace loss function during the training process to determine the loss value of the classification model; and/or, the batch samples selected for each training of the classification model The data is sample data with a preset proportion of the number of different target types selected from the sample data set by the data generator.
  10. 根据权利要求1至9任一项所述的方法,所述获取包含目标对象的至少一张待分类图像,包括:The method according to any one of claims 1 to 9, wherein the acquiring at least one image to be classified containing the target object comprises:
    分别从多张原始医学图像提取得到包含所述目标对象的待分类图像。The to-be-classified images containing the target object are respectively extracted from a plurality of original medical images.
  11. 根据权利要求10所述的方法,所述分别从多张原始医学图像提取得到包含所述目标对象的待分类图像,包括:The method according to claim 10, wherein the images to be classified including the target object are extracted from a plurality of original medical images respectively, comprising:
    确定所述原始医学图像中所述目标对象的初始区域,按照所述预设比例扩大所述初始区域,得到待提取区域;determining the initial area of the target object in the original medical image, and expanding the initial area according to the preset ratio to obtain the area to be extracted;
    从所述原始医学图像中提取所述待提取区域中的图像数据,得到所述待分类图像。The image data in the to-be-extracted area is extracted from the original medical image to obtain the to-be-classified image.
  12. 根据权利要求10或11所述的方法,所述分别从多张原始医学图像提取得到包含所述目标对象的待分类图像之前,所述方法还包括以下至少一个步骤:The method according to claim 10 or 11, before the image to be classified containing the target object is extracted from a plurality of original medical images, the method further comprises at least one of the following steps:
    将所述原始医学图像重采样至预设分辨率;resampling the original medical image to a preset resolution;
    调整所述原始医学图像中的像素值范围;adjusting the range of pixel values in the original medical image;
    将所述原始医学图像进行归一化处理;normalizing the original medical image;
    检测到第一原始医学图像未标注有所述目标对象的初始区域,利用第二原始医学图像上标注的所述目标对象的初始区域以及所述第二原始医学图像与所述第一原始医学图像的配准关系,确定所述第一原始医学图像上所述目标对象的初始区域。It is detected that the initial area of the target object is not marked in the first original medical image, and the initial area of the target object marked on the second original medical image and the second original medical image and the first original medical image are used. The registration relationship is determined, and the initial area of the target object on the first original medical image is determined.
  13. 根据权利要求10至12任一项所述的方法,所述原始医学图像和所述待分类图像为二维图像;或者,所述原始医学图像为三维图像,所述待分类图像为二维图像或三维图像。The method according to any one of claims 10 to 12, wherein the original medical image and the image to be classified are two-dimensional images; or, the original medical image is a three-dimensional image, and the image to be classified is a two-dimensional image or 3D images.
  14. 根据权利要求13所述的方法,所述原始医学图像为三维图像,所述待分类图像为对所述原始医学图像中所述目标对象最大面积所在层提取得到的二维图像。The method according to claim 13, wherein the original medical image is a three-dimensional image, and the image to be classified is a two-dimensional image obtained by extracting the layer where the target object has the largest area in the original medical image.
  15. 一种图像目标分类装置,包括:An image object classification device, comprising:
    图像获取模块,配置为获取包含目标对象的至少一张待分类图像,其中,所述至少一张待分类图像为属于至少一种扫描图像类别的医学图像;an image acquisition module, configured to acquire at least one image to be classified including the target object, wherein the at least one image to be classified is a medical image belonging to at least one scanned image category;
    目标分类模块,配置为利用分类模型,对所述至少一张待分类图像进行目标分类,得到所述目标对象的类型。The target classification module is configured to use a classification model to perform target classification on the at least one image to be classified to obtain the type of the target object.
  16. 一种电子设备,包括相互耦接的存储器和处理器,所述处理器用于执行所述存储器中存储的程序指令,以实现权利要求1至14任一项所述的图像目标分类方法。An electronic device includes a memory and a processor coupled to each other, the processor is configured to execute program instructions stored in the memory, so as to implement the image object classification method according to any one of claims 1 to 14.
  17. 一种计算机可读存储介质,其上存储有程序指令,所述程序指令被处理器执行时实现权利要求1至14任一项所述的图像目标分类方法。A computer-readable storage medium having program instructions stored thereon, the program instructions implement the image object classification method according to any one of claims 1 to 14 when the program instructions are executed by a processor.
  18. 一种计算机程序,所述计算机程序包括计算机可读代码,在所述计算机可读代码在电子设备中运行的情况下,所述电子设备的处理器执行用于实现如权利要求1至14任一项所述的图像目标分类方法。A computer program comprising computer readable codes, when the computer readable codes are executed in an electronic device, the processor of the electronic device executes the code for realizing any one of claims 1 to 14 The image object classification method described in item.
PCT/CN2020/139913 2020-11-03 2020-12-28 Image object classification method and apparatus, device, storage medium and program WO2022095258A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202011212261.1 2020-11-03
CN202011212261.1A CN112329844A (en) 2020-11-03 2020-11-03 Image object classification method and related device, equipment and storage medium

Publications (1)

Publication Number Publication Date
WO2022095258A1 true WO2022095258A1 (en) 2022-05-12

Family

ID=74324556

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/139913 WO2022095258A1 (en) 2020-11-03 2020-12-28 Image object classification method and apparatus, device, storage medium and program

Country Status (3)

Country Link
CN (1) CN112329844A (en)
TW (1) TW202219832A (en)
WO (1) WO2022095258A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116452957A (en) * 2023-06-20 2023-07-18 小米汽车科技有限公司 Quality detection method and device for image annotation data and electronic equipment
CN116524135A (en) * 2023-07-05 2023-08-01 方心科技股份有限公司 Three-dimensional model generation method and system based on image
CN117689660A (en) * 2024-02-02 2024-03-12 杭州百子尖科技股份有限公司 Vacuum cup temperature quality inspection method based on machine vision
CN117689660B (en) * 2024-02-02 2024-05-14 杭州百子尖科技股份有限公司 Vacuum cup temperature quality inspection method based on machine vision

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114663715B (en) * 2022-05-26 2022-08-26 浙江太美医疗科技股份有限公司 Medical image quality control and classification model training method and device and computer equipment

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109872306A (en) * 2019-01-28 2019-06-11 腾讯科技(深圳)有限公司 Medical image cutting method, device and storage medium
CN111476802A (en) * 2020-04-09 2020-07-31 山东财经大学 Medical image segmentation and tumor detection method and device based on dense convolution model and readable storage medium
CN111507381A (en) * 2020-03-31 2020-08-07 上海商汤智能科技有限公司 Image recognition method and related device and equipment
CN111755118A (en) * 2020-03-16 2020-10-09 腾讯科技(深圳)有限公司 Medical information processing method, medical information processing device, electronic equipment and storage medium

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110993094B (en) * 2019-11-19 2023-05-23 中国科学院深圳先进技术研究院 Intelligent auxiliary diagnosis method and terminal based on medical image

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109872306A (en) * 2019-01-28 2019-06-11 腾讯科技(深圳)有限公司 Medical image cutting method, device and storage medium
CN111755118A (en) * 2020-03-16 2020-10-09 腾讯科技(深圳)有限公司 Medical information processing method, medical information processing device, electronic equipment and storage medium
CN111507381A (en) * 2020-03-31 2020-08-07 上海商汤智能科技有限公司 Image recognition method and related device and equipment
CN111476802A (en) * 2020-04-09 2020-07-31 山东财经大学 Medical image segmentation and tumor detection method and device based on dense convolution model and readable storage medium

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116452957A (en) * 2023-06-20 2023-07-18 小米汽车科技有限公司 Quality detection method and device for image annotation data and electronic equipment
CN116452957B (en) * 2023-06-20 2023-09-26 小米汽车科技有限公司 Quality detection method and device for image annotation data and electronic equipment
CN116524135A (en) * 2023-07-05 2023-08-01 方心科技股份有限公司 Three-dimensional model generation method and system based on image
CN116524135B (en) * 2023-07-05 2023-09-15 方心科技股份有限公司 Three-dimensional model generation method and system based on image
CN117689660A (en) * 2024-02-02 2024-03-12 杭州百子尖科技股份有限公司 Vacuum cup temperature quality inspection method based on machine vision
CN117689660B (en) * 2024-02-02 2024-05-14 杭州百子尖科技股份有限公司 Vacuum cup temperature quality inspection method based on machine vision

Also Published As

Publication number Publication date
TW202219832A (en) 2022-05-16
CN112329844A (en) 2021-02-05

Similar Documents

Publication Publication Date Title
WO2022095258A1 (en) Image object classification method and apparatus, device, storage medium and program
Cai et al. A review of the application of deep learning in medical image classification and segmentation
CN110033456B (en) Medical image processing method, device, equipment and system
Zuo et al. R2AU-Net: attention recurrent residual convolutional neural network for multimodal medical image segmentation
WO2021196955A1 (en) Image recognition method and related apparatus, and device
JP2020518915A (en) System and method for automated fundus image analysis
Luo et al. Micro-vessel image segmentation based on the AD-UNet model
Zhao et al. SCOAT-Net: A novel network for segmenting COVID-19 lung opacification from CT images
CN111476793B (en) Dynamic enhanced magnetic resonance imaging processing method, system, storage medium and terminal
WO2022062590A1 (en) Image recognition method and apparatus, device, storage medium and program
CN110956632B (en) Method and device for automatically detecting pectoralis major region in molybdenum target image
JP2023540910A (en) Connected Machine Learning Model with Collaborative Training for Lesion Detection
CN112001895B (en) Thyroid calcification detection device
WO2019184851A1 (en) Image processing method and apparatus, and training method for neural network model
Tan et al. Automated vessel segmentation in lung CT and CTA images via deep neural networks
CN114782307A (en) Enhanced CT image colorectal cancer staging auxiliary diagnosis system based on deep learning
CN111260651A (en) Stomach low-quality MRI image segmentation method based on deep migration learning
WO2019223121A1 (en) Lesion site recognition method and apparatus, and computer apparatus and readable storage medium
CN111062909A (en) Method and equipment for judging benign and malignant breast tumor
CN116109610A (en) Method and system for segmenting breast tumor in ultrasonic examination report image
Rodríguez et al. Color segmentation applied to study of the angiogenesis. Part I
Sridhar et al. Lung Segment Anything Model (LuSAM): A Prompt-integrated Framework for Automated Lung Segmentation on ICU Chest X-Ray Images
CN113870194A (en) Deep layer characteristic and superficial layer LBP characteristic fused breast tumor ultrasonic image processing device
CN114972192B (en) Breast molybdenum target pectoral large muscle region segmentation method based on deep learning
Sreelekshmi et al. A Review on Multimodal Medical Image Fusion

Legal Events

Date Code Title Description
ENP Entry into the national phase

Ref document number: 2021578180

Country of ref document: JP

Kind code of ref document: A

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20960715

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 04.10.2023)