WO2022095258A1 - Procédé et appareil de classification d'objet d'image, dispositif, support de stockage et programme - Google Patents

Procédé et appareil de classification d'objet d'image, dispositif, support de stockage et programme Download PDF

Info

Publication number
WO2022095258A1
WO2022095258A1 PCT/CN2020/139913 CN2020139913W WO2022095258A1 WO 2022095258 A1 WO2022095258 A1 WO 2022095258A1 CN 2020139913 W CN2020139913 W CN 2020139913W WO 2022095258 A1 WO2022095258 A1 WO 2022095258A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
classified
target object
feature information
initial
Prior art date
Application number
PCT/CN2020/139913
Other languages
English (en)
Chinese (zh)
Inventor
朱雅靖
陈翼男
罗祥德
任家敏
Original Assignee
上海商汤智能科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 上海商汤智能科技有限公司 filed Critical 上海商汤智能科技有限公司
Publication of WO2022095258A1 publication Critical patent/WO2022095258A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features

Definitions

  • the present disclosure relates to the technical field of image processing, and in particular, to an image object classification method, apparatus, device, storage medium and program.
  • the scan image categories often include time-related pre-contrast scan, early arterial phase, late arterial phase, portal venous phase, and delayed phase, etc.
  • the scan image category can also include scan parameters related to the T1-weighted inverse imaging, T1-weighted in-phase imaging, T2-weighted imaging, diffusion-weighted imaging, and surface diffusion coefficient imaging, etc.
  • the doctor in the process of disease diagnosis and treatment, the doctor usually needs to repeatedly check the signs of the target object such as the tumor on the medical image, which makes the determination of the type of the tumor overly dependent on the professional level of the doctor, and the problem of low efficiency in determining the type of the tumor. .
  • the present disclosure provides at least an image object classification method, apparatus, device, storage medium and program.
  • the embodiment of the present disclosure provides an image object classification method, and the image object classification method includes:
  • the at least one image to be classified is a medical image belonging to at least one scanned image category
  • the at least one image to be classified is subjected to target classification to obtain the type of the target object.
  • the classification model is used to classify the at least one image to be classified, and the type of the target object is obtained. Because the classification model is used to classify the image to be classified, intelligence is realized. Target classification, and no manual target classification is required, which can reduce the dependence on manual work and improve the efficiency of target classification.
  • performing target classification on the at least one image to be classified to obtain the type of the target object includes:
  • the final feature information is classified to obtain the type of the target object.
  • the final feature information can be classified to obtain the type of the target object, so the target classification is realized by using the feature information of the target object.
  • the method before the target classification is performed on the at least one image to be classified and the type of the target object is obtained, the method further includes:
  • the final area to perform several layers of feature extraction on the at least one image to be classified, correspondingly obtain several sets of initial feature information; wherein, in the feature extraction process, the weight of the image to be classified corresponding to the final area is high the weight of other regions in the image to be classified; and/or, the features corresponding to the final region in the initial feature information are more abundant than the features of other regions.
  • the weight of the corresponding final region in the image to be classified is higher than the weights of other regions in the image to be classified, so that the classification model tends to extract features with more details for the final region; and /or, the features corresponding to the final region in the initial feature information are richer than those of other regions; thus, the classification model can learn the feature information of the target object itself better by using the initial feature information of the image to be classified, to a certain extent Reduce the impact of noise interference around the target object on target classification.
  • obtaining the final area of the target object based on the initial area corresponding to the target object in the image to be classified includes:
  • the union of the initial regions corresponding to the target object in the at least one image to be classified is obtained as the final region of the target object.
  • the final area of the target object is the union of the initial areas of the target object in the image to be classified
  • the final area is greater than or equal to any initial area, ensuring that the final area of the target object can contain different targets in the to-be-classified image.
  • the object corresponds to the area, so that the feature information of the target object can be paid attention to as much as possible when the feature extraction of the image to be classified is performed.
  • the at least one image to be classified includes a first image to be classified without an initial region of the target object and a second image to be classified with an initial region of the target object labeled; Before obtaining the final area of the target object based on the initial area corresponding to the target object in the image to be classified, the method further includes:
  • the classification model to detect that the first image to be classified is not marked with the initial area of the target object, and based on the initial area of the target object marked on the second to-be-classified image and the second to-be-classified image The registration relationship between the classified image and the first to-be-classified image determines the initial area of the target object on the first to-be-classified image.
  • the classification model can be used to determine the initial area of the target object for the first to-be-classified image that is not labeled with the initial area of the target object, so as to complete the labeling, so that the to-be-classified images all include the initial area.
  • the method before obtaining the final feature information based on at least one set of initial feature information in the several sets of initial feature information, the method further includes:
  • obtaining final feature information based on at least one set of initial feature information in the several groups of initial feature information including:
  • the at least one set of initial feature information is fused to obtain the final feature information.
  • each group of initial feature information is uniformly converted into a preset dimension, which facilitates subsequent acquisition of final feature information.
  • the weight of at least one set of initial feature information can be used to fuse the initial feature information of different sizes extracted from at least one layer of features to obtain the final feature information.
  • the initial feature information of small size may be compressed to remove important features.
  • the weight of each set of the initial feature information is determined during the training process of the classification model.
  • the weight of the initial feature information for fusion is determined through the iterative training of the classification model, so that the final feature information obtained by using the weight fusion can better reflect the characteristics of the target object and further improve the classification performance.
  • the preset dimension is one dimension.
  • each group of initial feature information can be converted into one-dimensional, data unification is realized, and subsequent fusion is facilitated.
  • the classification model uses an ArcFace loss function during the training process to determine the loss value of the classification model; and/or, the batch sample data selected for each training of the classification model is generated by using the data
  • the number of different target types selected by the processor from the sample data set is a preset proportion of sample data.
  • the ArcFace loss function to determine the loss value of the classification model can aggregate the feature information of the same target objects and keep the feature information of different types of target objects away, thereby improving the classification performance of the target objects.
  • the data generator is used to select sample data from the sample data set, and the sample data with a preset ratio of different target types is used as the batch sample data, so that the target types of the batch sample data for training the classification model are more balanced.
  • the acquiring at least one image to be classified including the target object includes:
  • the to-be-classified images containing the target object are respectively extracted from a plurality of original medical images.
  • the image to be classified is obtained, and the image to be classified can be extracted from the original medical image.
  • the image size of the subsequent classification can be reduced, and some backgrounds in the original medical image can be avoided to a certain extent. Therefore, the consumption of processing resources for subsequent classification can be reduced, and the classification performance can be improved.
  • the image to be classified containing the target object is extracted from multiple original medical images, including:
  • the image data in the to-be-extracted area is extracted from the original medical image to obtain the to-be-classified image.
  • the initial area is the area containing the target object, and the initial area of the target object is expanded according to a preset ratio, so that the obtained area to be extracted contains both the target object and some background information around the target object, so that the area to be extracted can be extracted.
  • the image to be classified can include the target object and some background information.
  • the method before the image to be classified including the target object is extracted from multiple original medical images, the method further includes at least one of the following steps:
  • the initial area of the target object is not marked in the first original medical image, and the initial area of the target object marked on the second original medical image and the second original medical image and the first original medical image are used.
  • the registration relationship is determined, and the initial area of the target object on the first original medical image is determined.
  • the original medical image can be preprocessed before the image to be classified is extracted from the original medical image, and the unclassified image can be unified. Image parameters of the image to improve the quality of the image to be classified.
  • the original medical image and the image to be classified are two-dimensional images; or, the original medical image is a three-dimensional image, and the image to be classified is a two-dimensional image or a three-dimensional image.
  • the image to be classified is extracted from the original medical image. If the original medical image is a two-dimensional image, the image to be classified is a two-dimensional image; and if the original medical image is a three-dimensional image, the image to be classified is a two-dimensional image.
  • the dimensions can be two-dimensional or three-dimensional.
  • the original medical image is a three-dimensional image
  • the image to be classified is a two-dimensional image obtained by extracting a layer where the target object has the largest area in the original medical image.
  • the layer where the maximum area of the target object is located in the original medical image can be extracted as the image to be classified, so that the extraction range of the target object in the image to be classified is larger. , contains more information about the target object and improves the classification accuracy of the target object.
  • the embodiment of the present disclosure also provides an image object classification device, and the image object classification device includes:
  • an image acquisition module configured to acquire at least one image to be classified including the target object, wherein the at least one image to be classified is a medical image belonging to at least one scanned image category;
  • the target classification module is configured to use a classification model to perform target classification on the at least one image to be classified to obtain the type of the target object.
  • the target classification module is configured to:
  • the final feature information is classified to obtain the type of the target object.
  • the target classification module is configured to:
  • the target classification module configured as:
  • the final area to perform several layers of feature extraction on the at least one image to be classified, correspondingly obtain several sets of initial feature information; wherein, in the feature extraction process, the weight of the image to be classified corresponding to the final area is high the weight of other regions in the image to be classified; and/or, the features corresponding to the final region in the initial feature information are more abundant than the features of other regions.
  • the target classification module is configured to:
  • the union of the initial regions corresponding to the target object in the at least one image to be classified is obtained as the final region of the target object.
  • the target classification module is configured to:
  • the classification model to detect that the first image to be classified is not marked with the initial area of the target object, and based on the initial area of the target object marked on the second to-be-classified image and the second to-be-classified image The registration relationship between the classified image and the first to-be-classified image determines the initial area of the target object on the first to-be-classified image.
  • the target classification module is configured to:
  • the target classification module configured as:
  • the at least one set of initial feature information is fused to obtain the final feature information.
  • the weight of each set of the initial feature information is determined during the training process of the classification model.
  • the preset dimension is one dimension.
  • the classification model uses an ArcFace loss function during the training process to determine the loss value of the classification model; and/or, the batch sample data selected for each training of the classification model is generated by using the data
  • the number of different target types selected by the processor from the sample data set is a preset proportion of sample data.
  • the image acquisition module is configured to:
  • the to-be-classified images containing the target object are respectively extracted from a plurality of original medical images.
  • the image acquisition module is configured to:
  • the image data in the to-be-extracted area is extracted from the original medical image to obtain the to-be-classified image.
  • the image acquisition module is configured to:
  • the initial area of the target object is not marked in the first original medical image, and the initial area of the target object marked on the second original medical image and the second original medical image and the first original medical image are used.
  • the registration relationship is determined, and the initial area of the target object on the first original medical image is determined.
  • the original medical image and the image to be classified are two-dimensional images; or, the original medical image is a three-dimensional image, and the image to be classified is a two-dimensional image or a three-dimensional image.
  • the original medical image is a three-dimensional image
  • the image to be classified is a two-dimensional image obtained by extracting a layer where the target object has the largest area in the original medical image.
  • An embodiment of the present disclosure also provides an electronic device, including a memory and a processor coupled to each other, where the processor is configured to execute program instructions stored in the memory, so as to implement the image object classification method provided in any of the previous embodiments .
  • the embodiments of the present disclosure also provide a computer-readable storage medium, which stores program instructions, and the program instructions are executed by a processor as the image object classification method provided in any of the previous embodiments.
  • An embodiment of the present disclosure also provides a computer program, the computer program includes computer-readable codes, and when the computer-readable codes are executed in an electronic device, the processor of the electronic device executes any of the preceding implementations The image object classification method described in the example.
  • an image target classification method based on artificial intelligence technology is proposed to achieve intelligent target classification. Because the classification model is used to classify the images to be classified, it not only makes the target classification process simpler, reduces the dependence on doctors, improves the speed and accuracy of target classification, but also combines artificial intelligence technology to achieve target classification, so as to assist doctors in intelligent diseases diagnosis and treatment.
  • FIG. 1 is a schematic flowchart of an image target classification method provided by an embodiment of the present disclosure
  • FIG. 2 is a schematic diagram of a system architecture to which the image object classification method according to the embodiment of the present disclosure can be applied;
  • FIG. 3 is a schematic flowchart of obtaining at least one image to be classified according to an embodiment of the present disclosure
  • FIG. 4 is a schematic flowchart of a target classification for at least one image to be classified according to an embodiment of the present disclosure
  • FIG. 5 is a schematic diagram of a network architecture used by a classification model in the image target classification method according to an embodiment of the present disclosure
  • FIG. 6 is a schematic diagram of a framework of an image object classification apparatus 60 provided by an embodiment of the present disclosure.
  • FIG. 7 is a schematic frame diagram of an electronic device 70 provided by an embodiment of the present disclosure.
  • FIG. 8 is a schematic diagram of a framework of a computer-readable storage medium 80 provided by an embodiment of the present disclosure.
  • three-dimensional imaging technology based on CT and MR plays a crucial role in medical imaging diagnosis, and is one of the main imaging examination methods for diagnosing, for example, liver diseases.
  • the scanning sequence of CT examination mainly includes the plain scan phase, the dynamic enhancement phase, the arterial phase, the portal venous phase and the delayed phase.
  • the plain scan period is generally used to observe changes in the liver surface, whether there are fatty liver, liver fibrosis, liver cirrhosis and other diseases.
  • phase images with dynamic enhancement can show the specific image features of the lesion.
  • HCC Hepatocellular Carcinoma
  • HCC mainly occurs in patients with chronic liver disease and liver cirrhosis, and the corresponding changes in liver surface morphology can be observed from the plain scan period. Or the same density as the liver parenchyma; after enhanced scanning, HCC in each phase showed: marked enhancement or inhomogeneous enhancement in arterial phase, accompanied by low-density capsule; contrast agent outflow in portal venous phase, showing enhanced capsule at the same time ; the delayed phase presents a delayed enhanced envelope. Therefore, in a feasible implementation manner, it can be determined whether the target tumor is HCC by identifying the imaging features exhibited by the images in multiple stages. Compared with the single-phase image, the judgment accuracy is higher, because the image characteristics of the small liver metastases with rich blood supply in the plain and arterial phase are similar to those of small HCC. classification task, which can further improve the accuracy of image classification.
  • medical image analysis generally has problems such as less labeled data, complex and difficult tasks, and at the same time, in order to better characterize lesions, it is necessary to analyze the correlation between sequences.
  • the existence of these problems limits the complexity and depth of deep learning networks to a certain extent, and some other strategies need to be introduced to solve the task of medical image analysis.
  • the image features of the tumor itself are the main basis for judging its type, and there may be various noises around the target tumor, which will mislead the deep learning network to learn some wrong features; liver tumors vary in size and small.
  • the network needs to be able to take into account the characteristics of liver tumors, and improve the identification ability of small tumors on the basis of ensuring high-precision classification and identification of large tumors; limited to the resolution of CT scan images,
  • the imaging features of liver tumors are not necessarily obvious.
  • the present disclosure provides at least one image target classification method, which uses a classification model to classify images to be classified, which not only makes the target classification process simpler, reduces the dependence on doctors, and improves the speed and accuracy of target classification, And combined with artificial intelligence technology to achieve target classification, in order to assist doctors in intelligent disease diagnosis and treatment.
  • FIG. 1 is a schematic flowchart of an image object classification method provided by an embodiment of the present disclosure. Specifically, the following steps can be included:
  • Step S11 Acquire at least one image to be classified that includes the target object.
  • At least one image to be classified is a medical image belonging to at least one scanned image category.
  • the images to be classified may be medical images, including but not limited to CT images and MR images, which are not limited herein.
  • the images to be classified may all be CT images, may all be MR images, and may also be partly CT images and partly MR images, which are not specifically limited herein.
  • CT images and MR images are multi-phase images or multi-sequence imaging. Each phase image or sequence shows different image information of the area where the target object is located or other areas. Combined effectively, the nature of the lesions can be more precisely defined.
  • the images to be classified may be obtained by scanning the abdomen, chest and other regions.
  • the image to be classified obtained by scanning the abdomen may include tissues and organs such as liver, spleen, and kidney
  • the image to be classified obtained by scanning the chest may include tissues and organs such as the heart and lung.
  • the images to be classified may be scanned according to the actual application. Images, not limited here.
  • the target object may be, but is not limited to, a liver tumor and other objects that need to be classified using the image object classification method of the embodiment of the present disclosure.
  • the at least one image to be classified may be a medical image belonging to at least one category of scanned images. Medical images of different scanned image categories can be used to display different characteristic information of target objects, thus improving the accuracy of image target classification.
  • the class of scanned images may also be referred to as the above-described images and/or sequences.
  • the images of the different scanned image categories may be timing-dependent and/or scan-parameter-dependent images.
  • the scan image category may include time-series-related pre-contrast scan, early arterial phase, late arterial phase, portal venous phase, and delayed phase, etc.; alternatively, the scan image category may also include T1-weighted inverse imaging, T1 Weighted in-phase imaging, T2-weighted imaging, diffusion-weighted imaging, and surface diffusion coefficient imaging, etc.
  • the early arterial stage can indicate that the portal vein has not been enhanced
  • the late arterial stage can indicate that the portal vein has been enhanced
  • the portal venous phase can indicate that the portal vein has been sufficiently enhanced and the liver blood vessels have been enhanced by forward blood flow.
  • the delay period can indicate that the portal vein and arteries are in an enhanced state and weaker than the portal venous phase
  • the liver parenchyma is in an enhanced state and weaker than the portal venous phase
  • other scanning image categories will not be listed one by one here.
  • Step S12 Using the classification model, perform target classification on at least one image to be classified to obtain the type of the target object.
  • the classification model is used to classify the at least one image to be classified, so as to obtain the type of the target object.
  • the classification model performs target classification on at least one image to be classified, obtains probabilities that the target objects belong to different types, and uses the types that satisfy the preset probability conditions as the types of the target objects.
  • the preset probability conditions include but are not limited to the maximum probability value and the like.
  • the probability that the target objects belong to different types can be obtained by training the classification model.
  • the batch sample data selected for each training of the classification model is sample data with a preset proportion of the number of different target types selected from the sample data set by the data generator. Since the data generator randomly selects sample data containing equal proportions of different target types as batch sample data, so as to avoid unbalanced classification performance due to too few sample data of a certain target type, the classification model is used for at least one image to be classified.
  • the target classification is obtained by training a large number of batch sample data, which can improve the classification performance of the classification model.
  • Using the classification model to obtain the type of the target object can assist the doctor in determining the type of the target object, save the doctor's time for reviewing the images to be classified, and thus can speed up the output of the report.
  • target classification is performed on at least one image to be classified, and when the type of the target object is obtained, several layers of feature extraction are performed on at least one image to be classified, and several sets of initial feature information are correspondingly obtained; At least one set of initial feature information in the feature information is used to obtain final feature information; the final feature information is classified to obtain the type of the target object.
  • the number of layers for feature extraction may be one layer, two layers or even more layers.
  • which layers to perform feature extraction on can be obtained through artificial settings, or can be determined through a large number of experiments when training a classification model, which is not specifically limited here.
  • a layer of feature extraction is performed on at least one image to be classified, and a set of initial feature information is correspondingly obtained.
  • Multi-layer feature extraction is performed on at least one image to be classified, and multiple sets of initial feature information are correspondingly obtained, wherein the multi-layer feature extraction may be continuous or discontinuous.
  • the initial feature information may be a feature map of the target object, reflecting the feature information of the target object in the image to be classified.
  • the classification model is a deep learning network
  • the deep learning network may include an encoder (encoder) or its variants, Resnet or its variants, and may be a neural network (Visual Geometry Group Network, VGG) 16 or its variants. , or other network model structures for classification.
  • the classification model performs feature extraction on at least one image to be classified through a convolution layer, and different convolution layers correspond to different layers of feature extraction to obtain different groups of initial feature information.
  • the classification model is used to classify the at least one image to be classified, and the type of the target object is obtained. Therefore, an image target classification method based on artificial intelligence technology is proposed, which realizes the Intelligent target classification, and no manual target classification is required, which can reduce the dependence on manual work and improve the efficiency of target classification.
  • At least one image to be classified including liver tumors is acquired, and a classification model is used to perform target classification on at least one image to be classified, so as to obtain the type of liver tumor, and no manual classification is required.
  • the images are classified, and the classification model can be used to realize the classification of liver tumors, so that the doctor can obtain the type of liver tumors.
  • FIG. 2 is a schematic diagram of a system architecture to which the image object classification method according to an embodiment of the present disclosure can be applied; as shown in FIG. 2 , the system architecture includes an image acquisition terminal 201 , a network 202 and an object classification terminal 203 .
  • the image acquisition terminal 201 and the target classification terminal 203 establish a communication connection through the network 202, and the image acquisition terminal 201 reports at least one image to be classified containing the target object to the target classification terminal 203 through the network 202, and the target classification The terminal 203 responds to the received at least one image to be classified, and uses the classification model to perform target classification on the at least one image to be classified to obtain the type of the target object.
  • the target classification terminal 203 uploads the type of the target object to the network 202 and sends it to the image acquisition terminal 201 through the network 202 .
  • the image acquisition terminal 201 may include an image acquisition device, and the target classification terminal 203 may include a vision processing device or a remote server with visual information processing capability.
  • Network 202 may employ wired or wireless connections.
  • the image acquisition terminal 201 can be connected to the visual processing device through a wired connection, such as data communication through a bus; when the target classification terminal 203 is a remote server, the image acquisition terminal 201 can perform data interaction with a remote server through a wireless network.
  • the image acquisition terminal 201 may be a vision processing device with an image acquisition module, which is specifically implemented as a host with a camera.
  • the image object classification method according to the embodiment of the present disclosure may be executed by the image acquisition terminal 201 , and the above-mentioned system architecture may not include the network 202 and the object classification terminal 203 .
  • FIG. 3 is a schematic flowchart of acquiring at least one image to be classified according to an embodiment of the present disclosure. Specifically, the following steps can be included:
  • Step S111 Resampling the original medical image to a preset resolution.
  • the size of the preset resolution can be customized, and the preset resolution corresponding to the target object can be set according to different target objects, so as to unify the resolution of the original medical image to the resolution with the best image effect.
  • Step S112 Adjust the pixel value range in the original medical image.
  • the brightness and color of the original medical image are made easier to display the target object.
  • the categories of the original medical images include, but are not limited to, CT images, MR images, and other images that can reflect the feature information of the target object, which are not limited herein.
  • the original medical image is a CT image
  • the original medical image can be unified to a preset window width and window level;
  • the gray value corresponding to a preset ratio (for example, 99.9%) under the grayscale cumulative distribution function can be used as the normalized preprocessing clamp value, so that the contrast of the MR image data can be enhanced and the subsequent image target can be improved. Classification accuracy.
  • Step S113 Normalize the original medical image.
  • the raw medical images may be normalized.
  • the normalization process includes, but is not limited to, normalizing the intensity or pixel values of the original medical image to a preset range (eg, a range of 0 to 1).
  • Step S114 Detecting the initial area of the first original medical image that is not marked with the target object, using the initial area of the target object marked on the second original medical image and the registration relationship between the second original medical image and the first original medical image, An initial region of the target object on the first original medical image is determined.
  • not all original medical images may be marked with the initial area of the target object. Therefore, in order to use more images to be classified including the target object to perform image object classification and improve the accuracy of image object classification, The initial area of the original medical image can be filled. After detecting the initial area of the first original medical image that is not marked with the target object, use the initial area of the target object marked on the second original medical image and the registration relationship between the second original medical image and the first original medical image to determine the first An initial region of the target object on the original medical image.
  • the above step of determining the initial area of the target object on the first original medical image may be performed by using a registration network.
  • the image target classification method may include several steps from the above steps S111 to S114.
  • the above steps S111 to S114 are only exemplary descriptions. In the disclosed embodiment, several steps can be selected to preprocess the original medical image as required, that is, the number of the above steps S111 to S114 can be arbitrarily selected, which is not specifically limited herein.
  • the original medical image can be preprocessed before the image to be classified is extracted from the original medical image, and the Image parameters to improve the quality of images to be classified.
  • the images to be classified including the target object can be extracted from the multiple original medical images respectively. For details, refer to steps S115 and S116 below.
  • Step S115 Determine the initial area of the target object in the original medical image, and expand the initial area according to a preset ratio to obtain the area to be extracted.
  • the characteristics of the target object itself are the main basis for judging its type, and there may be a variety of noise interference around the target object, which will mislead the classification of the target object.
  • the target object as a liver tumor as an example, the background of chronic liver disease or cirrhosis, other types of tumors, and blood vessels close to the liver tumor will affect the classification accuracy of the target object.
  • the area to be extracted is used as the area to be extracted, so that the area to be extracted contains the target object.
  • the initial area of the target object in the original medical image is determined.
  • the initial area may be expanded according to a preset ratio to obtain the area to be extracted.
  • the initial region is used to delineate the position of the target object in the original medical image.
  • an image segmentation technique can be used to determine the boundary contour of the target object in the original medical image, and mark the boundary contour to form an initial area.
  • Step S116 Extract the image data in the area to be extracted from the original medical image to obtain the image to be classified.
  • the image data is extracted from the original medical image by using the area to be extracted, and the obtained image to be classified includes the target object.
  • the original medical image can be a two-dimensional image or a three-dimensional image.
  • the image to be classified is a two-dimensional image.
  • the image to be classified may be a three-dimensional image, or the image to be classified may be a two-dimensional image.
  • the two-dimensional image of the layer where the target object has the largest area may be, but not limited to, the image to be classified.
  • the original medical image is a three-dimensional image
  • the image to be classified is a two-dimensional image obtained by extracting the layer where the maximum area of the target object is located in the original medical image, so that the layer where the maximum area of the target object is located in the original medical image can be extracted.
  • the extraction range of the target object in the to-be-classified image is larger and contains more information of the target object, thereby improving the classification accuracy of the target object.
  • the initial area of the target object in the original medical image is determined, and the initial area is expanded according to a preset ratio to obtain the area to be extracted; then the area to be extracted is extracted from the original medical image.
  • image data to obtain an image to be classified The initial area is the area containing the target object, and the initial area of the target object is expanded according to a preset ratio, so that the obtained area to be extracted contains both the target object and some background information around the target object, so that the image in the area to be extracted can be extracted.
  • the image to be classified can include the target object and some background information.
  • the to-be-classified images containing the target object are extracted from multiple original medical images respectively, so as to realize the acquisition of the to-be-classified images, and the to-be-classified images can be extracted from the original medical images.
  • the subsequent classification is reduced. It can avoid some background noise in the original medical image to a certain extent, so it can reduce the processing resource consumption of subsequent classification and improve the classification performance.
  • FIG. 4 is a schematic flowchart of a target classification for at least one image to be classified according to an embodiment of the present disclosure. Specifically, the following steps can be included:
  • Step S121 extracting several layers of features on at least one image to be classified, and correspondingly obtaining several sets of initial feature information.
  • the size of each group of initial feature information is different.
  • the number of layers for feature extraction may be one layer, two layers or even more layers.
  • Feature extraction can be implemented by convolutional layers, and each convolutional layer performs feature extraction on at least one image to be classified to obtain initial feature information.
  • which layers to perform feature extraction on can be obtained through manual settings, or can be determined through a large number of experiments when training a classification model, which is not specifically limited here.
  • a layer of feature extraction is performed on at least one image to be classified, and a set of initial feature information is correspondingly obtained, wherein the layer of feature extraction can be any layer, such as but not limited to the initial feature information obtained by extracting the last layer of features. as the basis for subsequent target classification.
  • Multi-layer feature extraction is performed on at least one image to be classified, and multiple sets of initial feature information are correspondingly obtained, wherein the multi-layer feature extraction may be continuous or discontinuous.
  • the initial feature information may be a feature map of the target object, reflecting the feature information of the target object in the image to be classified.
  • the size of each set of initial feature information is different, wherein the size includes dimension and/or resolution, so that the multiple sets of initial feature information respectively reflect different feature information of the target object.
  • the classification model is a deep learning network
  • the included network model structure can be an encoder or its variant, Resnet or its variant, VGG16 or its variant, or other network model structures for classification.
  • the classification model performs feature extraction on at least one image to be classified through a convolution layer, and different convolution layers correspond to different layers of feature extraction to obtain different groups of initial feature information.
  • the target object in the image to be classified, there may be noise interference around the target object.
  • the target object as a liver tumor as an example, the background of chronic liver disease or cirrhosis, other types of tumors, and blood vessels close to the liver tumor will affect the classification accuracy of the target object. Therefore, before using the classification model to perform target classification on at least one image to be classified to obtain the type of the target object, the final area of the target object can be obtained based on the initial area corresponding to the target object in the image to be classified.
  • an initial area can be used as the final area of the target object, or the final area of the target object can be obtained by combining the initial areas corresponding to the target object in at least one image to be classified.
  • the union of the initial regions corresponding to the target object in the image to be classified is regarded as the final region of the target object, which is not limited here.
  • the initial feature information of the image to be classified (such as the global features of the image to be classified, etc.) can be extracted.
  • the weight of the corresponding final area in the image to be classified is higher than the weight of other areas in the image to be classified, which makes the classification model tend to
  • the final region extracts features with richer details, so that the initial feature information output by the classification model corresponding to the final region can be as rich in features as possible; and/or, the features corresponding to the final region in the initial feature information are more abundant than the features of other regions.
  • the classification model is guided to pay more attention to the target object in the final area, so that the classification model can learn the target object.
  • the feature information of itself can reduce the influence of noise interference around the target object on the target classification.
  • a union of the initial regions corresponding to the target object in at least one image to be classified is obtained as the target object. Therefore, several layers of feature extraction can be performed on at least one image to be classified by using the final region, and several sets of initial feature information can be obtained correspondingly.
  • the final area of the target object is the union of the initial areas of the target object in the image to be classified, the final area is greater than or equal to any initial area, ensuring that the final area of the target object can contain the corresponding areas of the target objects in different images to be classified, Therefore, when the feature extraction of the image to be classified is performed, the feature information of the target object can be paid attention to as much as possible.
  • the at least one image to be classified includes a first image to be classified without an initial area of the target object and a second image to be classified with an initial area of the target object marked;
  • the classification model can also be used to detect the initial area of the first image to be classified that is not marked with the target object, and based on the initial area of the target object marked on the second image to be classified and the first image to be classified 2.
  • the registration relationship between the image to be classified and the first image to be classified determines the initial area of the target object on the first image to be classified.
  • the classification model can be used to determine the initial area of the target object for the first to-be-classified image that is not labeled with the initial area of the target object, so as to complete the labeling, so that the to-be-classified images all include the initial area.
  • a final area map including the final area of the target object may be generated, and the final area map and the image to be classified may be input into the classification model, so that the target classification is performed on at least one image to be classified by using the classification model,
  • the final region of the target object included in the final region map to perform several layers of feature extraction on at least one image to be classified can guide the network to pay more attention to the learning of the features of the final region and avoid the network to a certain extent. A lot of wrong feature information is learned, and the interference of noise around the target object on feature extraction is reduced. It can be understood that, before inputting the final area map and the image to be classified into the classification model, the sizes of the final area image and the image to be classified may be adjusted to a uniform size.
  • Step S122 Obtain final feature information based on at least one set of initial feature information in several sets of initial feature information.
  • Several layers of feature extraction are performed on at least one image to be classified, and after several sets of initial feature information are obtained, the final feature information can be obtained based on at least one set of initial feature information in several sets of initial feature information, and the selected initial feature information The information is different, and the final feature information obtained is different.
  • the number of groups of initial feature information and the parameter information such as the convolution layer corresponding to the classification model may be manually set, or may be determined during the training process of the classification model, which is not limited here. Fusion of multiple sets of initial feature information can improve the performance of the classification model and the accuracy of target classification, but the fusion of too much initial feature information will cause overfitting problems.
  • each set of initial feature information is different in dimension and resolution, and reflects different feature information of the target object, at least one set of initial feature information can be fused to obtain the final feature information.
  • the layer high-dimensional feature map is used as the final feature information, after multiple convolutions, some important feature information may be compressed, especially the target object with small area and blurred image features is missed.
  • Feature information the initial feature information obtained in different feature extraction stages can be spliced together to improve the accuracy of image target classification.
  • the weight of at least one set of initial feature information is used to fuse at least one set of initial feature information to obtain final feature information.
  • the weight of each set of initial feature information may be manually set, or may be determined during the training process of the classification model, which is not limited here. For example, first initialize the weight of each group of initial feature information, and continuously update the weight during the training process of the classification model. The above steps of updating weights are continuously repeated by using the training classification model, so that the training classification model continuously learns and updates the weight of each group of initial feature information, and obtains the trained classification model and the weight of each group of initial feature information.
  • the weights of each initial set of initial feature information may be the same or different, and the sum of the weights of each set of initial feature information is 1.
  • the weight of the initial feature information for fusion is determined, so that the final feature information obtained by using the weight fusion can better reflect the characteristics of the target object and further improve the classification performance.
  • the weights of different groups of initial feature information may be the same or different, and the sum of the weights of each group of initial feature information is 1.
  • the weight of the initial feature information can be used to fuse the initial feature information of different sizes extracted from at least one layer of features to obtain the final feature information, considering the initial feature information of smaller size.
  • a feature fusion network can be used to obtain final feature information based on at least one set of initial feature information in several sets of initial feature information, and initial feature information of multiple sizes can be spliced together as the final feature of the classification task.
  • each initial feature information is given a weight, and the weight is continuously updated during the model training process after initialization, so as to integrate multiple initial feature information to obtain a better feature representation of the target object, thereby improving the accuracy of target classification. performance.
  • each set of initial feature information may be converted into a preset dimension to facilitate subsequent acquisition of final feature information.
  • a feature extraction network is used to convert each set of initial feature information into a preset dimension.
  • the preset dimension can be set as required, for example, but not limited to, the preset dimension is one dimension.
  • Step S123 Classify the final feature information to obtain the type of the target object.
  • the final feature information carries the features of the target object, so that the final feature information is classified to obtain the type of the target object.
  • the classification model When determining the type of the target object, including but not limited to, the classification model performs target classification on at least one image to be classified, obtains the probability that the target object belongs to different types, and takes the type that satisfies the preset probability condition as the type of the target object.
  • the preset probability conditions include but are not limited to the maximum probability value and the like.
  • the classification model uses the ArcFace loss function to determine the loss value of the classification model during the training process, and the ArcFace loss function is used to shorten the distance of similar target objects and shorten the distance of heterogeneous target objects, thereby increasing the confusion of target objects. classification ability.
  • the ArcFace loss function is simple and easy to use, and can be well applied to the network structure of the classification model without being combined with other loss functions. At the same time, the overfitting problem is reduced to a certain extent, thereby improving the classification performance of the target object.
  • the training result of the classification model can be the cosine of the angle between the weight of the first fully connected layer and the feature entering the first fully connected layer. value.
  • the dot product between the features entering the first fully-connected layer of the classification model and the weights of the first fully-connected layer can be equal to the normalized cosine distance between the feature and the weight, so that the angular cosine function can be used to calculate the normalized
  • the target object as a liver tumor as an example, considering that the characteristic information of the liver tumor itself is the main basis for judging its type, but the size of the liver tumor varies, ranging from less than 0.5 cm to more than 20 cm, plus the size of the tumor outside the target object. Influencing factors, such as the low resolution of the image to be classified, other types of tumors around the liver tumor, blood vessels with similar characteristics to the target object, chronic liver disease or liver cirrhosis background, etc.
  • the ArcFace loss function can learn better The feature representation of liver tumors can realize the aggregation of similar tumors and the distance of heterogeneous tumors, and can effectively improve the classification performance of tumors.
  • the effect of using the ArcFace loss function to determine the loss value of the classification model in the training process of the classification model is similar, and no examples will be given here.
  • the ArcFace loss function is a loss function that uses margin to expand the distance between different classes.
  • the predicted value is the cosine of the angle between the weight of the first fully connected layer and the feature entering the first fully connected layer.
  • the principle and operation process are as follows: first, the dot product between the feature entering the first fully connected layer and the weight of the first fully connected layer is equal to the cosine distance after the normalization of the feature and the weight, and secondly, using the angular cosine function (arc-cosine function) to calculate the angle between normalized features and normalized weights; then, add an additional angular margin (additive angular margin) to the target angle, and then obtain through the cosine function
  • the logit of the target then rescales all logits with a fixed feature norm, and the subsequent steps are exactly the same as in the softmax loss.
  • the writing order of each step does not mean a strict execution order but constitutes any limitation on the implementation process, and the specific execution order of each step should be based on its function and possible Internal logic is determined.
  • At least one image to be classified is preprocessed, and a corresponding two-dimensional or three-dimensional multi-phase image tumor sub-image block is extracted, that is, a multi-phase image tumor patch image, and corresponding
  • the mask image, the mask patch image is fed together into a deep learning classification network.
  • FIG. 5 it is a schematic diagram of the network architecture used by the classification model in the image target classification method according to the embodiment of the present disclosure; wherein, the batch data input to the classification model 501 randomly includes data of different types of tumors in equal proportions, including stage Like 1, phase like 2, ..., phase like m, and multi-phase like the union of lesion masks.
  • 502 is the CNN backbone network, that is, CNN backbone, which can be the encoder of U-Net or its variant, Resnet or its variant, VGG16 or its variant, or other CNN structures for classification;
  • 503 is Feature Block, which includes Adaptive average pooling, FC and Relu; the previously obtained feature map is subjected to adaptive average pooling, full connection and Relu activation to obtain a one-dimensional feature; at the same time, each Feature Block corresponds to A feature_1.
  • Feature Fusion a feature fusion layer
  • Feature Fusion splices multiple one-dimensional features, each feature has a corresponding weight coefficient, and the coefficient can be learned; including: the weight coefficient of feature_1_ 1, weight coefficient_2 of feature_2, ..., weight coefficient_n of feature_n.
  • the feature map from any convolutional layer in the CNN backbone needs to enter the feature block and feature fusion layer, which can be determined by experimenting in the training process. In the experiment process of this scheme, it is found that the introduction of feature fusion layers can improve the performance of the model; however, fusing too many feature maps will cause over-fitting problems, especially the feature maps obtained by fusing the front convolutional layers.
  • 505 is a fully connected layer (Fully Connected), that is, the fused features are sent to FC, and converted into classification probability values of each tumor category through softmax.
  • 506 is the predicted probability value.
  • FIG. 6 is a schematic frame diagram of an image object classification apparatus 60 provided by an embodiment of the present disclosure.
  • the image object classification device 60 includes an image acquisition module 61 and an object classification module 62 .
  • the image acquisition module 61 is configured to: acquire at least one image to be classified including the target object, wherein at least one image to be classified is a medical image belonging to at least one scanned image category;
  • the target classification module 62 is configured to: use the classification The model performs target classification on at least one image to be classified to obtain the type of the target object.
  • the target classification module 62 is configured to: perform several layers of feature extraction on at least one image to be classified, and correspondingly obtain several sets of initial feature information; wherein, the size of each set of initial feature information is different; At least one group of initial feature information in the group of initial feature information is obtained to obtain final feature information; the final feature information is classified to obtain the type of the target object.
  • the target classification module 62 is configured to: obtain the final area of the target object based on the initial area corresponding to the target object in the image to be classified; correspondingly, the target classification module 62 is configured to: use the final area to At least one image to be classified is subjected to several layers of feature extraction, corresponding to several sets of initial feature information; wherein, in the feature extraction process, the weight of the corresponding final area in the image to be classified is higher than the weight of other areas in the image to be classified; and/ Or, the features corresponding to the final region in the initial feature information are more abundant than the features of other regions.
  • the target classification module 62 is configured to obtain a union of initial regions corresponding to the target object in at least one image to be classified, as the final region of the target object.
  • the target classification module 62 is configured to: use the classification model to detect the initial area of the first image to be classified that is not marked with the target object, and based on the initial area of the target object marked on the second image to be classified and the registration relationship between the second to-be-classified image and the first to-be-classified image to determine the initial area of the target object on the first to-be-classified image.
  • the target classification module 62 is configured to: convert each set of initial feature information into a preset dimension; and/or, the target classification module 62 is configured to: use the weight of at least one set of initial feature information to classify At least one set of initial feature information is fused to obtain final feature information.
  • the weight of each set of initial feature information is determined during the training process of the classification model.
  • the preset dimension is one dimension.
  • the classification model adopts the ArcFace loss function during the training process to determine the loss value of the classification model; and/or, the batch sample data selected for each training of the classification model is selected from the sample data set using a data generator
  • the number of different target types is a preset ratio of sample data.
  • the image acquisition module 61 is configured to: extract images to be classified including the target object from a plurality of original medical images, respectively.
  • the image acquisition module 61 is configured to: determine the initial area of the target object in the original medical image, expand the initial area according to a preset ratio to obtain the area to be extracted; extract the area to be extracted from the original medical image image data to obtain the image to be classified.
  • the image acquisition module 61 is configured to: resample the original medical image to a preset resolution; adjust the range of pixel values in the original medical image; normalize the original medical image; detect The first original medical image is not marked with the initial area of the target object, and the first original medical image is determined by using the initial area of the target object marked on the second original medical image and the registration relationship between the second original medical image and the first original medical image. The initial area of the target object on the image.
  • the original medical image and the image to be classified are two-dimensional images; or, the original medical image is a three-dimensional image, and the image to be classified is a two-dimensional image or a three-dimensional image.
  • the object classification module 62 uses the classification model to perform object classification on the at least one image to be classified to obtain the target object Therefore, an image target classification method based on artificial intelligence technology is proposed to achieve intelligent target classification. Because the classification model is used to classify the images to be classified, it not only makes the target classification process simpler, reduces the dependence on doctors, and improves the speed of target classification, but also combines artificial intelligence technology to achieve target classification, so as to assist doctors in intelligent disease diagnosis and treatment.
  • FIG. 7 is a schematic diagram of a framework of an embodiment of an electronic device 70 of the present disclosure.
  • the electronic device 70 includes a memory 71 and a processor 72 coupled to each other, and the processor 72 is configured to execute program instructions stored in the memory 71 to implement the steps of any of the image object classification method embodiments described above.
  • the electronic device 70 may include, but is not limited to, a microcomputer and a server.
  • the electronic device 70 may also include mobile devices such as a notebook computer and a tablet computer, which are not limited herein.
  • the processor 72 is configured to control itself and the memory 71 to implement the steps of any of the image object classification method embodiments described above.
  • the processor 72 may also be referred to as a central processing unit (Central Processing Unit, CPU).
  • the processor 72 may be an integrated circuit chip with signal processing capability.
  • the processor 72 may also be a general-purpose processor, a digital signal processor (Digital Signal Processor, DSP), an application-specific integrated circuit (Application Specific Integrated Circuit, ASIC), a Field-Programmable Gate Array (Field-Programmable Gate Array, FPGA) or other Programmable logic devices, discrete gate or transistor logic devices, discrete hardware components.
  • a general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
  • the processor 72 may be jointly implemented by an integrated circuit chip.
  • the electronic device 70 after acquiring at least one image to be classified including a target object, uses a classification model to classify the at least one image to be classified, and obtains the type of the target object. Therefore, an artificial intelligence technology-based method is proposed.
  • Image object classification method to achieve intelligent object classification. Because the classification model is used to classify the images to be classified, it not only makes the target classification process simpler, reduces the dependence on doctors, and improves the speed of target classification, but also combines artificial intelligence technology to achieve target classification, so as to assist doctors in intelligent disease diagnosis and treatment.
  • FIG. 8 is a schematic diagram of a framework of an embodiment of a computer-readable storage medium 80 of the present disclosure.
  • the computer-readable storage medium 80 stores program instructions 801 that can be executed by the processor, and the program instructions 801 are used to implement the steps of any of the foregoing image object classification method embodiments.
  • the classification model is used to classify the at least one image to be classified to obtain the type of the target object.
  • the image target classification method of intelligent technology realizes intelligent target classification. Because the classification model is used to classify the images to be classified, it not only makes the target classification process simpler, reduces the dependence on doctors, and improves the speed of target classification, but also combines artificial intelligence technology to achieve target classification, so as to assist doctors in intelligent disease diagnosis and treatment.
  • An embodiment of the present disclosure further provides a computer program, where the computer program includes a computer-readable code, and when the computer-readable code is run in an electronic device, the processor of the electronic device executes the program for implementing any of the foregoing embodiments.
  • Image object classification methods can be specifically implemented by hardware, software or a combination thereof.
  • the computer program product is embodied as a computer storage medium, and in some embodiments of the present disclosure, the computer program product is embodied as a software product, such as a software development kit (Software Development Kit, SDK), etc. .
  • the functions or modules included in the image object classification apparatus provided in the embodiments of the present disclosure may be used to execute the methods described in the above method embodiments, and the specific implementation may refer to the above method embodiments. It is concise and will not be repeated here.
  • the disclosed method and apparatus may be implemented in other manners.
  • the device implementations described above are only illustrative.
  • the division of modules or units is only a logical function division. In actual implementation, there may be other divisions.
  • units or components may be combined or integrated. to another system, or some features can be ignored, or not implemented.
  • the shown or discussed mutual coupling or direct coupling or communication connection may be through some interfaces, indirect coupling or communication connection of devices or units, which may be in electrical, mechanical or other forms.
  • each functional unit in each embodiment of the present disclosure may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit.
  • the above-mentioned integrated units may be implemented in the form of hardware, or may be implemented in the form of software functional units.
  • the integrated unit if implemented as a software functional unit and sold or used as a stand-alone product, may be stored in a computer-readable storage medium.
  • the technical solutions of the present disclosure can be embodied in the form of software products in essence, or the part that contributes to the prior art, or all or part of the technical solutions, and the computer software product is stored in a storage medium , including several instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) or a processor (processor) to execute all or part of the steps of the methods of the various embodiments of the present disclosure.
  • the aforementioned storage medium includes: U disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk or optical disk and other media that can store program codes .
  • the present disclosure provides an image object classification method, device, device, storage medium and program, wherein the image object classification method includes: acquiring at least one image to be classified including a target object, wherein the at least one image to be classified is a medical image belonging to at least one type of scanned image; using a classification model, perform target classification on the at least one image to be classified to obtain the type of the target object.

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

Sont divulgués dans des modes de réalisation un procédé et un appareil de classification d'objet d'image, un dispositif, un support de stockage et un programme Le procédé de classification d'objet d'image consiste : à acquérir au moins une image à classifier comprenant un objet cible, l'au moins une image étant une image médicale appartenant à au moins un type d'images de balayage ; et à réaliser une classification d'objet sur l'au moins une image à l'aide d'un modèle de classification afin d'obtenir le type de l'objet cible. La solution décrite peut être appliquée à une image médicale comprenant au moins une phase d'une tumeur, de manière à déterminer le type de la tumeur dans l'image médicale, c'est-à-dire qu'elle permet d'obtenir une classification d'objet intelligente et d'améliorer l'efficacité de classification d'objet.
PCT/CN2020/139913 2020-11-03 2020-12-28 Procédé et appareil de classification d'objet d'image, dispositif, support de stockage et programme WO2022095258A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202011212261.1 2020-11-03
CN202011212261.1A CN112329844A (zh) 2020-11-03 2020-11-03 图像目标分类方法及相关装置、设备、存储介质

Publications (1)

Publication Number Publication Date
WO2022095258A1 true WO2022095258A1 (fr) 2022-05-12

Family

ID=74324556

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/139913 WO2022095258A1 (fr) 2020-11-03 2020-12-28 Procédé et appareil de classification d'objet d'image, dispositif, support de stockage et programme

Country Status (3)

Country Link
CN (1) CN112329844A (fr)
TW (1) TW202219832A (fr)
WO (1) WO2022095258A1 (fr)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116452957A (zh) * 2023-06-20 2023-07-18 小米汽车科技有限公司 图像标注数据的质量检测方法、装置及电子设备
CN116524135A (zh) * 2023-07-05 2023-08-01 方心科技股份有限公司 一种基于图像的三维模型生成方法及系统
CN117689660A (zh) * 2024-02-02 2024-03-12 杭州百子尖科技股份有限公司 基于机器视觉的保温杯温度质检方法

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114663715B (zh) * 2022-05-26 2022-08-26 浙江太美医疗科技股份有限公司 医学图像质控、分类模型训练方法、装置及计算机设备

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109872306A (zh) * 2019-01-28 2019-06-11 腾讯科技(深圳)有限公司 医学图像分割方法、装置和存储介质
CN111476802A (zh) * 2020-04-09 2020-07-31 山东财经大学 一种基于稠密卷积模型的医学图像分割与肿瘤检测方法,设备及可读存储介质
CN111507381A (zh) * 2020-03-31 2020-08-07 上海商汤智能科技有限公司 图像识别方法及相关装置、设备
CN111755118A (zh) * 2020-03-16 2020-10-09 腾讯科技(深圳)有限公司 医疗信息处理方法、装置、电子设备及存储介质

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110993094B (zh) * 2019-11-19 2023-05-23 中国科学院深圳先进技术研究院 一种基于医学图像的智能辅助诊断方法及终端

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109872306A (zh) * 2019-01-28 2019-06-11 腾讯科技(深圳)有限公司 医学图像分割方法、装置和存储介质
CN111755118A (zh) * 2020-03-16 2020-10-09 腾讯科技(深圳)有限公司 医疗信息处理方法、装置、电子设备及存储介质
CN111507381A (zh) * 2020-03-31 2020-08-07 上海商汤智能科技有限公司 图像识别方法及相关装置、设备
CN111476802A (zh) * 2020-04-09 2020-07-31 山东财经大学 一种基于稠密卷积模型的医学图像分割与肿瘤检测方法,设备及可读存储介质

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116452957A (zh) * 2023-06-20 2023-07-18 小米汽车科技有限公司 图像标注数据的质量检测方法、装置及电子设备
CN116452957B (zh) * 2023-06-20 2023-09-26 小米汽车科技有限公司 图像标注数据的质量检测方法、装置及电子设备
CN116524135A (zh) * 2023-07-05 2023-08-01 方心科技股份有限公司 一种基于图像的三维模型生成方法及系统
CN116524135B (zh) * 2023-07-05 2023-09-15 方心科技股份有限公司 一种基于图像的三维模型生成方法及系统
CN117689660A (zh) * 2024-02-02 2024-03-12 杭州百子尖科技股份有限公司 基于机器视觉的保温杯温度质检方法
CN117689660B (zh) * 2024-02-02 2024-05-14 杭州百子尖科技股份有限公司 基于机器视觉的保温杯温度质检方法

Also Published As

Publication number Publication date
TW202219832A (zh) 2022-05-16
CN112329844A (zh) 2021-02-05

Similar Documents

Publication Publication Date Title
WO2022095258A1 (fr) Procédé et appareil de classification d'objet d'image, dispositif, support de stockage et programme
Cai et al. A review of the application of deep learning in medical image classification and segmentation
Zuo et al. R2AU-Net: attention recurrent residual convolutional neural network for multimodal medical image segmentation
WO2021196955A1 (fr) Procédé de reconnaissance d'image et appareil associé et dispositif
CN110188792B (zh) 前列腺mri三维图像的图像特征获取方法
JP2020518915A (ja) 自動眼底画像分析用のシステムおよび方法
Luo et al. Micro-vessel image segmentation based on the AD-UNet model
Zhao et al. SCOAT-Net: A novel network for segmenting COVID-19 lung opacification from CT images
CN111476793B (zh) 动态增强磁共振成像处理方法、系统、存储介质、终端
WO2022062590A1 (fr) Procédé et appareil de reconnaissance d'image, dispositif, support d'enregistrement et programme
JP2023540910A (ja) 病変検出のための共同訓練を伴う接続機械学習モデル
CN110956632B (zh) 钼靶图像中胸大肌区域自动检测方法及装置
CN112001895B (zh) 一种甲状腺钙化检测装置
WO2019184851A1 (fr) Procédé et appareil de traitement d'image, et procédé d'entraînement pour modèle de réseau neuronal
Tan et al. Automated vessel segmentation in lung CT and CTA images via deep neural networks
CN114782307A (zh) 基于深度学习的增强ct影像直肠癌分期辅助诊断系统
CN111260651A (zh) 一种基于深度迁移学习的胃部低质量mri图像分割方法
WO2019223121A1 (fr) Procédé et appareil de reconnaissance de site de lésion, appareil informatique et support de stockage lisible
CN111062909A (zh) 乳腺肿块良恶性判断方法及设备
CN116109610A (zh) 一种超声检查报告图像中乳腺肿瘤分割方法和系统
Rodríguez et al. Color segmentation applied to study of the angiogenesis. Part I
Zhou et al. Eso-net: A novel 2.5 D segmentation network with the multi-structure response filter for the cancerous esophagus
Sridhar et al. Lung Segment Anything Model (LuSAM): A Prompt-integrated Framework for Automated Lung Segmentation on ICU Chest X-Ray Images
CN113870194A (zh) 深层特征和浅层lbp特征融合的乳腺肿瘤超声图像处理装置
CN114972192B (zh) 一种基于深度学习的乳腺钼靶胸大肌区域分割方法

Legal Events

Date Code Title Description
ENP Entry into the national phase

Ref document number: 2021578180

Country of ref document: JP

Kind code of ref document: A

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20960715

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 04.10.2023)