CN113469942B - CT image lesion detection method - Google Patents

CT image lesion detection method Download PDF

Info

Publication number
CN113469942B
CN113469942B CN202110608053.1A CN202110608053A CN113469942B CN 113469942 B CN113469942 B CN 113469942B CN 202110608053 A CN202110608053 A CN 202110608053A CN 113469942 B CN113469942 B CN 113469942B
Authority
CN
China
Prior art keywords
network
loss
regression
rcnn
rpn
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110608053.1A
Other languages
Chinese (zh)
Other versions
CN113469942A (en
Inventor
侯永宏
刘传玉
李岳阳
王拓
苏晓雨
郭子慧
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sichuan Xuanguang Liying Medical Technology Co Ltd
Original Assignee
Tianjin University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianjin University filed Critical Tianjin University
Priority to CN202110608053.1A priority Critical patent/CN113469942B/en
Publication of CN113469942A publication Critical patent/CN113469942A/en
Application granted granted Critical
Publication of CN113469942B publication Critical patent/CN113469942B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/20ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10081Computed x-ray tomography [CT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30096Tumor; Lesion

Abstract

The invention provides a CT image lesion detection method, which is used for preprocessing and enhancing data of a medical CT image; the method comprises the steps of constructing a network model for detecting lesions of multiple continuous slices, wherein the model is based on a Mask-RCNN network and comprises a Convolutional Neural Network (CNN), an RPN network and an RCNN network, detecting the lesions in two stages, constructing a multi-scale local axial self-attention MSLASA module in the Convolutional Neural Network (CNN), and realizing information combination of pixel points in different local areas by using the MSLASA module, so that each pixel point in a characteristic diagram can obtain a self weight value according to a proper local space. Through an attention mechanism, information related to pathological changes is enhanced and highlighted to a great extent, and non-pathological features are subjected to desalination and filtration, so that the task of network discrimination at the next stage is relieved, and the number of false positive samples is reduced to a great extent.

Description

CT image lesion detection method
Technical Field
The invention belongs to the fields of computer vision, deep learning and the like, relates to a target detection technology, and particularly relates to a CT image lesion detection method.
Background
Lesion detection in Computed Tomography (CT) images is an important basis and guarantee for doctors in clinical practice, in hospitals, doctors analyze lesions in given CT images based on their own medical knowledge and daily empirical interpretation, and the accuracy of film reading is limited by the experience and knowledge levels of different doctors, thus causing differences. Moreover, for reading a large number of CT pictures, the doctor can influence the mental state to cause misjudgment and the like after working for a long time. Therefore, for the work of taking a lot of time and energy of the radiologist, the present invention hopes to reduce the burden of the radiologist by using a computer-aided system (CAD) to provide objective and real-time reference information to assist the physician in judging the lesion, and the present invention believes that the present invention is a trend of modern technology development, and the CAD system will be used as the best tool of the physician to provide quasi-safe, convenient and fast diagnosis and treatment for the patient in the future.
The advent of Convolutional Neural Networks (CNN) has made it possible to obtain better picture feature representations for natural pictures under their processing, and thus, there is a great breakthrough in the results of detection and segmentation tasks. The medical picture and the natural picture have certain field difference, the main embodiment is that the natural picture is mostly a scene picture in daily life of people, the picture characteristic is obvious, the prospect to be detected generally has higher discrimination, and at least the naked eye can judge rapidly. However, the lesion to be detected and the surrounding area in the medical CT image have high similarity, and even most non-professional persons cannot locate the lesion position from the similarity. The convolutional neural network is used for extracting picture characteristics of the medical image and then carrying out further lesion location and classification, and the powerful CNN is also greatly improved on the medical image and can detect most lesion positions. Most studies use CNNs to detect lesions in specific areas, such as the abdomen, chest or pelvis. However, the present invention expects that the CNN-based CAD system can process CT images from any part, that is, for CT medical images from all parts of the human body, the CNN should well extract image features for further classification and detection processing, which brings great challenges to the CNN designed for natural images.
Although the method of deep learning is able to detect the vast majority of lesions. However, due to the high similarity between the lesion in the CT image and its surrounding background, many non-lesion features passing through the CNN can be misjudged as lesions, resulting in many false positive samples.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and provide a method for detecting lesions in a CT image, which can obtain the pixel weight value of the image characteristic through local attention and enhance the discrimination of the CT image characteristic, thereby reducing the number of false positive samples obtained by a network.
The innovation points of the invention are as follows:
the structure innovation in the convolutional neural network model is that the shallow feature map is used for obtaining the weight values of feature spaces under different levels, and the weight values are applied to the feature map after the shallow feature map and the deep feature map are combined.
The MSLASA module is utilized to realize information combination of the pixel points in different local areas, so that each pixel point in the characteristic diagram can acquire a weight value of the pixel point according to a proper local space.
In the self-attention module, the square area of the local space is converted into high-axis and wide-axis pixel points, so that the calculation complexity is O (N)2) Reduced to o (2N).
And adding position information to the query value and the key value in the self-attention module, so that the spatial information is better reflected.
The technical scheme for realizing the purpose of the invention is as follows:
a CT image lesion detection method comprises the following steps:
(1) the medical CT image is preprocessed and data enhanced, and the window width of each CT slice is converted to be (-1024) -3071HU) so as to cover the characteristic representation of a plurality of positions. The size of each CT slice is basically (512 ), and the data enhancement is realized by cutting, scaling and rotating the picture.
(2) And constructing a network model for detecting the lesion of the plurality of continuous slices, wherein the model is based on a Mask-RCNN network and comprises a Convolutional Neural Network (CNN), an RPN network and an RCNN network. The model has accurate effect by detecting the lesion in two stages.
(3) And constructing a multi-scale local axial self-attention MSLASA module in the convolutional neural network CNN, and enhancing the identification degree between pathological changes and non-pathological changes in the CT image. The CNN adopts a characteristic pyramid form, obtains shallow features and deep features of the CT picture through Bottom-up and Top-down Top-down network modes, and establishes fusion of the deep features and the shallow features through jump connection. In addition, with the shallow features as input, feature weights of the CT image are acquired in the local axial self-attention module mslas at multiple scales and mapped into the output features.
(4) The RPN is responsible for regression and classification of the detection frames in the first stage in the model, a fixed number of anchors are generated at each point in the feature map according to the size of the feature map and the size and proportion of a preset rectangular frame, and the RPN predicts the bias and category regression of each anchor to obtain the regression frame in the first stage.
(5) And selecting a part of the score according to the score of the score in the anchors obtained in the first stage, and sending the part of the score to the RCNN for regression and classification in the second stage. And obtaining the optimal model parameters.
(6) The network model respectively calculates the loss once in two stages, the loss of each stage is regression loss and classification loss, the final loss function is the sum of the loss of the two stages, and the loss function is as follows:
class loss:
Figure GDA0003462054370000021
wherein L isclsRepresents the class loss, n represents the total number of samples, x represents the samples, yiRepresenting the actual label and a the predicted output.
Regression loss:
Figure GDA0003462054370000031
wherein L isregRepresents regression loss, n represents total number of samples, x represents samples, y represents actual coordinates, and f (x) represents predicted coordinate values.
Total loss:
L=LRPN,cls+LRPN,reg+LRCNN,cls+LRCNN,reg
where L is the final loss of the network model. L isRPN,clsAnd LRPN,regRespectively refer to classification and regression in the first stage RPN networkloss,LRCNN,clsAnd LRCNN,regRespectively, classification and regression loss in the second stage RCNN network.
(7) Under the two-stage processing, the model outputs the category and the coordinates of the regression frame which is finally predicted, outputs the accuracy of the experiment under different FPs, and outputs the CT medical picture of the detection regression frame which is additionally predicted.
The invention has the advantages and beneficial effects that:
(1) the lesion detection in the CT image based on the attention mechanism is applied to processing the CNN in research, so that the features of the obtained CT image can have higher discrimination, the information related to the lesion is enhanced and highlighted to a great extent through the attention mechanism, and the non-lesion features are subjected to desalination and filtration, so that the task of network discrimination in the next stage is reduced, and the number of false positive samples is reduced to a great extent.
(2) The present invention proposes to adaptively combine multiple local regions with different proportions for each pixel using local self-attention modules at multiple scales to find the best fusion information. The module utilizes the characteristic information under different scales, brings obvious progress in the lesion detection of the CT image, and can be applied to more fields.
(3) The present invention centers on each pixel and assigns a certain number of squares to them to provide local self-attention. In view of computational complexity and memory limitations, the present invention selects 1D axial self-attention instead of 2D axial self-attention.
(4) The present invention develops a position sensitive module to enrich the spatial information of each pixel. Then, similar to the radiologist's diagnostic procedure, the mslas a module of the present invention can help the CNN to exclude many redundant pixels and capture a better characterization. The MSLASA module introduces few parameters and has low calculation overhead, thereby greatly improving the detection precision. Numerous experiments performed on the deep version basis demonstrated the effectiveness of the mslas module. More importantly, the invention provides a new method for implementing local self-attention to a plurality of local spaces.
Drawings
FIG. 1 is a network overall structure diagram of a CT image lesion detection method according to an embodiment of the present invention;
FIG. 2 is a model diagram of a multi-scale local axial self-attention module;
FIG. 3a is a graph of the results of a MULAN method on a picture of a human abdomen;
FIG. 3b is a graph of the results of the MULAN method on a picture of the liver of a human;
FIG. 3c is a graph of the result of the MULAN method on a human pelvis image;
FIG. 4a is a graph showing the results of a test conducted on a picture of the abdomen of a human body according to the method of the present invention;
FIG. 4b is a graph showing the results of the test of the present invention on a picture of the liver;
FIG. 4c is a graph showing the results of the test of the method of the present invention on a human pelvis image.
Detailed Description
The present invention is further illustrated by the following specific examples, which are intended to be illustrative, not limiting and are not intended to limit the scope of the invention.
The invention provides a method for detecting lesions in a CT image, wherein a network used is trained end to end.
As shown in FIG. 1, embodiments of the overall structure of the present invention are described in the following 1 to 5.
1) Since the CT image is a single-channel slice unlike the natural image, the present invention combines 3 single-channel slices as a 3-channel picture, so that the pretrained model of ImageNet can be used in the CNN network, and the data set has 32735 labeled lesion instances, 70% of which is used as the training set, 15% of which is used as the validation set, and the remaining 15% of which is used as the test set.
2) And constructing a feature extraction network model, wherein the network takes DenseNet-121 as a dense residual block as a basic component element and uses a DenseNet-121 pre-training model. Considering that the shallow feature has a larger size to display more picture details and the deep feature has more high-level semantic information to display more picture details as the convolutional neural network deepens, both of which have important roles in judging the pathological changes in the CT picture, the invention adopts a network structure of a Feature Pyramid (FPN) and uses jump connection to fuse the shallow feature and the deep semantic information to obtain a comprehensive feature map. In addition, the invention designs a self-attention Module (MSLASA), obtains weight information on a feature map space through local attention of shallow features, and maps the weight information on the feature map after the fusion of the shallow features and the deep features. The self-attention module will be described with emphasis in fig. 2.
3) And processing the picture characteristics acquired by the network as the input of the RPN network. First, for each point on the feature map, a fixed number of anchors are generated according to the proportion and area of the preset rectangular box. Meanwhile, the feature map is sent to an RPN network to obtain the corresponding category and regression bias of each rectangular frame, so that the invention can obtain the corresponding category and correction coordinate according to the basic coordinate and bias of each rectangular frame, and then 2000 proposals with the maximum probability can be selected by using non-maximum suppression (NMS) to be used as the detection of the first stage.
4) Rpn, finding a region with corresponding size and position on the feature map as a feature of the propsall, sending the region to ROI Align to obtain a feature with a fixed size, similarly obtaining a category score and a regression bias of the propsall through activation functions such as convolution and relu of the RCNN network, and correcting coordinates of the propsals again to obtain a second-stage final coordinate output.
5) It should be noted that not all anchors and propofol participate in the calculation of loss during the two-stage loss calculation. Under the assistance of the labeling information, the anchors obtained by the RPN can obtain the label of each anchor by calculating the cross-over ratio of the anchor to the label (the cross-over ratio is more than 0.5 and is set as 1, otherwise, the cross-over ratio is set as 0), and 32 samples are selected from the samples with the labels of 0 and 1 according to the ratio of 1:1 to calculate the loss. The second stage of RCNN also performs the above processing on propofol in calculating loss, except that the total number of positive and negative samples selected is 64. In addition, both positive and negative samples will participate in the calculation of the classification loss, but only the positive sample will participate in the calculation of the regression loss.
As shown in FIG. 2, the present invention introduces a multi-scale local axial self attention (MSLASA) module implementation. It should be noted that the feature map is used as an input of the mslas a module, and the mslas a module obtains a pixel weight value of each point in the feature map according to local information of a plurality of scales around each pixel point by taking the pixel point as a center.
1) And (3) carrying out 1x1 convolution on the input feature diagram F belonging to BxCxHxW (B: batch size, C: channel, H: height, W: width) to obtain the query value of each point with the feature diagram size unchanged and the channel number C', and dividing the feature diagram into 4 parts according to the channel number, wherein the obtained query value corresponds to (3,5,7,9)4 different scales. And simultaneously, dividing the feature map into 4 equal parts in the channel dimension, corresponding to 4 different scales, performing edge filling on each feature map according to the different scales, wherein the filling is 0, and then convolving each feature map to obtain a key value corresponding to a pixel point under each scale.
2) Because the local information key value corresponding to the query value of each pixel point is a plurality of squares taking the local information key value as the center, the calculation complexity and the subsequent calculation amount are higher, when the key value range is selected, for each pixel point, the invention only uses the high axis and the wide axis which pass through the pixel point and are vertical to the side length of the square, and the complexity is O (N)2) Reduced to O (2N).
3) And respectively adding position information to the query and the key in consideration of the influence of the spatial position information on the query and the key. For query, the coordinates of the upper left corner and the lower right corner of the whole feature map are (-1, -1) and (1,1), position coordinates of each point of the whole feature map are obtained, the position embedding of the feature map is set to be 1x2xHxW (2 represents horizontal and vertical coordinates), the position embedding is carried out with 1x1 convolution, the dimension size and the channel number are not changed, and the query value of each pixel point of the feature map is added with the value obtained after the convolution of the horizontal and vertical coordinates of the corresponding position to serve as a final query value. For the position information of the key value, only the high axis and the wide axis of 3,5,7 and 9 are adopted by the invention. Therefore, for each scale (N is 3,5,7,9), the present invention sets one end to be-1, the other end to be 1, and the position embedding to be 1x1xN, and performs 1-dimensional convolution on the position information to obtain position information, which is added to the key values on the horizontal axis and the vertical axis as the final key value.
4) In each scale, the query value of each point in the feature map and the corresponding key value are subjected to point multiplication, then the result is divided by the number of pixels of the high axis and the wide axis to be taken as a mean value, and the feature maps in all scales are cascaded on a channel to obtain a new feature map G.
5) Through convolution of 1x1, the size of the feature graph G is not changed, the number of channels is reduced to 1, and then relu processing is carried out, so that the final weight information feature graph is obtained.
6) The concrete implementation formula is as follows:
Figure GDA0003462054370000061
wherein xoFinal output weight, f, representing o pointConvRepresents a convolution of 1x1 + relu function, phi represents a channel cascade, p represents a local region at different scales,
Figure GDA0003462054370000062
representing the calculated quantity at different scales of o points, qoRepresents the query value of the o point,
Figure GDA0003462054370000063
and
Figure GDA0003462054370000064
representing the location information of the horizontal and vertical coordinates, k, corresponding to the querypRepresenting the key values of the pixel points at different scales,
Figure GDA0003462054370000065
representing the position information corresponding to the key.
The following is a description of the experimental results of the present invention on the deepversion dataset and the real images:
the invention trains and predicts under a Pythrch deep learning framework. Training and testing were performed on the DeepLesion dataset with sensitivity sensitivities with different FPs as the standard. The results of the experiments on the data set are shown in the table below
Figure GDA0003462054370000066
It can be seen that the method provided by the invention can obtain results superior to other methods in evaluation indexes. The present invention was tested in real images and visually compared to other mainstream algorithms, as shown in figures 3 a-3 c, and figures 4 a-4 c below.
FIGS. 3 a-3 c show the results of the MULAN method, and FIGS. 4 a-4 c show the results of the experiments according to the present invention. The solid line frame represents the real lesion label, and the dotted line frame represents the predicted lesion position and the predicted score, so that the method reduces a plurality of false positive samples while accurately predicting.
The foregoing is only a preferred embodiment of the present invention, and it should be noted that, for those skilled in the art, various changes and modifications can be made without departing from the inventive concept, and these changes and modifications are all within the scope of the present invention.

Claims (5)

1. A CT image lesion detection method comprises the following steps:
preprocessing and data enhancing the medical CT image;
constructing a network model for detecting the lesion of a plurality of continuous slices, wherein the model is based on a Mask-RCNN network and comprises a Convolutional Neural Network (CNN), an RPN network and an RCNN network, and the model is divided into two stages for detecting the lesion;
constructing a multi-scale local axial self-attention MSLASA module in a Convolutional Neural Network (CNN), wherein the CNN adopts a characteristic pyramid form, obtains shallow features and deep features of a CT picture in a Bottom-up and Top-down (Top-down) network mode, establishes fusion of the deep features and the shallow features through jump connection, takes the shallow features as input, and obtains the characteristic weight of the CT image in the multi-scale local axial self-attention module MSLASA and maps the characteristic weight to output features;
the RPN is responsible for regression and classification of the detection frames in the first stage in the model, a fixed number of anchors are generated at each point in the feature map according to the size of the feature map and the size and proportion of a preset rectangular frame, and the RPN predicts the bias and category regression of each anchor to obtain the regression frame in the first stage; selecting a part of the score according to the score in the anchors obtained in the first stage, and sending the part of the score into an RCNN network for regression and classification in the second stage to obtain the optimal model parameters;
under the two-stage processing, the model outputs the category and the coordinates of the regression frame which is finally predicted, outputs the accuracy of the experiment under different FPs, and outputs the CT medical picture of the detection regression frame which is additionally predicted.
2. The detection method according to claim 1, characterized in that: the RPN and the RCNN respectively calculate the loss once in two detection stages, the loss in each stage is regression loss and classification loss, the final loss function is the sum of the losses in the two stages, and the loss function is as follows:
class loss:
Figure FDA0003462054360000011
wherein L isclsRepresents the class loss, n represents the total number of samples, x represents the samples, yiRepresents the actual label, a represents the predicted output;
regression loss:
Figure FDA0003462054360000012
wherein L isregRepresents regression loss, n represents total number of samples, x represents samples, y represents actual coordinates, and f (x) represents predicted coordinate values;
total loss:
L=LRPN,cls+LRPN,reg+LRCNN,cls+LRCNN,reg
where L is the final loss of the network model, LRPN,clsAnd LRPN,regRespectively refer to classification and regression loss, L in the first stage RPN networkRCNN,clsAnd LRCNN,regRespectively, classification and regression loss in the second stage RCNN network.
3. The detection method according to claim 1, characterized in that: the characteristic graph is used as the input of the MSLASA module, the MSLASA module takes each pixel point as the center, and the pixel weight value of each point in the characteristic graph is obtained according to local information of the MSLASA module under a plurality of scales around the MSLASA module.
4. The detection method according to claim 3, characterized in that: the specific implementation steps of the multi-scale local axial self-attention module MSLASA are as follows:
equally dividing a feature graph input into the MSLASA module into a plurality of parts in a channel dimension, correspondingly calculating feature information under a plurality of different scales, and performing fusion calculation on each pixel point and local pixels under the peripheral fixed scale of the pixel point to obtain fusion information of the pixel under the scale;
firstly, carrying out 1x1 convolution on the characteristic graph to obtain a characteristic display query value of each pixel point, simultaneously carrying out 1x1 convolution on the characteristic graph to obtain characteristic key values of different positions, distributing a local area under a fixed scale with the query value as a center to each pixel point on the characteristic graph, carrying out dot multiplication on the query value of each pixel point and the key of the corresponding local area, averaging according to the number of the local pixel points, taking the output of each point as characteristic output under the local area, selecting the pixel points on a horizontal axis and a vertical axis which pass through the center of a square and are perpendicular to the side length of the square as the local area, and adding position embedding information to the pixel points;
each feature map is processed, feature outputs under various scales are obtained for each pixel point, 4 feature maps are cascaded on a channel to serve as one feature map, and features under various scales are fused through convolution of 1x1 and a Relu activation function to obtain weight outputs of each point.
5. The detection method according to claim 4, characterized in that: the implementation formula of the local axial self-attention module mslas at multiple scales is as follows:
Figure FDA0003462054360000021
wherein xoFinal output weight, f, representing o pointConvRepresents a convolution of 1x1 + Relu activation function, phi represents channel concatenation,prepresenting local regions at different scales of a scale,
Figure FDA0003462054360000022
representing the calculated quantity at different scales of o points, qoRepresents the query value of the o point,
Figure FDA0003462054360000023
and
Figure FDA0003462054360000024
representing the location information of the horizontal and vertical coordinates, k, corresponding to the querypRepresenting the key values of the pixel points at different scales,
Figure FDA0003462054360000025
representing the position information corresponding to the key.
CN202110608053.1A 2021-06-01 2021-06-01 CT image lesion detection method Active CN113469942B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110608053.1A CN113469942B (en) 2021-06-01 2021-06-01 CT image lesion detection method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110608053.1A CN113469942B (en) 2021-06-01 2021-06-01 CT image lesion detection method

Publications (2)

Publication Number Publication Date
CN113469942A CN113469942A (en) 2021-10-01
CN113469942B true CN113469942B (en) 2022-02-22

Family

ID=77871995

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110608053.1A Active CN113469942B (en) 2021-06-01 2021-06-01 CT image lesion detection method

Country Status (1)

Country Link
CN (1) CN113469942B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114639156B (en) * 2022-05-17 2022-07-22 武汉大学 Depression angle face recognition method and system based on axial attention weight distribution network

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11158048B2 (en) * 2019-06-28 2021-10-26 Shandong University Of Science And Technology CT lymph node detection system based on spatial-temporal recurrent attention mechanism
CN111401201B (en) * 2020-03-10 2023-06-20 南京信息工程大学 Aerial image multi-scale target detection method based on spatial pyramid attention drive
CN111429447A (en) * 2020-04-03 2020-07-17 深圳前海微众银行股份有限公司 Focal region detection method, device, equipment and storage medium
CN111915613B (en) * 2020-08-11 2023-06-13 华侨大学 Image instance segmentation method, device, equipment and storage medium
CN111967399A (en) * 2020-08-19 2020-11-20 辽宁科技大学 Improved fast RCNN behavior identification method

Also Published As

Publication number Publication date
CN113469942A (en) 2021-10-01

Similar Documents

Publication Publication Date Title
US11887311B2 (en) Method and apparatus for segmenting a medical image, and storage medium
Kim et al. Machine-learning-based automatic identification of fetal abdominal circumference from ultrasound images
Ni et al. Standard plane localization in ultrasound by radial component model and selective search
CN109363699B (en) Method and device for identifying focus of breast image
CN111340827A (en) Lung CT image data processing and analyzing method and system
CN109363697B (en) Method and device for identifying focus of breast image
Ni et al. Selective search and sequential detection for standard plane localization in ultrasound
Płotka et al. Deep learning fetal ultrasound video model match human observers in biometric measurements
CN113782184A (en) Cerebral apoplexy auxiliary evaluation system based on facial key point and feature pre-learning
Xiao et al. A cascade and heterogeneous neural network for CT pulmonary nodule detection and its evaluation on both phantom and patient data
Wen et al. Pulmonary nodule detection based on convolutional block attention module
CN113469942B (en) CT image lesion detection method
Nayan et al. A deep learning approach for brain tumor detection using magnetic resonance imaging
Li et al. Developing an image-based deep learning framework for automatic scoring of the pentagon drawing test
CN105551042B (en) A kind of scanning bed mark point method for determining position and device
CN114757873A (en) Rib fracture detection method and device, terminal equipment and readable storage medium
WO2022110525A1 (en) Comprehensive detection apparatus and method for cancerous region
CN111724356B (en) Image processing method and system for CT image pneumonia recognition
CN111275754B (en) Face acne mark proportion calculation method based on deep learning
CN112529900A (en) Method, device, terminal and storage medium for matching ROI in mammary gland image
Bagheri et al. Semantic segmentation of lesions from dermoscopic images using Yolo-DeepLab networks
Zhou et al. A universal approach for automatic organ segmentations on 3D CT images based on organ localization and 3D GrabCut
Guo et al. LesionTalk: Core Data Extraction and Multi-class Lesion Detection in IoT-based Intelligent Healthcare
CN114332858A (en) Focus detection method and device and focus detection model acquisition method
Cudek et al. Automatic system for classification of melanocytic skin lesions based on images recognition

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right

Effective date of registration: 20220804

Address after: 071000 200m north of the intersection of Dingzhou commercial street and Xingding Road, Baoding City, Hebei Province (No. 1910, 19th floor, building 3, Jueshan community)

Patentee after: Hebei Kaitong Information Technology Service Co.,Ltd.

Address before: No.92 Weijin Road, Nankai District, Tianjin 300071

Patentee before: Tianjin University

Effective date of registration: 20220804

Address after: 1F8-14, 2F8-10, Building 18, "Strait Small and Medium Enterprise Incubation Park", No. 188, West Section of Kexing Road, Chengdu Cross-Strait Science and Technology Industry Development Park, Wenjiang District, Chengdu City, Sichuan Province, China

Patentee after: Sichuan Xuanguang Liying Medical Technology Co., Ltd.

Address before: 071000 200m north of the intersection of Dingzhou commercial street and Xingding Road, Baoding City, Hebei Province (No. 1910, 19th floor, building 3, Jueshan community)

Patentee before: Hebei Kaitong Information Technology Service Co.,Ltd.

TR01 Transfer of patent right